{ "best_metric": null, "best_model_checkpoint": null, "epoch": 44.91017964071856, "eval_steps": 500, "global_step": 30000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0014970059880239522, "grad_norm": 12.25, "learning_rate": 0.0008, "loss": 3.8621, "step": 1 }, { "epoch": 0.0029940119760479044, "grad_norm": 147.0, "learning_rate": 0.0008, "loss": 17.8731, "step": 2 }, { "epoch": 0.004491017964071856, "grad_norm": 19.375, "learning_rate": 0.0008, "loss": 3.9524, "step": 3 }, { "epoch": 0.005988023952095809, "grad_norm": 39.25, "learning_rate": 0.0008, "loss": 9.6708, "step": 4 }, { "epoch": 0.0074850299401197605, "grad_norm": 8.5, "learning_rate": 0.0008, "loss": 2.3872, "step": 5 }, { "epoch": 0.008982035928143712, "grad_norm": 25.0, "learning_rate": 0.0008, "loss": 3.4757, "step": 6 }, { "epoch": 0.010479041916167664, "grad_norm": 6.6875, "learning_rate": 0.0008, "loss": 2.1171, "step": 7 }, { "epoch": 0.011976047904191617, "grad_norm": 29.0, "learning_rate": 0.0008, "loss": 3.6517, "step": 8 }, { "epoch": 0.01347305389221557, "grad_norm": 33.5, "learning_rate": 0.0008, "loss": 4.0679, "step": 9 }, { "epoch": 0.014970059880239521, "grad_norm": 15.1875, "learning_rate": 0.0008, "loss": 2.6297, "step": 10 }, { "epoch": 0.016467065868263474, "grad_norm": 5.4375, "learning_rate": 0.0008, "loss": 2.1564, "step": 11 }, { "epoch": 0.017964071856287425, "grad_norm": 6.5, "learning_rate": 0.0008, "loss": 2.2034, "step": 12 }, { "epoch": 0.019461077844311378, "grad_norm": 0.98828125, "learning_rate": 0.0008, "loss": 2.0095, "step": 13 }, { "epoch": 0.020958083832335328, "grad_norm": 5.5, "learning_rate": 0.0008, "loss": 2.2219, "step": 14 }, { "epoch": 0.02245508982035928, "grad_norm": 3.359375, "learning_rate": 0.0008, "loss": 2.1824, "step": 15 }, { "epoch": 0.023952095808383235, "grad_norm": 2.765625, "learning_rate": 0.0008, "loss": 2.0574, "step": 16 }, { "epoch": 0.025449101796407185, "grad_norm": 1.9296875, "learning_rate": 0.0008, "loss": 2.0102, "step": 17 }, { "epoch": 0.02694610778443114, "grad_norm": 2.171875, "learning_rate": 0.0008, "loss": 2.0261, "step": 18 }, { "epoch": 0.02844311377245509, "grad_norm": 1.3671875, "learning_rate": 0.0008, "loss": 1.9611, "step": 19 }, { "epoch": 0.029940119760479042, "grad_norm": 2.078125, "learning_rate": 0.0008, "loss": 1.9503, "step": 20 }, { "epoch": 0.03143712574850299, "grad_norm": 1.4765625, "learning_rate": 0.0008, "loss": 1.9254, "step": 21 }, { "epoch": 0.03293413173652695, "grad_norm": 1.671875, "learning_rate": 0.0008, "loss": 1.9061, "step": 22 }, { "epoch": 0.0344311377245509, "grad_norm": 1.578125, "learning_rate": 0.0008, "loss": 1.875, "step": 23 }, { "epoch": 0.03592814371257485, "grad_norm": 0.6171875, "learning_rate": 0.0008, "loss": 1.8203, "step": 24 }, { "epoch": 0.0374251497005988, "grad_norm": 0.98828125, "learning_rate": 0.0008, "loss": 1.7867, "step": 25 }, { "epoch": 0.038922155688622756, "grad_norm": 1.296875, "learning_rate": 0.0008, "loss": 1.7805, "step": 26 }, { "epoch": 0.040419161676646706, "grad_norm": 1.25, "learning_rate": 0.0008, "loss": 1.7473, "step": 27 }, { "epoch": 0.041916167664670656, "grad_norm": 1.1875, "learning_rate": 0.0008, "loss": 1.7456, "step": 28 }, { "epoch": 0.04341317365269461, "grad_norm": 1.125, "learning_rate": 0.0008, "loss": 1.776, "step": 29 }, { "epoch": 0.04491017964071856, "grad_norm": 0.859375, "learning_rate": 0.0008, "loss": 1.722, "step": 30 }, { "epoch": 0.04640718562874251, "grad_norm": 0.85546875, "learning_rate": 0.0008, "loss": 1.6853, "step": 31 }, { "epoch": 0.04790419161676647, "grad_norm": 0.52734375, "learning_rate": 0.0008, "loss": 1.649, "step": 32 }, { "epoch": 0.04940119760479042, "grad_norm": 0.482421875, "learning_rate": 0.0008, "loss": 1.6499, "step": 33 }, { "epoch": 0.05089820359281437, "grad_norm": 0.41015625, "learning_rate": 0.0008, "loss": 1.6225, "step": 34 }, { "epoch": 0.05239520958083832, "grad_norm": 0.419921875, "learning_rate": 0.0008, "loss": 1.6077, "step": 35 }, { "epoch": 0.05389221556886228, "grad_norm": 0.4296875, "learning_rate": 0.0008, "loss": 1.5821, "step": 36 }, { "epoch": 0.05538922155688623, "grad_norm": 0.39453125, "learning_rate": 0.0008, "loss": 1.5864, "step": 37 }, { "epoch": 0.05688622754491018, "grad_norm": 0.482421875, "learning_rate": 0.0008, "loss": 1.5684, "step": 38 }, { "epoch": 0.058383233532934134, "grad_norm": 0.326171875, "learning_rate": 0.0008, "loss": 1.5676, "step": 39 }, { "epoch": 0.059880239520958084, "grad_norm": 0.27734375, "learning_rate": 0.0008, "loss": 1.554, "step": 40 }, { "epoch": 0.061377245508982034, "grad_norm": 0.279296875, "learning_rate": 0.0008, "loss": 1.5503, "step": 41 }, { "epoch": 0.06287425149700598, "grad_norm": 0.26953125, "learning_rate": 0.0008, "loss": 1.5388, "step": 42 }, { "epoch": 0.06437125748502993, "grad_norm": 0.232421875, "learning_rate": 0.0008, "loss": 1.547, "step": 43 }, { "epoch": 0.0658682634730539, "grad_norm": 0.2578125, "learning_rate": 0.0008, "loss": 1.5204, "step": 44 }, { "epoch": 0.06736526946107785, "grad_norm": 0.279296875, "learning_rate": 0.0008, "loss": 1.511, "step": 45 }, { "epoch": 0.0688622754491018, "grad_norm": 0.423828125, "learning_rate": 0.0008, "loss": 1.5279, "step": 46 }, { "epoch": 0.07035928143712575, "grad_norm": 0.330078125, "learning_rate": 0.0008, "loss": 1.5328, "step": 47 }, { "epoch": 0.0718562874251497, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.5189, "step": 48 }, { "epoch": 0.07335329341317365, "grad_norm": 0.2158203125, "learning_rate": 0.0008, "loss": 1.5183, "step": 49 }, { "epoch": 0.0748502994011976, "grad_norm": 0.2470703125, "learning_rate": 0.0008, "loss": 1.5107, "step": 50 }, { "epoch": 0.07634730538922156, "grad_norm": 0.2216796875, "learning_rate": 0.0008, "loss": 1.507, "step": 51 }, { "epoch": 0.07784431137724551, "grad_norm": 0.1884765625, "learning_rate": 0.0008, "loss": 1.483, "step": 52 }, { "epoch": 0.07934131736526946, "grad_norm": 0.2470703125, "learning_rate": 0.0008, "loss": 1.4858, "step": 53 }, { "epoch": 0.08083832335329341, "grad_norm": 0.275390625, "learning_rate": 0.0008, "loss": 1.4907, "step": 54 }, { "epoch": 0.08233532934131736, "grad_norm": 0.32421875, "learning_rate": 0.0008, "loss": 1.4804, "step": 55 }, { "epoch": 0.08383233532934131, "grad_norm": 0.376953125, "learning_rate": 0.0008, "loss": 1.4833, "step": 56 }, { "epoch": 0.08532934131736528, "grad_norm": 0.462890625, "learning_rate": 0.0008, "loss": 1.4743, "step": 57 }, { "epoch": 0.08682634730538923, "grad_norm": 0.51171875, "learning_rate": 0.0008, "loss": 1.4808, "step": 58 }, { "epoch": 0.08832335329341318, "grad_norm": 0.578125, "learning_rate": 0.0008, "loss": 1.4926, "step": 59 }, { "epoch": 0.08982035928143713, "grad_norm": 0.65234375, "learning_rate": 0.0008, "loss": 1.4869, "step": 60 }, { "epoch": 0.09131736526946108, "grad_norm": 0.423828125, "learning_rate": 0.0008, "loss": 1.4629, "step": 61 }, { "epoch": 0.09281437125748503, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.4634, "step": 62 }, { "epoch": 0.09431137724550898, "grad_norm": 0.455078125, "learning_rate": 0.0008, "loss": 1.4597, "step": 63 }, { "epoch": 0.09580838323353294, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.4483, "step": 64 }, { "epoch": 0.09730538922155689, "grad_norm": 0.48046875, "learning_rate": 0.0008, "loss": 1.452, "step": 65 }, { "epoch": 0.09880239520958084, "grad_norm": 0.361328125, "learning_rate": 0.0008, "loss": 1.4557, "step": 66 }, { "epoch": 0.10029940119760479, "grad_norm": 0.212890625, "learning_rate": 0.0008, "loss": 1.4346, "step": 67 }, { "epoch": 0.10179640718562874, "grad_norm": 0.447265625, "learning_rate": 0.0008, "loss": 1.4549, "step": 68 }, { "epoch": 0.10329341317365269, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.4367, "step": 69 }, { "epoch": 0.10479041916167664, "grad_norm": 0.267578125, "learning_rate": 0.0008, "loss": 1.4298, "step": 70 }, { "epoch": 0.1062874251497006, "grad_norm": 0.453125, "learning_rate": 0.0008, "loss": 1.4434, "step": 71 }, { "epoch": 0.10778443113772455, "grad_norm": 0.23046875, "learning_rate": 0.0008, "loss": 1.4333, "step": 72 }, { "epoch": 0.1092814371257485, "grad_norm": 0.2353515625, "learning_rate": 0.0008, "loss": 1.4331, "step": 73 }, { "epoch": 0.11077844311377245, "grad_norm": 0.1923828125, "learning_rate": 0.0008, "loss": 1.4263, "step": 74 }, { "epoch": 0.1122754491017964, "grad_norm": 0.220703125, "learning_rate": 0.0008, "loss": 1.4307, "step": 75 }, { "epoch": 0.11377245508982035, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.4165, "step": 76 }, { "epoch": 0.11526946107784432, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.4138, "step": 77 }, { "epoch": 0.11676646706586827, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.4204, "step": 78 }, { "epoch": 0.11826347305389222, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.4122, "step": 79 }, { "epoch": 0.11976047904191617, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.4143, "step": 80 }, { "epoch": 0.12125748502994012, "grad_norm": 0.18359375, "learning_rate": 0.0008, "loss": 1.4032, "step": 81 }, { "epoch": 0.12275449101796407, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.4069, "step": 82 }, { "epoch": 0.12425149700598802, "grad_norm": 0.1923828125, "learning_rate": 0.0008, "loss": 1.4059, "step": 83 }, { "epoch": 0.12574850299401197, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.4065, "step": 84 }, { "epoch": 0.12724550898203593, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.3974, "step": 85 }, { "epoch": 0.12874251497005987, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.3896, "step": 86 }, { "epoch": 0.13023952095808383, "grad_norm": 0.23046875, "learning_rate": 0.0008, "loss": 1.4011, "step": 87 }, { "epoch": 0.1317365269461078, "grad_norm": 0.376953125, "learning_rate": 0.0008, "loss": 1.391, "step": 88 }, { "epoch": 0.13323353293413173, "grad_norm": 0.498046875, "learning_rate": 0.0008, "loss": 1.3908, "step": 89 }, { "epoch": 0.1347305389221557, "grad_norm": 1.3046875, "learning_rate": 0.0008, "loss": 1.4263, "step": 90 }, { "epoch": 0.13622754491017963, "grad_norm": 5.34375, "learning_rate": 0.0008, "loss": 1.5911, "step": 91 }, { "epoch": 0.1377245508982036, "grad_norm": 0.86328125, "learning_rate": 0.0008, "loss": 1.4479, "step": 92 }, { "epoch": 0.13922155688622753, "grad_norm": 0.90625, "learning_rate": 0.0008, "loss": 1.5098, "step": 93 }, { "epoch": 0.1407185628742515, "grad_norm": 0.51953125, "learning_rate": 0.0008, "loss": 1.4637, "step": 94 }, { "epoch": 0.14221556886227546, "grad_norm": 0.52734375, "learning_rate": 0.0008, "loss": 1.4494, "step": 95 }, { "epoch": 0.1437125748502994, "grad_norm": 0.71875, "learning_rate": 0.0008, "loss": 1.4602, "step": 96 }, { "epoch": 0.14520958083832336, "grad_norm": 1.7109375, "learning_rate": 0.0008, "loss": 1.4749, "step": 97 }, { "epoch": 0.1467065868263473, "grad_norm": 0.9765625, "learning_rate": 0.0008, "loss": 1.4621, "step": 98 }, { "epoch": 0.14820359281437126, "grad_norm": 0.466796875, "learning_rate": 0.0008, "loss": 1.4306, "step": 99 }, { "epoch": 0.1497005988023952, "grad_norm": 0.46484375, "learning_rate": 0.0008, "loss": 1.4282, "step": 100 }, { "epoch": 0.15119760479041916, "grad_norm": 0.41015625, "learning_rate": 0.0008, "loss": 1.4251, "step": 101 }, { "epoch": 0.15269461077844312, "grad_norm": 0.30078125, "learning_rate": 0.0008, "loss": 1.4319, "step": 102 }, { "epoch": 0.15419161676646706, "grad_norm": 0.2890625, "learning_rate": 0.0008, "loss": 1.4123, "step": 103 }, { "epoch": 0.15568862275449102, "grad_norm": 0.265625, "learning_rate": 0.0008, "loss": 1.4124, "step": 104 }, { "epoch": 0.15718562874251496, "grad_norm": 0.2890625, "learning_rate": 0.0008, "loss": 1.4222, "step": 105 }, { "epoch": 0.15868263473053892, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.4084, "step": 106 }, { "epoch": 0.1601796407185629, "grad_norm": 0.20703125, "learning_rate": 0.0008, "loss": 1.4113, "step": 107 }, { "epoch": 0.16167664670658682, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.403, "step": 108 }, { "epoch": 0.1631736526946108, "grad_norm": 0.185546875, "learning_rate": 0.0008, "loss": 1.3906, "step": 109 }, { "epoch": 0.16467065868263472, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.3873, "step": 110 }, { "epoch": 0.1661676646706587, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.3932, "step": 111 }, { "epoch": 0.16766467065868262, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.397, "step": 112 }, { "epoch": 0.1691616766467066, "grad_norm": 0.185546875, "learning_rate": 0.0008, "loss": 1.388, "step": 113 }, { "epoch": 0.17065868263473055, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.3821, "step": 114 }, { "epoch": 0.1721556886227545, "grad_norm": 0.212890625, "learning_rate": 0.0008, "loss": 1.3754, "step": 115 }, { "epoch": 0.17365269461077845, "grad_norm": 0.365234375, "learning_rate": 0.0008, "loss": 1.3932, "step": 116 }, { "epoch": 0.1751497005988024, "grad_norm": 0.82421875, "learning_rate": 0.0008, "loss": 1.3986, "step": 117 }, { "epoch": 0.17664670658682635, "grad_norm": 1.1953125, "learning_rate": 0.0008, "loss": 1.4088, "step": 118 }, { "epoch": 0.1781437125748503, "grad_norm": 1.2890625, "learning_rate": 0.0008, "loss": 1.4281, "step": 119 }, { "epoch": 0.17964071856287425, "grad_norm": 0.37890625, "learning_rate": 0.0008, "loss": 1.3933, "step": 120 }, { "epoch": 0.18113772455089822, "grad_norm": 1.015625, "learning_rate": 0.0008, "loss": 1.4178, "step": 121 }, { "epoch": 0.18263473053892215, "grad_norm": 0.578125, "learning_rate": 0.0008, "loss": 1.3884, "step": 122 }, { "epoch": 0.18413173652694612, "grad_norm": 0.4140625, "learning_rate": 0.0008, "loss": 1.3867, "step": 123 }, { "epoch": 0.18562874251497005, "grad_norm": 0.34765625, "learning_rate": 0.0008, "loss": 1.3795, "step": 124 }, { "epoch": 0.18712574850299402, "grad_norm": 0.33203125, "learning_rate": 0.0008, "loss": 1.3872, "step": 125 }, { "epoch": 0.18862275449101795, "grad_norm": 0.267578125, "learning_rate": 0.0008, "loss": 1.3764, "step": 126 }, { "epoch": 0.19011976047904192, "grad_norm": 0.2734375, "learning_rate": 0.0008, "loss": 1.3773, "step": 127 }, { "epoch": 0.19161676646706588, "grad_norm": 0.26171875, "learning_rate": 0.0008, "loss": 1.377, "step": 128 }, { "epoch": 0.19311377245508982, "grad_norm": 0.2001953125, "learning_rate": 0.0008, "loss": 1.3746, "step": 129 }, { "epoch": 0.19461077844311378, "grad_norm": 0.236328125, "learning_rate": 0.0008, "loss": 1.3691, "step": 130 }, { "epoch": 0.19610778443113772, "grad_norm": 0.2236328125, "learning_rate": 0.0008, "loss": 1.373, "step": 131 }, { "epoch": 0.19760479041916168, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.3726, "step": 132 }, { "epoch": 0.19910179640718562, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.3653, "step": 133 }, { "epoch": 0.20059880239520958, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.3572, "step": 134 }, { "epoch": 0.20209580838323354, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.3518, "step": 135 }, { "epoch": 0.20359281437125748, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.3573, "step": 136 }, { "epoch": 0.20508982035928144, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.3523, "step": 137 }, { "epoch": 0.20658682634730538, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.3393, "step": 138 }, { "epoch": 0.20808383233532934, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.3482, "step": 139 }, { "epoch": 0.20958083832335328, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.3397, "step": 140 }, { "epoch": 0.21107784431137724, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.3446, "step": 141 }, { "epoch": 0.2125748502994012, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.3432, "step": 142 }, { "epoch": 0.21407185628742514, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.3407, "step": 143 }, { "epoch": 0.2155688622754491, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.334, "step": 144 }, { "epoch": 0.21706586826347304, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.3358, "step": 145 }, { "epoch": 0.218562874251497, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.3459, "step": 146 }, { "epoch": 0.22005988023952097, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.3315, "step": 147 }, { "epoch": 0.2215568862275449, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.3374, "step": 148 }, { "epoch": 0.22305389221556887, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.3344, "step": 149 }, { "epoch": 0.2245508982035928, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.345, "step": 150 }, { "epoch": 0.22604790419161677, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.3301, "step": 151 }, { "epoch": 0.2275449101796407, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.331, "step": 152 }, { "epoch": 0.22904191616766467, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.322, "step": 153 }, { "epoch": 0.23053892215568864, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.3297, "step": 154 }, { "epoch": 0.23203592814371257, "grad_norm": 0.31640625, "learning_rate": 0.0008, "loss": 1.3389, "step": 155 }, { "epoch": 0.23353293413173654, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.3189, "step": 156 }, { "epoch": 0.23502994011976047, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.3325, "step": 157 }, { "epoch": 0.23652694610778444, "grad_norm": 0.28515625, "learning_rate": 0.0008, "loss": 1.3364, "step": 158 }, { "epoch": 0.23802395209580837, "grad_norm": 0.318359375, "learning_rate": 0.0008, "loss": 1.3401, "step": 159 }, { "epoch": 0.23952095808383234, "grad_norm": 0.3515625, "learning_rate": 0.0008, "loss": 1.3334, "step": 160 }, { "epoch": 0.2410179640718563, "grad_norm": 0.75390625, "learning_rate": 0.0008, "loss": 1.3468, "step": 161 }, { "epoch": 0.24251497005988024, "grad_norm": 1.921875, "learning_rate": 0.0008, "loss": 1.407, "step": 162 }, { "epoch": 0.2440119760479042, "grad_norm": 0.439453125, "learning_rate": 0.0008, "loss": 1.3459, "step": 163 }, { "epoch": 0.24550898203592814, "grad_norm": 1.03125, "learning_rate": 0.0008, "loss": 1.3732, "step": 164 }, { "epoch": 0.2470059880239521, "grad_norm": 1.1640625, "learning_rate": 0.0008, "loss": 1.3869, "step": 165 }, { "epoch": 0.24850299401197604, "grad_norm": 0.515625, "learning_rate": 0.0008, "loss": 1.3686, "step": 166 }, { "epoch": 0.25, "grad_norm": 0.365234375, "learning_rate": 0.0008, "loss": 1.3486, "step": 167 }, { "epoch": 0.25149700598802394, "grad_norm": 0.431640625, "learning_rate": 0.0008, "loss": 1.357, "step": 168 }, { "epoch": 0.25299401197604793, "grad_norm": 0.310546875, "learning_rate": 0.0008, "loss": 1.3466, "step": 169 }, { "epoch": 0.25449101796407186, "grad_norm": 0.271484375, "learning_rate": 0.0008, "loss": 1.3512, "step": 170 }, { "epoch": 0.2559880239520958, "grad_norm": 0.34765625, "learning_rate": 0.0008, "loss": 1.3607, "step": 171 }, { "epoch": 0.25748502994011974, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.3398, "step": 172 }, { "epoch": 0.25898203592814373, "grad_norm": 0.2080078125, "learning_rate": 0.0008, "loss": 1.3372, "step": 173 }, { "epoch": 0.26047904191616766, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.3342, "step": 174 }, { "epoch": 0.2619760479041916, "grad_norm": 0.203125, "learning_rate": 0.0008, "loss": 1.3536, "step": 175 }, { "epoch": 0.2634730538922156, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.3416, "step": 176 }, { "epoch": 0.26497005988023953, "grad_norm": 0.17578125, "learning_rate": 0.0008, "loss": 1.3383, "step": 177 }, { "epoch": 0.26646706586826346, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.3285, "step": 178 }, { "epoch": 0.2679640718562874, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.3252, "step": 179 }, { "epoch": 0.2694610778443114, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.3281, "step": 180 }, { "epoch": 0.27095808383233533, "grad_norm": 0.1943359375, "learning_rate": 0.0008, "loss": 1.3373, "step": 181 }, { "epoch": 0.27245508982035926, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.3247, "step": 182 }, { "epoch": 0.27395209580838326, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.3321, "step": 183 }, { "epoch": 0.2754491017964072, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.3243, "step": 184 }, { "epoch": 0.27694610778443113, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.323, "step": 185 }, { "epoch": 0.27844311377245506, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.3149, "step": 186 }, { "epoch": 0.27994011976047906, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.3184, "step": 187 }, { "epoch": 0.281437125748503, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.3214, "step": 188 }, { "epoch": 0.28293413173652693, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.3092, "step": 189 }, { "epoch": 0.2844311377245509, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.3186, "step": 190 }, { "epoch": 0.28592814371257486, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.3194, "step": 191 }, { "epoch": 0.2874251497005988, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.3081, "step": 192 }, { "epoch": 0.28892215568862273, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.3161, "step": 193 }, { "epoch": 0.2904191616766467, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.3115, "step": 194 }, { "epoch": 0.29191616766467066, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.3152, "step": 195 }, { "epoch": 0.2934131736526946, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.3126, "step": 196 }, { "epoch": 0.2949101796407186, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.3046, "step": 197 }, { "epoch": 0.2964071856287425, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.3053, "step": 198 }, { "epoch": 0.29790419161676646, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.3166, "step": 199 }, { "epoch": 0.2994011976047904, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.3115, "step": 200 }, { "epoch": 0.3008982035928144, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.3097, "step": 201 }, { "epoch": 0.3023952095808383, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.3051, "step": 202 }, { "epoch": 0.30389221556886226, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.3034, "step": 203 }, { "epoch": 0.30538922155688625, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.305, "step": 204 }, { "epoch": 0.3068862275449102, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2913, "step": 205 }, { "epoch": 0.3083832335329341, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.3031, "step": 206 }, { "epoch": 0.30988023952095806, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.3113, "step": 207 }, { "epoch": 0.31137724550898205, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.3012, "step": 208 }, { "epoch": 0.312874251497006, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.2968, "step": 209 }, { "epoch": 0.3143712574850299, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2994, "step": 210 }, { "epoch": 0.3158682634730539, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.3021, "step": 211 }, { "epoch": 0.31736526946107785, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.3061, "step": 212 }, { "epoch": 0.3188622754491018, "grad_norm": 0.44140625, "learning_rate": 0.0008, "loss": 1.3072, "step": 213 }, { "epoch": 0.3203592814371258, "grad_norm": 0.73046875, "learning_rate": 0.0008, "loss": 1.3489, "step": 214 }, { "epoch": 0.3218562874251497, "grad_norm": 0.47265625, "learning_rate": 0.0008, "loss": 1.3112, "step": 215 }, { "epoch": 0.32335329341317365, "grad_norm": 1.4765625, "learning_rate": 0.0008, "loss": 1.3582, "step": 216 }, { "epoch": 0.3248502994011976, "grad_norm": 1.2890625, "learning_rate": 0.0008, "loss": 1.3626, "step": 217 }, { "epoch": 0.3263473053892216, "grad_norm": 0.328125, "learning_rate": 0.0008, "loss": 1.3273, "step": 218 }, { "epoch": 0.3278443113772455, "grad_norm": 0.77734375, "learning_rate": 0.0008, "loss": 1.3443, "step": 219 }, { "epoch": 0.32934131736526945, "grad_norm": 0.69921875, "learning_rate": 0.0008, "loss": 1.3579, "step": 220 }, { "epoch": 0.33083832335329344, "grad_norm": 0.390625, "learning_rate": 0.0008, "loss": 1.3535, "step": 221 }, { "epoch": 0.3323353293413174, "grad_norm": 0.4140625, "learning_rate": 0.0008, "loss": 1.3344, "step": 222 }, { "epoch": 0.3338323353293413, "grad_norm": 0.341796875, "learning_rate": 0.0008, "loss": 1.3299, "step": 223 }, { "epoch": 0.33532934131736525, "grad_norm": 0.298828125, "learning_rate": 0.0008, "loss": 1.3354, "step": 224 }, { "epoch": 0.33682634730538924, "grad_norm": 0.2216796875, "learning_rate": 0.0008, "loss": 1.3313, "step": 225 }, { "epoch": 0.3383233532934132, "grad_norm": 0.28125, "learning_rate": 0.0008, "loss": 1.3231, "step": 226 }, { "epoch": 0.3398203592814371, "grad_norm": 0.220703125, "learning_rate": 0.0008, "loss": 1.3302, "step": 227 }, { "epoch": 0.3413173652694611, "grad_norm": 0.216796875, "learning_rate": 0.0008, "loss": 1.3175, "step": 228 }, { "epoch": 0.34281437125748504, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.3205, "step": 229 }, { "epoch": 0.344311377245509, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.3124, "step": 230 }, { "epoch": 0.3458083832335329, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.3172, "step": 231 }, { "epoch": 0.3473053892215569, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.3148, "step": 232 }, { "epoch": 0.34880239520958084, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.3087, "step": 233 }, { "epoch": 0.3502994011976048, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.3094, "step": 234 }, { "epoch": 0.35179640718562877, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.3113, "step": 235 }, { "epoch": 0.3532934131736527, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.3054, "step": 236 }, { "epoch": 0.35479041916167664, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.3145, "step": 237 }, { "epoch": 0.3562874251497006, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.304, "step": 238 }, { "epoch": 0.35778443113772457, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.3063, "step": 239 }, { "epoch": 0.3592814371257485, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.3018, "step": 240 }, { "epoch": 0.36077844311377244, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.3051, "step": 241 }, { "epoch": 0.36227544910179643, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2974, "step": 242 }, { "epoch": 0.36377245508982037, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.303, "step": 243 }, { "epoch": 0.3652694610778443, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.3004, "step": 244 }, { "epoch": 0.36676646706586824, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2908, "step": 245 }, { "epoch": 0.36826347305389223, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2944, "step": 246 }, { "epoch": 0.36976047904191617, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.3007, "step": 247 }, { "epoch": 0.3712574850299401, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2976, "step": 248 }, { "epoch": 0.3727544910179641, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.293, "step": 249 }, { "epoch": 0.37425149700598803, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.2994, "step": 250 }, { "epoch": 0.37574850299401197, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.2936, "step": 251 }, { "epoch": 0.3772455089820359, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.2927, "step": 252 }, { "epoch": 0.3787425149700599, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2853, "step": 253 }, { "epoch": 0.38023952095808383, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.285, "step": 254 }, { "epoch": 0.38173652694610777, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2995, "step": 255 }, { "epoch": 0.38323353293413176, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.2804, "step": 256 }, { "epoch": 0.3847305389221557, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2919, "step": 257 }, { "epoch": 0.38622754491017963, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.2924, "step": 258 }, { "epoch": 0.38772455089820357, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2885, "step": 259 }, { "epoch": 0.38922155688622756, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2904, "step": 260 }, { "epoch": 0.3907185628742515, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.2872, "step": 261 }, { "epoch": 0.39221556886227543, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2919, "step": 262 }, { "epoch": 0.3937125748502994, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.2957, "step": 263 }, { "epoch": 0.39520958083832336, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2855, "step": 264 }, { "epoch": 0.3967065868263473, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.2825, "step": 265 }, { "epoch": 0.39820359281437123, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2848, "step": 266 }, { "epoch": 0.3997005988023952, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.2872, "step": 267 }, { "epoch": 0.40119760479041916, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2917, "step": 268 }, { "epoch": 0.4026946107784431, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.2817, "step": 269 }, { "epoch": 0.4041916167664671, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.2888, "step": 270 }, { "epoch": 0.405688622754491, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.287, "step": 271 }, { "epoch": 0.40718562874251496, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2871, "step": 272 }, { "epoch": 0.4086826347305389, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2878, "step": 273 }, { "epoch": 0.4101796407185629, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2845, "step": 274 }, { "epoch": 0.4116766467065868, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.274, "step": 275 }, { "epoch": 0.41317365269461076, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.2769, "step": 276 }, { "epoch": 0.41467065868263475, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.284, "step": 277 }, { "epoch": 0.4161676646706587, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.2755, "step": 278 }, { "epoch": 0.4176646706586826, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.278, "step": 279 }, { "epoch": 0.41916167664670656, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.2831, "step": 280 }, { "epoch": 0.42065868263473055, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2768, "step": 281 }, { "epoch": 0.4221556886227545, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.2748, "step": 282 }, { "epoch": 0.4236526946107784, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.2761, "step": 283 }, { "epoch": 0.4251497005988024, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.2679, "step": 284 }, { "epoch": 0.42664670658682635, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.276, "step": 285 }, { "epoch": 0.4281437125748503, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.2812, "step": 286 }, { "epoch": 0.4296407185628742, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2695, "step": 287 }, { "epoch": 0.4311377245508982, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2785, "step": 288 }, { "epoch": 0.43263473053892215, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.2744, "step": 289 }, { "epoch": 0.4341317365269461, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2805, "step": 290 }, { "epoch": 0.4356287425149701, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.2771, "step": 291 }, { "epoch": 0.437125748502994, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.2746, "step": 292 }, { "epoch": 0.43862275449101795, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.284, "step": 293 }, { "epoch": 0.44011976047904194, "grad_norm": 0.1982421875, "learning_rate": 0.0008, "loss": 1.2814, "step": 294 }, { "epoch": 0.4416167664670659, "grad_norm": 0.1982421875, "learning_rate": 0.0008, "loss": 1.2742, "step": 295 }, { "epoch": 0.4431137724550898, "grad_norm": 0.302734375, "learning_rate": 0.0008, "loss": 1.2815, "step": 296 }, { "epoch": 0.44461077844311375, "grad_norm": 0.625, "learning_rate": 0.0008, "loss": 1.3037, "step": 297 }, { "epoch": 0.44610778443113774, "grad_norm": 0.97265625, "learning_rate": 0.0008, "loss": 1.3257, "step": 298 }, { "epoch": 0.4476047904191617, "grad_norm": 1.640625, "learning_rate": 0.0008, "loss": 1.3587, "step": 299 }, { "epoch": 0.4491017964071856, "grad_norm": 0.75390625, "learning_rate": 0.0008, "loss": 1.3257, "step": 300 }, { "epoch": 0.4505988023952096, "grad_norm": 0.455078125, "learning_rate": 0.0008, "loss": 1.3205, "step": 301 }, { "epoch": 0.45209580838323354, "grad_norm": 0.384765625, "learning_rate": 0.0008, "loss": 1.3186, "step": 302 }, { "epoch": 0.4535928143712575, "grad_norm": 0.4375, "learning_rate": 0.0008, "loss": 1.3042, "step": 303 }, { "epoch": 0.4550898203592814, "grad_norm": 0.248046875, "learning_rate": 0.0008, "loss": 1.299, "step": 304 }, { "epoch": 0.4565868263473054, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.2962, "step": 305 }, { "epoch": 0.45808383233532934, "grad_norm": 0.25, "learning_rate": 0.0008, "loss": 1.2882, "step": 306 }, { "epoch": 0.4595808383233533, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.2876, "step": 307 }, { "epoch": 0.46107784431137727, "grad_norm": 0.181640625, "learning_rate": 0.0008, "loss": 1.2961, "step": 308 }, { "epoch": 0.4625748502994012, "grad_norm": 0.2216796875, "learning_rate": 0.0008, "loss": 1.295, "step": 309 }, { "epoch": 0.46407185628742514, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.2868, "step": 310 }, { "epoch": 0.4655688622754491, "grad_norm": 0.17578125, "learning_rate": 0.0008, "loss": 1.2954, "step": 311 }, { "epoch": 0.46706586826347307, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.2825, "step": 312 }, { "epoch": 0.468562874251497, "grad_norm": 0.2021484375, "learning_rate": 0.0008, "loss": 1.2827, "step": 313 }, { "epoch": 0.47005988023952094, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.2897, "step": 314 }, { "epoch": 0.47155688622754494, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.2906, "step": 315 }, { "epoch": 0.47305389221556887, "grad_norm": 0.201171875, "learning_rate": 0.0008, "loss": 1.2799, "step": 316 }, { "epoch": 0.4745508982035928, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.2748, "step": 317 }, { "epoch": 0.47604790419161674, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.2831, "step": 318 }, { "epoch": 0.47754491017964074, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.2808, "step": 319 }, { "epoch": 0.47904191616766467, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.2816, "step": 320 }, { "epoch": 0.4805389221556886, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2835, "step": 321 }, { "epoch": 0.4820359281437126, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.2818, "step": 322 }, { "epoch": 0.48353293413173654, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2763, "step": 323 }, { "epoch": 0.48502994011976047, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2805, "step": 324 }, { "epoch": 0.4865269461077844, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.2805, "step": 325 }, { "epoch": 0.4880239520958084, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.276, "step": 326 }, { "epoch": 0.48952095808383234, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.2827, "step": 327 }, { "epoch": 0.49101796407185627, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2658, "step": 328 }, { "epoch": 0.49251497005988026, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2749, "step": 329 }, { "epoch": 0.4940119760479042, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2715, "step": 330 }, { "epoch": 0.49550898203592814, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.2661, "step": 331 }, { "epoch": 0.49700598802395207, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2714, "step": 332 }, { "epoch": 0.49850299401197606, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.2666, "step": 333 }, { "epoch": 0.5, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2837, "step": 334 }, { "epoch": 0.5014970059880239, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2692, "step": 335 }, { "epoch": 0.5029940119760479, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.268, "step": 336 }, { "epoch": 0.5044910179640718, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2758, "step": 337 }, { "epoch": 0.5059880239520959, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.2652, "step": 338 }, { "epoch": 0.5074850299401198, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2729, "step": 339 }, { "epoch": 0.5089820359281437, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2636, "step": 340 }, { "epoch": 0.5104790419161677, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2692, "step": 341 }, { "epoch": 0.5119760479041916, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.2694, "step": 342 }, { "epoch": 0.5134730538922155, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2742, "step": 343 }, { "epoch": 0.5149700598802395, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.267, "step": 344 }, { "epoch": 0.5164670658682635, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2677, "step": 345 }, { "epoch": 0.5179640718562875, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.2671, "step": 346 }, { "epoch": 0.5194610778443114, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2737, "step": 347 }, { "epoch": 0.5209580838323353, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.2797, "step": 348 }, { "epoch": 0.5224550898203593, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.2638, "step": 349 }, { "epoch": 0.5239520958083832, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2663, "step": 350 }, { "epoch": 0.5254491017964071, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.2715, "step": 351 }, { "epoch": 0.5269461077844312, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2662, "step": 352 }, { "epoch": 0.5284431137724551, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.2748, "step": 353 }, { "epoch": 0.5299401197604791, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2716, "step": 354 }, { "epoch": 0.531437125748503, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.2581, "step": 355 }, { "epoch": 0.5329341317365269, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2679, "step": 356 }, { "epoch": 0.5344311377245509, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.2615, "step": 357 }, { "epoch": 0.5359281437125748, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.2642, "step": 358 }, { "epoch": 0.5374251497005988, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.267, "step": 359 }, { "epoch": 0.5389221556886228, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.2599, "step": 360 }, { "epoch": 0.5404191616766467, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.2684, "step": 361 }, { "epoch": 0.5419161676646707, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2617, "step": 362 }, { "epoch": 0.5434131736526946, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.2601, "step": 363 }, { "epoch": 0.5449101796407185, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2675, "step": 364 }, { "epoch": 0.5464071856287425, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.269, "step": 365 }, { "epoch": 0.5479041916167665, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.264, "step": 366 }, { "epoch": 0.5494011976047904, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2626, "step": 367 }, { "epoch": 0.5508982035928144, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.2651, "step": 368 }, { "epoch": 0.5523952095808383, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2659, "step": 369 }, { "epoch": 0.5538922155688623, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.2623, "step": 370 }, { "epoch": 0.5553892215568862, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.2614, "step": 371 }, { "epoch": 0.5568862275449101, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2689, "step": 372 }, { "epoch": 0.5583832335329342, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2632, "step": 373 }, { "epoch": 0.5598802395209581, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.2591, "step": 374 }, { "epoch": 0.561377245508982, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.2601, "step": 375 }, { "epoch": 0.562874251497006, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2628, "step": 376 }, { "epoch": 0.5643712574850299, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.2577, "step": 377 }, { "epoch": 0.5658682634730539, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2571, "step": 378 }, { "epoch": 0.5673652694610778, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.2581, "step": 379 }, { "epoch": 0.5688622754491018, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.2596, "step": 380 }, { "epoch": 0.5703592814371258, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.2629, "step": 381 }, { "epoch": 0.5718562874251497, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2597, "step": 382 }, { "epoch": 0.5733532934131736, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2626, "step": 383 }, { "epoch": 0.5748502994011976, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2557, "step": 384 }, { "epoch": 0.5763473053892215, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.2703, "step": 385 }, { "epoch": 0.5778443113772455, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.2595, "step": 386 }, { "epoch": 0.5793413173652695, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2534, "step": 387 }, { "epoch": 0.5808383233532934, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.2662, "step": 388 }, { "epoch": 0.5823353293413174, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.2494, "step": 389 }, { "epoch": 0.5838323353293413, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.2616, "step": 390 }, { "epoch": 0.5853293413173652, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.2516, "step": 391 }, { "epoch": 0.5868263473053892, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.2502, "step": 392 }, { "epoch": 0.5883233532934131, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2538, "step": 393 }, { "epoch": 0.5898203592814372, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2571, "step": 394 }, { "epoch": 0.5913173652694611, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.2501, "step": 395 }, { "epoch": 0.592814371257485, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.2572, "step": 396 }, { "epoch": 0.594311377245509, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.2573, "step": 397 }, { "epoch": 0.5958083832335329, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2487, "step": 398 }, { "epoch": 0.5973053892215568, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.2464, "step": 399 }, { "epoch": 0.5988023952095808, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.251, "step": 400 }, { "epoch": 0.6002994011976048, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2577, "step": 401 }, { "epoch": 0.6017964071856288, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2532, "step": 402 }, { "epoch": 0.6032934131736527, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.2586, "step": 403 }, { "epoch": 0.6047904191616766, "grad_norm": 0.216796875, "learning_rate": 0.0008, "loss": 1.2574, "step": 404 }, { "epoch": 0.6062874251497006, "grad_norm": 0.365234375, "learning_rate": 0.0008, "loss": 1.2516, "step": 405 }, { "epoch": 0.6077844311377245, "grad_norm": 1.171875, "learning_rate": 0.0008, "loss": 1.2707, "step": 406 }, { "epoch": 0.6092814371257484, "grad_norm": 1.890625, "learning_rate": 0.0008, "loss": 1.3835, "step": 407 }, { "epoch": 0.6107784431137725, "grad_norm": 0.81640625, "learning_rate": 0.0008, "loss": 1.3279, "step": 408 }, { "epoch": 0.6122754491017964, "grad_norm": 2.703125, "learning_rate": 0.0008, "loss": 1.4153, "step": 409 }, { "epoch": 0.6137724550898204, "grad_norm": 0.76953125, "learning_rate": 0.0008, "loss": 1.3757, "step": 410 }, { "epoch": 0.6152694610778443, "grad_norm": 1.2265625, "learning_rate": 0.0008, "loss": 1.4161, "step": 411 }, { "epoch": 0.6167664670658682, "grad_norm": 0.96875, "learning_rate": 0.0008, "loss": 1.3939, "step": 412 }, { "epoch": 0.6182634730538922, "grad_norm": 0.64453125, "learning_rate": 0.0008, "loss": 1.3756, "step": 413 }, { "epoch": 0.6197604790419161, "grad_norm": 0.515625, "learning_rate": 0.0008, "loss": 1.373, "step": 414 }, { "epoch": 0.6212574850299402, "grad_norm": 0.423828125, "learning_rate": 0.0008, "loss": 1.3487, "step": 415 }, { "epoch": 0.6227544910179641, "grad_norm": 0.373046875, "learning_rate": 0.0008, "loss": 1.3556, "step": 416 }, { "epoch": 0.624251497005988, "grad_norm": 0.3203125, "learning_rate": 0.0008, "loss": 1.3364, "step": 417 }, { "epoch": 0.625748502994012, "grad_norm": 0.3125, "learning_rate": 0.0008, "loss": 1.3407, "step": 418 }, { "epoch": 0.6272455089820359, "grad_norm": 0.26171875, "learning_rate": 0.0008, "loss": 1.3387, "step": 419 }, { "epoch": 0.6287425149700598, "grad_norm": 0.2216796875, "learning_rate": 0.0008, "loss": 1.3285, "step": 420 }, { "epoch": 0.6302395209580839, "grad_norm": 0.2451171875, "learning_rate": 0.0008, "loss": 1.3225, "step": 421 }, { "epoch": 0.6317365269461078, "grad_norm": 0.2099609375, "learning_rate": 0.0008, "loss": 1.3073, "step": 422 }, { "epoch": 0.6332335329341318, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.3177, "step": 423 }, { "epoch": 0.6347305389221557, "grad_norm": 0.2216796875, "learning_rate": 0.0008, "loss": 1.3074, "step": 424 }, { "epoch": 0.6362275449101796, "grad_norm": 0.21875, "learning_rate": 0.0008, "loss": 1.2996, "step": 425 }, { "epoch": 0.6377245508982036, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.3035, "step": 426 }, { "epoch": 0.6392215568862275, "grad_norm": 0.2197265625, "learning_rate": 0.0008, "loss": 1.3063, "step": 427 }, { "epoch": 0.6407185628742516, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.3045, "step": 428 }, { "epoch": 0.6422155688622755, "grad_norm": 0.1572265625, "learning_rate": 0.0008, "loss": 1.2989, "step": 429 }, { "epoch": 0.6437125748502994, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.293, "step": 430 }, { "epoch": 0.6452095808383234, "grad_norm": 0.216796875, "learning_rate": 0.0008, "loss": 1.2885, "step": 431 }, { "epoch": 0.6467065868263473, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.2915, "step": 432 }, { "epoch": 0.6482035928143712, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.2892, "step": 433 }, { "epoch": 0.6497005988023952, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.2859, "step": 434 }, { "epoch": 0.6511976047904192, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.2829, "step": 435 }, { "epoch": 0.6526946107784432, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.2888, "step": 436 }, { "epoch": 0.6541916167664671, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.278, "step": 437 }, { "epoch": 0.655688622754491, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.2785, "step": 438 }, { "epoch": 0.657185628742515, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2866, "step": 439 }, { "epoch": 0.6586826347305389, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.2802, "step": 440 }, { "epoch": 0.6601796407185628, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.2791, "step": 441 }, { "epoch": 0.6616766467065869, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2747, "step": 442 }, { "epoch": 0.6631736526946108, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.2801, "step": 443 }, { "epoch": 0.6646706586826348, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.2771, "step": 444 }, { "epoch": 0.6661676646706587, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2773, "step": 445 }, { "epoch": 0.6676646706586826, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2719, "step": 446 }, { "epoch": 0.6691616766467066, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.2724, "step": 447 }, { "epoch": 0.6706586826347305, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2643, "step": 448 }, { "epoch": 0.6721556886227545, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.2621, "step": 449 }, { "epoch": 0.6736526946107785, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.272, "step": 450 }, { "epoch": 0.6751497005988024, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2579, "step": 451 }, { "epoch": 0.6766467065868264, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.2697, "step": 452 }, { "epoch": 0.6781437125748503, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.2567, "step": 453 }, { "epoch": 0.6796407185628742, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2649, "step": 454 }, { "epoch": 0.6811377245508982, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2707, "step": 455 }, { "epoch": 0.6826347305389222, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2685, "step": 456 }, { "epoch": 0.6841317365269461, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2598, "step": 457 }, { "epoch": 0.6856287425149701, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2675, "step": 458 }, { "epoch": 0.687125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2612, "step": 459 }, { "epoch": 0.688622754491018, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2551, "step": 460 }, { "epoch": 0.6901197604790419, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2577, "step": 461 }, { "epoch": 0.6916167664670658, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2606, "step": 462 }, { "epoch": 0.6931137724550899, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2602, "step": 463 }, { "epoch": 0.6946107784431138, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.2544, "step": 464 }, { "epoch": 0.6961077844311377, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.2547, "step": 465 }, { "epoch": 0.6976047904191617, "grad_norm": 0.337890625, "learning_rate": 0.0008, "loss": 1.2575, "step": 466 }, { "epoch": 0.6991017964071856, "grad_norm": 0.71484375, "learning_rate": 0.0008, "loss": 1.2726, "step": 467 }, { "epoch": 0.7005988023952096, "grad_norm": 1.34375, "learning_rate": 0.0008, "loss": 1.3157, "step": 468 }, { "epoch": 0.7020958083832335, "grad_norm": 0.609375, "learning_rate": 0.0008, "loss": 1.2837, "step": 469 }, { "epoch": 0.7035928143712575, "grad_norm": 0.640625, "learning_rate": 0.0008, "loss": 1.3059, "step": 470 }, { "epoch": 0.7050898203592815, "grad_norm": 0.52734375, "learning_rate": 0.0008, "loss": 1.3061, "step": 471 }, { "epoch": 0.7065868263473054, "grad_norm": 0.4921875, "learning_rate": 0.0008, "loss": 1.2931, "step": 472 }, { "epoch": 0.7080838323353293, "grad_norm": 0.48046875, "learning_rate": 0.0008, "loss": 1.2869, "step": 473 }, { "epoch": 0.7095808383233533, "grad_norm": 0.353515625, "learning_rate": 0.0008, "loss": 1.2761, "step": 474 }, { "epoch": 0.7110778443113772, "grad_norm": 0.333984375, "learning_rate": 0.0008, "loss": 1.2764, "step": 475 }, { "epoch": 0.7125748502994012, "grad_norm": 0.294921875, "learning_rate": 0.0008, "loss": 1.2857, "step": 476 }, { "epoch": 0.7140718562874252, "grad_norm": 0.2431640625, "learning_rate": 0.0008, "loss": 1.2777, "step": 477 }, { "epoch": 0.7155688622754491, "grad_norm": 0.1982421875, "learning_rate": 0.0008, "loss": 1.2747, "step": 478 }, { "epoch": 0.7170658682634731, "grad_norm": 0.2255859375, "learning_rate": 0.0008, "loss": 1.2732, "step": 479 }, { "epoch": 0.718562874251497, "grad_norm": 0.1826171875, "learning_rate": 0.0008, "loss": 1.2695, "step": 480 }, { "epoch": 0.7200598802395209, "grad_norm": 0.193359375, "learning_rate": 0.0008, "loss": 1.2706, "step": 481 }, { "epoch": 0.7215568862275449, "grad_norm": 0.1826171875, "learning_rate": 0.0008, "loss": 1.2783, "step": 482 }, { "epoch": 0.7230538922155688, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.2682, "step": 483 }, { "epoch": 0.7245508982035929, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.2582, "step": 484 }, { "epoch": 0.7260479041916168, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.275, "step": 485 }, { "epoch": 0.7275449101796407, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.2671, "step": 486 }, { "epoch": 0.7290419161676647, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.2655, "step": 487 }, { "epoch": 0.7305389221556886, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.2631, "step": 488 }, { "epoch": 0.7320359281437125, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.2566, "step": 489 }, { "epoch": 0.7335329341317365, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2566, "step": 490 }, { "epoch": 0.7350299401197605, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.2554, "step": 491 }, { "epoch": 0.7365269461077845, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.2556, "step": 492 }, { "epoch": 0.7380239520958084, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.2574, "step": 493 }, { "epoch": 0.7395209580838323, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.257, "step": 494 }, { "epoch": 0.7410179640718563, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2622, "step": 495 }, { "epoch": 0.7425149700598802, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2635, "step": 496 }, { "epoch": 0.7440119760479041, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2536, "step": 497 }, { "epoch": 0.7455089820359282, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.2551, "step": 498 }, { "epoch": 0.7470059880239521, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2509, "step": 499 }, { "epoch": 0.7485029940119761, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2617, "step": 500 }, { "epoch": 0.75, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2542, "step": 501 }, { "epoch": 0.7514970059880239, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.254, "step": 502 }, { "epoch": 0.7529940119760479, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2516, "step": 503 }, { "epoch": 0.7544910179640718, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.2534, "step": 504 }, { "epoch": 0.7559880239520959, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2536, "step": 505 }, { "epoch": 0.7574850299401198, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2457, "step": 506 }, { "epoch": 0.7589820359281437, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.2576, "step": 507 }, { "epoch": 0.7604790419161677, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2521, "step": 508 }, { "epoch": 0.7619760479041916, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2629, "step": 509 }, { "epoch": 0.7634730538922155, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2444, "step": 510 }, { "epoch": 0.7649700598802395, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2592, "step": 511 }, { "epoch": 0.7664670658682635, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2509, "step": 512 }, { "epoch": 0.7679640718562875, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2494, "step": 513 }, { "epoch": 0.7694610778443114, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2442, "step": 514 }, { "epoch": 0.7709580838323353, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2492, "step": 515 }, { "epoch": 0.7724550898203593, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.2487, "step": 516 }, { "epoch": 0.7739520958083832, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.2504, "step": 517 }, { "epoch": 0.7754491017964071, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.2428, "step": 518 }, { "epoch": 0.7769461077844312, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.2555, "step": 519 }, { "epoch": 0.7784431137724551, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2431, "step": 520 }, { "epoch": 0.7799401197604791, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2415, "step": 521 }, { "epoch": 0.781437125748503, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.2507, "step": 522 }, { "epoch": 0.7829341317365269, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.2531, "step": 523 }, { "epoch": 0.7844311377245509, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2447, "step": 524 }, { "epoch": 0.7859281437125748, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.2391, "step": 525 }, { "epoch": 0.7874251497005988, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2424, "step": 526 }, { "epoch": 0.7889221556886228, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2469, "step": 527 }, { "epoch": 0.7904191616766467, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2487, "step": 528 }, { "epoch": 0.7919161676646707, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.2405, "step": 529 }, { "epoch": 0.7934131736526946, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.2493, "step": 530 }, { "epoch": 0.7949101796407185, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.2407, "step": 531 }, { "epoch": 0.7964071856287425, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.2494, "step": 532 }, { "epoch": 0.7979041916167665, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2468, "step": 533 }, { "epoch": 0.7994011976047904, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.2456, "step": 534 }, { "epoch": 0.8008982035928144, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2483, "step": 535 }, { "epoch": 0.8023952095808383, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.2478, "step": 536 }, { "epoch": 0.8038922155688623, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2463, "step": 537 }, { "epoch": 0.8053892215568862, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2434, "step": 538 }, { "epoch": 0.8068862275449101, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2472, "step": 539 }, { "epoch": 0.8083832335329342, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2514, "step": 540 }, { "epoch": 0.8098802395209581, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2483, "step": 541 }, { "epoch": 0.811377245508982, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.2396, "step": 542 }, { "epoch": 0.812874251497006, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2467, "step": 543 }, { "epoch": 0.8143712574850299, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.2467, "step": 544 }, { "epoch": 0.8158682634730539, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.2465, "step": 545 }, { "epoch": 0.8173652694610778, "grad_norm": 0.2099609375, "learning_rate": 0.0008, "loss": 1.2495, "step": 546 }, { "epoch": 0.8188622754491018, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.2516, "step": 547 }, { "epoch": 0.8203592814371258, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.2468, "step": 548 }, { "epoch": 0.8218562874251497, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.2523, "step": 549 }, { "epoch": 0.8233532934131736, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.2377, "step": 550 }, { "epoch": 0.8248502994011976, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.2487, "step": 551 }, { "epoch": 0.8263473053892215, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.2451, "step": 552 }, { "epoch": 0.8278443113772455, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.2477, "step": 553 }, { "epoch": 0.8293413173652695, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.2495, "step": 554 }, { "epoch": 0.8308383233532934, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.2438, "step": 555 }, { "epoch": 0.8323353293413174, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2469, "step": 556 }, { "epoch": 0.8338323353293413, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.2444, "step": 557 }, { "epoch": 0.8353293413173652, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.2592, "step": 558 }, { "epoch": 0.8368263473053892, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.2508, "step": 559 }, { "epoch": 0.8383233532934131, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.2474, "step": 560 }, { "epoch": 0.8398203592814372, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.2452, "step": 561 }, { "epoch": 0.8413173652694611, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.2421, "step": 562 }, { "epoch": 0.842814371257485, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.2452, "step": 563 }, { "epoch": 0.844311377245509, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.2469, "step": 564 }, { "epoch": 0.8458083832335329, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2491, "step": 565 }, { "epoch": 0.8473053892215568, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.2406, "step": 566 }, { "epoch": 0.8488023952095808, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2382, "step": 567 }, { "epoch": 0.8502994011976048, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2385, "step": 568 }, { "epoch": 0.8517964071856288, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.2477, "step": 569 }, { "epoch": 0.8532934131736527, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2428, "step": 570 }, { "epoch": 0.8547904191616766, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2465, "step": 571 }, { "epoch": 0.8562874251497006, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.2548, "step": 572 }, { "epoch": 0.8577844311377245, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2468, "step": 573 }, { "epoch": 0.8592814371257484, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2416, "step": 574 }, { "epoch": 0.8607784431137725, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2437, "step": 575 }, { "epoch": 0.8622754491017964, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2291, "step": 576 }, { "epoch": 0.8637724550898204, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.2391, "step": 577 }, { "epoch": 0.8652694610778443, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2451, "step": 578 }, { "epoch": 0.8667664670658682, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2325, "step": 579 }, { "epoch": 0.8682634730538922, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2394, "step": 580 }, { "epoch": 0.8697604790419161, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2332, "step": 581 }, { "epoch": 0.8712574850299402, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.239, "step": 582 }, { "epoch": 0.8727544910179641, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.2363, "step": 583 }, { "epoch": 0.874251497005988, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2463, "step": 584 }, { "epoch": 0.875748502994012, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.2459, "step": 585 }, { "epoch": 0.8772455089820359, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.2403, "step": 586 }, { "epoch": 0.8787425149700598, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.249, "step": 587 }, { "epoch": 0.8802395209580839, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2359, "step": 588 }, { "epoch": 0.8817365269461078, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.2448, "step": 589 }, { "epoch": 0.8832335329341318, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.2473, "step": 590 }, { "epoch": 0.8847305389221557, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.2437, "step": 591 }, { "epoch": 0.8862275449101796, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.243, "step": 592 }, { "epoch": 0.8877245508982036, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.2377, "step": 593 }, { "epoch": 0.8892215568862275, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2365, "step": 594 }, { "epoch": 0.8907185628742516, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.2386, "step": 595 }, { "epoch": 0.8922155688622755, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.2347, "step": 596 }, { "epoch": 0.8937125748502994, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.2459, "step": 597 }, { "epoch": 0.8952095808383234, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.2417, "step": 598 }, { "epoch": 0.8967065868263473, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.2378, "step": 599 }, { "epoch": 0.8982035928143712, "grad_norm": 0.1572265625, "learning_rate": 0.0008, "loss": 1.2402, "step": 600 }, { "epoch": 0.8997005988023952, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.2324, "step": 601 }, { "epoch": 0.9011976047904192, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.2395, "step": 602 }, { "epoch": 0.9026946107784432, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.2374, "step": 603 }, { "epoch": 0.9041916167664671, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.2418, "step": 604 }, { "epoch": 0.905688622754491, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.2431, "step": 605 }, { "epoch": 0.907185628742515, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.2416, "step": 606 }, { "epoch": 0.9086826347305389, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.2369, "step": 607 }, { "epoch": 0.9101796407185628, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.2399, "step": 608 }, { "epoch": 0.9116766467065869, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2445, "step": 609 }, { "epoch": 0.9131736526946108, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.2343, "step": 610 }, { "epoch": 0.9146706586826348, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.2317, "step": 611 }, { "epoch": 0.9161676646706587, "grad_norm": 0.3203125, "learning_rate": 0.0008, "loss": 1.2381, "step": 612 }, { "epoch": 0.9176646706586826, "grad_norm": 0.7421875, "learning_rate": 0.0008, "loss": 1.2429, "step": 613 }, { "epoch": 0.9191616766467066, "grad_norm": 1.625, "learning_rate": 0.0008, "loss": 1.3109, "step": 614 }, { "epoch": 0.9206586826347305, "grad_norm": 0.8125, "learning_rate": 0.0008, "loss": 1.2647, "step": 615 }, { "epoch": 0.9221556886227545, "grad_norm": 0.71484375, "learning_rate": 0.0008, "loss": 1.2848, "step": 616 }, { "epoch": 0.9236526946107785, "grad_norm": 0.484375, "learning_rate": 0.0008, "loss": 1.2868, "step": 617 }, { "epoch": 0.9251497005988024, "grad_norm": 0.353515625, "learning_rate": 0.0008, "loss": 1.2641, "step": 618 }, { "epoch": 0.9266467065868264, "grad_norm": 0.298828125, "learning_rate": 0.0008, "loss": 1.261, "step": 619 }, { "epoch": 0.9281437125748503, "grad_norm": 0.349609375, "learning_rate": 0.0008, "loss": 1.2685, "step": 620 }, { "epoch": 0.9296407185628742, "grad_norm": 0.2431640625, "learning_rate": 0.0008, "loss": 1.2581, "step": 621 }, { "epoch": 0.9311377245508982, "grad_norm": 0.267578125, "learning_rate": 0.0008, "loss": 1.2602, "step": 622 }, { "epoch": 0.9326347305389222, "grad_norm": 0.2333984375, "learning_rate": 0.0008, "loss": 1.2504, "step": 623 }, { "epoch": 0.9341317365269461, "grad_norm": 0.185546875, "learning_rate": 0.0008, "loss": 1.2536, "step": 624 }, { "epoch": 0.9356287425149701, "grad_norm": 0.21875, "learning_rate": 0.0008, "loss": 1.2571, "step": 625 }, { "epoch": 0.937125748502994, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2459, "step": 626 }, { "epoch": 0.938622754491018, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.2506, "step": 627 }, { "epoch": 0.9401197604790419, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.2524, "step": 628 }, { "epoch": 0.9416167664670658, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.2505, "step": 629 }, { "epoch": 0.9431137724550899, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.2487, "step": 630 }, { "epoch": 0.9446107784431138, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.2488, "step": 631 }, { "epoch": 0.9461077844311377, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.2459, "step": 632 }, { "epoch": 0.9476047904191617, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.2413, "step": 633 }, { "epoch": 0.9491017964071856, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2407, "step": 634 }, { "epoch": 0.9505988023952096, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2363, "step": 635 }, { "epoch": 0.9520958083832335, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.2352, "step": 636 }, { "epoch": 0.9535928143712575, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2365, "step": 637 }, { "epoch": 0.9550898203592815, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.2447, "step": 638 }, { "epoch": 0.9565868263473054, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.245, "step": 639 }, { "epoch": 0.9580838323353293, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2441, "step": 640 }, { "epoch": 0.9595808383233533, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2406, "step": 641 }, { "epoch": 0.9610778443113772, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2385, "step": 642 }, { "epoch": 0.9625748502994012, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.2329, "step": 643 }, { "epoch": 0.9640718562874252, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.2412, "step": 644 }, { "epoch": 0.9655688622754491, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2366, "step": 645 }, { "epoch": 0.9670658682634731, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.2435, "step": 646 }, { "epoch": 0.968562874251497, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2438, "step": 647 }, { "epoch": 0.9700598802395209, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2343, "step": 648 }, { "epoch": 0.9715568862275449, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.2347, "step": 649 }, { "epoch": 0.9730538922155688, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2401, "step": 650 }, { "epoch": 0.9745508982035929, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.235, "step": 651 }, { "epoch": 0.9760479041916168, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2344, "step": 652 }, { "epoch": 0.9775449101796407, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2315, "step": 653 }, { "epoch": 0.9790419161676647, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2327, "step": 654 }, { "epoch": 0.9805389221556886, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.236, "step": 655 }, { "epoch": 0.9820359281437125, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.2296, "step": 656 }, { "epoch": 0.9835329341317365, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.2338, "step": 657 }, { "epoch": 0.9850299401197605, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2435, "step": 658 }, { "epoch": 0.9865269461077845, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.2294, "step": 659 }, { "epoch": 0.9880239520958084, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.232, "step": 660 }, { "epoch": 0.9895209580838323, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.2238, "step": 661 }, { "epoch": 0.9910179640718563, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2328, "step": 662 }, { "epoch": 0.9925149700598802, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2334, "step": 663 }, { "epoch": 0.9940119760479041, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2371, "step": 664 }, { "epoch": 0.9955089820359282, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2422, "step": 665 }, { "epoch": 0.9970059880239521, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2314, "step": 666 }, { "epoch": 0.9985029940119761, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2306, "step": 667 }, { "epoch": 1.0, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2275, "step": 668 }, { "epoch": 1.001497005988024, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.2372, "step": 669 }, { "epoch": 1.0029940119760479, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2366, "step": 670 }, { "epoch": 1.0044910179640718, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2293, "step": 671 }, { "epoch": 1.0059880239520957, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.2365, "step": 672 }, { "epoch": 1.0074850299401197, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2288, "step": 673 }, { "epoch": 1.0089820359281436, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2309, "step": 674 }, { "epoch": 1.0104790419161676, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2281, "step": 675 }, { "epoch": 1.0119760479041917, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2286, "step": 676 }, { "epoch": 1.0134730538922156, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2302, "step": 677 }, { "epoch": 1.0149700598802396, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2238, "step": 678 }, { "epoch": 1.0164670658682635, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.2237, "step": 679 }, { "epoch": 1.0179640718562875, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.222, "step": 680 }, { "epoch": 1.0194610778443114, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2323, "step": 681 }, { "epoch": 1.0209580838323353, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2246, "step": 682 }, { "epoch": 1.0224550898203593, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2249, "step": 683 }, { "epoch": 1.0239520958083832, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2348, "step": 684 }, { "epoch": 1.0254491017964071, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2309, "step": 685 }, { "epoch": 1.026946107784431, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.2295, "step": 686 }, { "epoch": 1.028443113772455, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2291, "step": 687 }, { "epoch": 1.029940119760479, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.2245, "step": 688 }, { "epoch": 1.031437125748503, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2247, "step": 689 }, { "epoch": 1.032934131736527, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2322, "step": 690 }, { "epoch": 1.034431137724551, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.2273, "step": 691 }, { "epoch": 1.035928143712575, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.223, "step": 692 }, { "epoch": 1.0374251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2276, "step": 693 }, { "epoch": 1.0389221556886228, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.221, "step": 694 }, { "epoch": 1.0404191616766467, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2212, "step": 695 }, { "epoch": 1.0419161676646707, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.222, "step": 696 }, { "epoch": 1.0434131736526946, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2294, "step": 697 }, { "epoch": 1.0449101796407185, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2231, "step": 698 }, { "epoch": 1.0464071856287425, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.2274, "step": 699 }, { "epoch": 1.0479041916167664, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2232, "step": 700 }, { "epoch": 1.0494011976047903, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2189, "step": 701 }, { "epoch": 1.0508982035928143, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2336, "step": 702 }, { "epoch": 1.0523952095808382, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.23, "step": 703 }, { "epoch": 1.0538922155688624, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2317, "step": 704 }, { "epoch": 1.0553892215568863, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2261, "step": 705 }, { "epoch": 1.0568862275449102, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.2241, "step": 706 }, { "epoch": 1.0583832335329342, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.2223, "step": 707 }, { "epoch": 1.0598802395209581, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2178, "step": 708 }, { "epoch": 1.061377245508982, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.2348, "step": 709 }, { "epoch": 1.062874251497006, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.2284, "step": 710 }, { "epoch": 1.06437125748503, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.2272, "step": 711 }, { "epoch": 1.0658682634730539, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.2345, "step": 712 }, { "epoch": 1.0673652694610778, "grad_norm": 0.1953125, "learning_rate": 0.0008, "loss": 1.2282, "step": 713 }, { "epoch": 1.0688622754491017, "grad_norm": 0.1748046875, "learning_rate": 0.0008, "loss": 1.2278, "step": 714 }, { "epoch": 1.0703592814371257, "grad_norm": 0.1953125, "learning_rate": 0.0008, "loss": 1.2277, "step": 715 }, { "epoch": 1.0718562874251496, "grad_norm": 0.189453125, "learning_rate": 0.0008, "loss": 1.2264, "step": 716 }, { "epoch": 1.0733532934131738, "grad_norm": 0.267578125, "learning_rate": 0.0008, "loss": 1.2238, "step": 717 }, { "epoch": 1.0748502994011977, "grad_norm": 0.3671875, "learning_rate": 0.0008, "loss": 1.218, "step": 718 }, { "epoch": 1.0763473053892216, "grad_norm": 0.828125, "learning_rate": 0.0008, "loss": 1.2397, "step": 719 }, { "epoch": 1.0778443113772456, "grad_norm": 1.8203125, "learning_rate": 0.0008, "loss": 1.3052, "step": 720 }, { "epoch": 1.0793413173652695, "grad_norm": 0.2470703125, "learning_rate": 0.0008, "loss": 1.2408, "step": 721 }, { "epoch": 1.0808383233532934, "grad_norm": 1.1875, "learning_rate": 0.0008, "loss": 1.2891, "step": 722 }, { "epoch": 1.0823353293413174, "grad_norm": 0.53125, "learning_rate": 0.0008, "loss": 1.2595, "step": 723 }, { "epoch": 1.0838323353293413, "grad_norm": 0.5, "learning_rate": 0.0008, "loss": 1.2605, "step": 724 }, { "epoch": 1.0853293413173652, "grad_norm": 0.294921875, "learning_rate": 0.0008, "loss": 1.2537, "step": 725 }, { "epoch": 1.0868263473053892, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.2526, "step": 726 }, { "epoch": 1.0883233532934131, "grad_norm": 0.2373046875, "learning_rate": 0.0008, "loss": 1.2418, "step": 727 }, { "epoch": 1.089820359281437, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.2439, "step": 728 }, { "epoch": 1.091317365269461, "grad_norm": 0.2109375, "learning_rate": 0.0008, "loss": 1.2389, "step": 729 }, { "epoch": 1.092814371257485, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.2388, "step": 730 }, { "epoch": 1.0943113772455089, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.2394, "step": 731 }, { "epoch": 1.095808383233533, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.2443, "step": 732 }, { "epoch": 1.097305389221557, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.2308, "step": 733 }, { "epoch": 1.098802395209581, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.23, "step": 734 }, { "epoch": 1.1002994011976048, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.2342, "step": 735 }, { "epoch": 1.1017964071856288, "grad_norm": 0.1689453125, "learning_rate": 0.0008, "loss": 1.2372, "step": 736 }, { "epoch": 1.1032934131736527, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.2339, "step": 737 }, { "epoch": 1.1047904191616766, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.2252, "step": 738 }, { "epoch": 1.1062874251497006, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.2298, "step": 739 }, { "epoch": 1.1077844311377245, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.2372, "step": 740 }, { "epoch": 1.1092814371257484, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.2223, "step": 741 }, { "epoch": 1.1107784431137724, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2322, "step": 742 }, { "epoch": 1.1122754491017963, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2261, "step": 743 }, { "epoch": 1.1137724550898203, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2329, "step": 744 }, { "epoch": 1.1152694610778444, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2268, "step": 745 }, { "epoch": 1.1167664670658684, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2296, "step": 746 }, { "epoch": 1.1182634730538923, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2252, "step": 747 }, { "epoch": 1.1197604790419162, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2342, "step": 748 }, { "epoch": 1.1212574850299402, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.2307, "step": 749 }, { "epoch": 1.122754491017964, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.224, "step": 750 }, { "epoch": 1.124251497005988, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.2255, "step": 751 }, { "epoch": 1.125748502994012, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2305, "step": 752 }, { "epoch": 1.127245508982036, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.228, "step": 753 }, { "epoch": 1.1287425149700598, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2216, "step": 754 }, { "epoch": 1.1302395209580838, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2284, "step": 755 }, { "epoch": 1.1317365269461077, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2271, "step": 756 }, { "epoch": 1.1332335329341316, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.2271, "step": 757 }, { "epoch": 1.1347305389221556, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.2278, "step": 758 }, { "epoch": 1.1362275449101795, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2244, "step": 759 }, { "epoch": 1.1377245508982037, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.2322, "step": 760 }, { "epoch": 1.1392215568862276, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2234, "step": 761 }, { "epoch": 1.1407185628742516, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2225, "step": 762 }, { "epoch": 1.1422155688622755, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2199, "step": 763 }, { "epoch": 1.1437125748502994, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.224, "step": 764 }, { "epoch": 1.1452095808383234, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2217, "step": 765 }, { "epoch": 1.1467065868263473, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2292, "step": 766 }, { "epoch": 1.1482035928143712, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2188, "step": 767 }, { "epoch": 1.1497005988023952, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.2298, "step": 768 }, { "epoch": 1.151197604790419, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2207, "step": 769 }, { "epoch": 1.152694610778443, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2295, "step": 770 }, { "epoch": 1.154191616766467, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.2203, "step": 771 }, { "epoch": 1.1556886227544911, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2198, "step": 772 }, { "epoch": 1.157185628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2228, "step": 773 }, { "epoch": 1.158682634730539, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.2141, "step": 774 }, { "epoch": 1.160179640718563, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.22, "step": 775 }, { "epoch": 1.1616766467065869, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2225, "step": 776 }, { "epoch": 1.1631736526946108, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2187, "step": 777 }, { "epoch": 1.1646706586826348, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.2214, "step": 778 }, { "epoch": 1.1661676646706587, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2203, "step": 779 }, { "epoch": 1.1676646706586826, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.2223, "step": 780 }, { "epoch": 1.1691616766467066, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.219, "step": 781 }, { "epoch": 1.1706586826347305, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.2217, "step": 782 }, { "epoch": 1.1721556886227544, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2189, "step": 783 }, { "epoch": 1.1736526946107784, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.2181, "step": 784 }, { "epoch": 1.1751497005988023, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2276, "step": 785 }, { "epoch": 1.1766467065868262, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.2178, "step": 786 }, { "epoch": 1.1781437125748502, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2254, "step": 787 }, { "epoch": 1.1796407185628743, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2179, "step": 788 }, { "epoch": 1.1811377245508983, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.219, "step": 789 }, { "epoch": 1.1826347305389222, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2253, "step": 790 }, { "epoch": 1.1841317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2195, "step": 791 }, { "epoch": 1.18562874251497, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2154, "step": 792 }, { "epoch": 1.187125748502994, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.2129, "step": 793 }, { "epoch": 1.188622754491018, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2196, "step": 794 }, { "epoch": 1.1901197604790419, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2216, "step": 795 }, { "epoch": 1.1916167664670658, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2218, "step": 796 }, { "epoch": 1.1931137724550898, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2223, "step": 797 }, { "epoch": 1.1946107784431137, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.2176, "step": 798 }, { "epoch": 1.1961077844311376, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2178, "step": 799 }, { "epoch": 1.1976047904191618, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2196, "step": 800 }, { "epoch": 1.1991017964071857, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.222, "step": 801 }, { "epoch": 1.2005988023952097, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2207, "step": 802 }, { "epoch": 1.2020958083832336, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2174, "step": 803 }, { "epoch": 1.2035928143712575, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2221, "step": 804 }, { "epoch": 1.2050898203592815, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2173, "step": 805 }, { "epoch": 1.2065868263473054, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2134, "step": 806 }, { "epoch": 1.2080838323353293, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2161, "step": 807 }, { "epoch": 1.2095808383233533, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.2128, "step": 808 }, { "epoch": 1.2110778443113772, "grad_norm": 0.2734375, "learning_rate": 0.0008, "loss": 1.2226, "step": 809 }, { "epoch": 1.2125748502994012, "grad_norm": 0.9140625, "learning_rate": 0.0008, "loss": 1.2246, "step": 810 }, { "epoch": 1.214071856287425, "grad_norm": 1.9140625, "learning_rate": 0.0008, "loss": 1.3315, "step": 811 }, { "epoch": 1.215568862275449, "grad_norm": 0.4453125, "learning_rate": 0.0008, "loss": 1.2585, "step": 812 }, { "epoch": 1.217065868263473, "grad_norm": 0.91015625, "learning_rate": 0.0008, "loss": 1.2749, "step": 813 }, { "epoch": 1.218562874251497, "grad_norm": 1.4375, "learning_rate": 0.0008, "loss": 1.2944, "step": 814 }, { "epoch": 1.220059880239521, "grad_norm": 0.4609375, "learning_rate": 0.0008, "loss": 1.2678, "step": 815 }, { "epoch": 1.221556886227545, "grad_norm": 0.3984375, "learning_rate": 0.0008, "loss": 1.2712, "step": 816 }, { "epoch": 1.223053892215569, "grad_norm": 0.359375, "learning_rate": 0.0008, "loss": 1.2548, "step": 817 }, { "epoch": 1.2245508982035929, "grad_norm": 0.318359375, "learning_rate": 0.0008, "loss": 1.2598, "step": 818 }, { "epoch": 1.2260479041916168, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.2494, "step": 819 }, { "epoch": 1.2275449101796407, "grad_norm": 0.2265625, "learning_rate": 0.0008, "loss": 1.258, "step": 820 }, { "epoch": 1.2290419161676647, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.2467, "step": 821 }, { "epoch": 1.2305389221556886, "grad_norm": 0.19140625, "learning_rate": 0.0008, "loss": 1.238, "step": 822 }, { "epoch": 1.2320359281437125, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.2271, "step": 823 }, { "epoch": 1.2335329341317365, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.2371, "step": 824 }, { "epoch": 1.2350299401197604, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.2355, "step": 825 }, { "epoch": 1.2365269461077844, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.2355, "step": 826 }, { "epoch": 1.2380239520958083, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2356, "step": 827 }, { "epoch": 1.2395209580838324, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.2343, "step": 828 }, { "epoch": 1.2410179640718564, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.2363, "step": 829 }, { "epoch": 1.2425149700598803, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.2363, "step": 830 }, { "epoch": 1.2440119760479043, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.22, "step": 831 }, { "epoch": 1.2455089820359282, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.2313, "step": 832 }, { "epoch": 1.2470059880239521, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2283, "step": 833 }, { "epoch": 1.248502994011976, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.232, "step": 834 }, { "epoch": 1.25, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2278, "step": 835 }, { "epoch": 1.251497005988024, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2336, "step": 836 }, { "epoch": 1.2529940119760479, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.218, "step": 837 }, { "epoch": 1.2544910179640718, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2327, "step": 838 }, { "epoch": 1.2559880239520957, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2223, "step": 839 }, { "epoch": 1.2574850299401197, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.2284, "step": 840 }, { "epoch": 1.2589820359281436, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2221, "step": 841 }, { "epoch": 1.2604790419161676, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.2212, "step": 842 }, { "epoch": 1.2619760479041915, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.2208, "step": 843 }, { "epoch": 1.2634730538922156, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2243, "step": 844 }, { "epoch": 1.2649700598802396, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2203, "step": 845 }, { "epoch": 1.2664670658682635, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.2163, "step": 846 }, { "epoch": 1.2679640718562875, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2256, "step": 847 }, { "epoch": 1.2694610778443114, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.2152, "step": 848 }, { "epoch": 1.2709580838323353, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2169, "step": 849 }, { "epoch": 1.2724550898203593, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2325, "step": 850 }, { "epoch": 1.2739520958083832, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2178, "step": 851 }, { "epoch": 1.2754491017964071, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.2087, "step": 852 }, { "epoch": 1.276946107784431, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2163, "step": 853 }, { "epoch": 1.278443113772455, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.2198, "step": 854 }, { "epoch": 1.2799401197604792, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2201, "step": 855 }, { "epoch": 1.281437125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.2171, "step": 856 }, { "epoch": 1.282934131736527, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.213, "step": 857 }, { "epoch": 1.284431137724551, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2188, "step": 858 }, { "epoch": 1.285928143712575, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2155, "step": 859 }, { "epoch": 1.2874251497005988, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2163, "step": 860 }, { "epoch": 1.2889221556886228, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2242, "step": 861 }, { "epoch": 1.2904191616766467, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2176, "step": 862 }, { "epoch": 1.2919161676646707, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.2127, "step": 863 }, { "epoch": 1.2934131736526946, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.2145, "step": 864 }, { "epoch": 1.2949101796407185, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2258, "step": 865 }, { "epoch": 1.2964071856287425, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.219, "step": 866 }, { "epoch": 1.2979041916167664, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2148, "step": 867 }, { "epoch": 1.2994011976047903, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.215, "step": 868 }, { "epoch": 1.3008982035928143, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.207, "step": 869 }, { "epoch": 1.3023952095808382, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2146, "step": 870 }, { "epoch": 1.3038922155688621, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.2173, "step": 871 }, { "epoch": 1.3053892215568863, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2191, "step": 872 }, { "epoch": 1.3068862275449102, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.2156, "step": 873 }, { "epoch": 1.3083832335329342, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.2124, "step": 874 }, { "epoch": 1.3098802395209581, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2191, "step": 875 }, { "epoch": 1.311377245508982, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.2193, "step": 876 }, { "epoch": 1.312874251497006, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2146, "step": 877 }, { "epoch": 1.31437125748503, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.2121, "step": 878 }, { "epoch": 1.3158682634730539, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2162, "step": 879 }, { "epoch": 1.3173652694610778, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2154, "step": 880 }, { "epoch": 1.3188622754491017, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.2199, "step": 881 }, { "epoch": 1.3203592814371259, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2234, "step": 882 }, { "epoch": 1.3218562874251498, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.222, "step": 883 }, { "epoch": 1.3233532934131738, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2121, "step": 884 }, { "epoch": 1.3248502994011977, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2114, "step": 885 }, { "epoch": 1.3263473053892216, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.2124, "step": 886 }, { "epoch": 1.3278443113772456, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2143, "step": 887 }, { "epoch": 1.3293413173652695, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.2189, "step": 888 }, { "epoch": 1.3308383233532934, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2093, "step": 889 }, { "epoch": 1.3323353293413174, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2117, "step": 890 }, { "epoch": 1.3338323353293413, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.2157, "step": 891 }, { "epoch": 1.3353293413173652, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2195, "step": 892 }, { "epoch": 1.3368263473053892, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1979, "step": 893 }, { "epoch": 1.3383233532934131, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.215, "step": 894 }, { "epoch": 1.339820359281437, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2084, "step": 895 }, { "epoch": 1.341317365269461, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2044, "step": 896 }, { "epoch": 1.342814371257485, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2095, "step": 897 }, { "epoch": 1.3443113772455089, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.2104, "step": 898 }, { "epoch": 1.3458083832335328, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2132, "step": 899 }, { "epoch": 1.347305389221557, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2128, "step": 900 }, { "epoch": 1.348802395209581, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2178, "step": 901 }, { "epoch": 1.3502994011976048, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2073, "step": 902 }, { "epoch": 1.3517964071856288, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2028, "step": 903 }, { "epoch": 1.3532934131736527, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.208, "step": 904 }, { "epoch": 1.3547904191616766, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2178, "step": 905 }, { "epoch": 1.3562874251497006, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.2105, "step": 906 }, { "epoch": 1.3577844311377245, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.213, "step": 907 }, { "epoch": 1.3592814371257484, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.2116, "step": 908 }, { "epoch": 1.3607784431137724, "grad_norm": 0.1982421875, "learning_rate": 0.0008, "loss": 1.2187, "step": 909 }, { "epoch": 1.3622754491017965, "grad_norm": 0.400390625, "learning_rate": 0.0008, "loss": 1.219, "step": 910 }, { "epoch": 1.3637724550898205, "grad_norm": 0.9140625, "learning_rate": 0.0008, "loss": 1.2323, "step": 911 }, { "epoch": 1.3652694610778444, "grad_norm": 1.8671875, "learning_rate": 0.0008, "loss": 1.2959, "step": 912 }, { "epoch": 1.3667664670658684, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.2367, "step": 913 }, { "epoch": 1.3682634730538923, "grad_norm": 1.0234375, "learning_rate": 0.0008, "loss": 1.2636, "step": 914 }, { "epoch": 1.3697604790419162, "grad_norm": 0.50390625, "learning_rate": 0.0008, "loss": 1.2372, "step": 915 }, { "epoch": 1.3712574850299402, "grad_norm": 0.5390625, "learning_rate": 0.0008, "loss": 1.2443, "step": 916 }, { "epoch": 1.372754491017964, "grad_norm": 0.30859375, "learning_rate": 0.0008, "loss": 1.2336, "step": 917 }, { "epoch": 1.374251497005988, "grad_norm": 0.392578125, "learning_rate": 0.0008, "loss": 1.2485, "step": 918 }, { "epoch": 1.375748502994012, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.2288, "step": 919 }, { "epoch": 1.377245508982036, "grad_norm": 0.28125, "learning_rate": 0.0008, "loss": 1.2385, "step": 920 }, { "epoch": 1.3787425149700598, "grad_norm": 0.232421875, "learning_rate": 0.0008, "loss": 1.2306, "step": 921 }, { "epoch": 1.3802395209580838, "grad_norm": 0.171875, "learning_rate": 0.0008, "loss": 1.2302, "step": 922 }, { "epoch": 1.3817365269461077, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.2309, "step": 923 }, { "epoch": 1.3832335329341316, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.2244, "step": 924 }, { "epoch": 1.3847305389221556, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.2262, "step": 925 }, { "epoch": 1.3862275449101795, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.208, "step": 926 }, { "epoch": 1.3877245508982035, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.2209, "step": 927 }, { "epoch": 1.3892215568862276, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.2219, "step": 928 }, { "epoch": 1.3907185628742516, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.216, "step": 929 }, { "epoch": 1.3922155688622755, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.2252, "step": 930 }, { "epoch": 1.3937125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2221, "step": 931 }, { "epoch": 1.3952095808383234, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2244, "step": 932 }, { "epoch": 1.3967065868263473, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2251, "step": 933 }, { "epoch": 1.3982035928143712, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.217, "step": 934 }, { "epoch": 1.3997005988023952, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.2183, "step": 935 }, { "epoch": 1.401197604790419, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.213, "step": 936 }, { "epoch": 1.402694610778443, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2141, "step": 937 }, { "epoch": 1.4041916167664672, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.2129, "step": 938 }, { "epoch": 1.4056886227544911, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2105, "step": 939 }, { "epoch": 1.407185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2148, "step": 940 }, { "epoch": 1.408682634730539, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2099, "step": 941 }, { "epoch": 1.410179640718563, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2095, "step": 942 }, { "epoch": 1.4116766467065869, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2154, "step": 943 }, { "epoch": 1.4131736526946108, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.217, "step": 944 }, { "epoch": 1.4146706586826348, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.2108, "step": 945 }, { "epoch": 1.4161676646706587, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.2119, "step": 946 }, { "epoch": 1.4176646706586826, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2122, "step": 947 }, { "epoch": 1.4191616766467066, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.2096, "step": 948 }, { "epoch": 1.4206586826347305, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.2052, "step": 949 }, { "epoch": 1.4221556886227544, "grad_norm": 0.051025390625, "learning_rate": 0.0008, "loss": 1.2113, "step": 950 }, { "epoch": 1.4236526946107784, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.2125, "step": 951 }, { "epoch": 1.4251497005988023, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.2084, "step": 952 }, { "epoch": 1.4266467065868262, "grad_norm": 0.049072265625, "learning_rate": 0.0008, "loss": 1.2112, "step": 953 }, { "epoch": 1.4281437125748502, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2088, "step": 954 }, { "epoch": 1.4296407185628741, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.2017, "step": 955 }, { "epoch": 1.4311377245508983, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2095, "step": 956 }, { "epoch": 1.4326347305389222, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.2201, "step": 957 }, { "epoch": 1.4341317365269461, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2119, "step": 958 }, { "epoch": 1.43562874251497, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.2162, "step": 959 }, { "epoch": 1.437125748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2035, "step": 960 }, { "epoch": 1.438622754491018, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.2084, "step": 961 }, { "epoch": 1.4401197604790419, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2118, "step": 962 }, { "epoch": 1.4416167664670658, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.2041, "step": 963 }, { "epoch": 1.4431137724550898, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.21, "step": 964 }, { "epoch": 1.4446107784431137, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.2105, "step": 965 }, { "epoch": 1.4461077844311379, "grad_norm": 0.046875, "learning_rate": 0.0008, "loss": 1.2103, "step": 966 }, { "epoch": 1.4476047904191618, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.2033, "step": 967 }, { "epoch": 1.4491017964071857, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.2092, "step": 968 }, { "epoch": 1.4505988023952097, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.2062, "step": 969 }, { "epoch": 1.4520958083832336, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.2027, "step": 970 }, { "epoch": 1.4535928143712575, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2035, "step": 971 }, { "epoch": 1.4550898203592815, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.2088, "step": 972 }, { "epoch": 1.4565868263473054, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2028, "step": 973 }, { "epoch": 1.4580838323353293, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2031, "step": 974 }, { "epoch": 1.4595808383233533, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2023, "step": 975 }, { "epoch": 1.4610778443113772, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.205, "step": 976 }, { "epoch": 1.4625748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2101, "step": 977 }, { "epoch": 1.464071856287425, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.2114, "step": 978 }, { "epoch": 1.465568862275449, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1998, "step": 979 }, { "epoch": 1.467065868263473, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2041, "step": 980 }, { "epoch": 1.468562874251497, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.2183, "step": 981 }, { "epoch": 1.4700598802395208, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2202, "step": 982 }, { "epoch": 1.471556886227545, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2125, "step": 983 }, { "epoch": 1.473053892215569, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2027, "step": 984 }, { "epoch": 1.4745508982035929, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2034, "step": 985 }, { "epoch": 1.4760479041916168, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.2009, "step": 986 }, { "epoch": 1.4775449101796407, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2107, "step": 987 }, { "epoch": 1.4790419161676647, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2079, "step": 988 }, { "epoch": 1.4805389221556886, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.2113, "step": 989 }, { "epoch": 1.4820359281437125, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.2029, "step": 990 }, { "epoch": 1.4835329341317365, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2102, "step": 991 }, { "epoch": 1.4850299401197604, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.2034, "step": 992 }, { "epoch": 1.4865269461077844, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.2087, "step": 993 }, { "epoch": 1.4880239520958085, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.2089, "step": 994 }, { "epoch": 1.4895209580838324, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.2062, "step": 995 }, { "epoch": 1.4910179640718564, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.2015, "step": 996 }, { "epoch": 1.4925149700598803, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2039, "step": 997 }, { "epoch": 1.4940119760479043, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.2031, "step": 998 }, { "epoch": 1.4955089820359282, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.2136, "step": 999 }, { "epoch": 1.4970059880239521, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2024, "step": 1000 }, { "epoch": 1.498502994011976, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.2109, "step": 1001 }, { "epoch": 1.5, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2178, "step": 1002 }, { "epoch": 1.501497005988024, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2105, "step": 1003 }, { "epoch": 1.5029940119760479, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2054, "step": 1004 }, { "epoch": 1.5044910179640718, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2028, "step": 1005 }, { "epoch": 1.5059880239520957, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2002, "step": 1006 }, { "epoch": 1.5074850299401197, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2092, "step": 1007 }, { "epoch": 1.5089820359281436, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2041, "step": 1008 }, { "epoch": 1.5104790419161676, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.2079, "step": 1009 }, { "epoch": 1.5119760479041915, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1992, "step": 1010 }, { "epoch": 1.5134730538922154, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2069, "step": 1011 }, { "epoch": 1.5149700598802394, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.2013, "step": 1012 }, { "epoch": 1.5164670658682635, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2034, "step": 1013 }, { "epoch": 1.5179640718562875, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2027, "step": 1014 }, { "epoch": 1.5194610778443114, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2092, "step": 1015 }, { "epoch": 1.5209580838323353, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.2086, "step": 1016 }, { "epoch": 1.5224550898203593, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2071, "step": 1017 }, { "epoch": 1.5239520958083832, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.2035, "step": 1018 }, { "epoch": 1.5254491017964071, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.203, "step": 1019 }, { "epoch": 1.5269461077844313, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.2158, "step": 1020 }, { "epoch": 1.5284431137724552, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2044, "step": 1021 }, { "epoch": 1.5299401197604792, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2039, "step": 1022 }, { "epoch": 1.531437125748503, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1994, "step": 1023 }, { "epoch": 1.532934131736527, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2034, "step": 1024 }, { "epoch": 1.534431137724551, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.2066, "step": 1025 }, { "epoch": 1.535928143712575, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2026, "step": 1026 }, { "epoch": 1.5374251497005988, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2077, "step": 1027 }, { "epoch": 1.5389221556886228, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2121, "step": 1028 }, { "epoch": 1.5404191616766467, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.2023, "step": 1029 }, { "epoch": 1.5419161676646707, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2043, "step": 1030 }, { "epoch": 1.5434131736526946, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.2075, "step": 1031 }, { "epoch": 1.5449101796407185, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2064, "step": 1032 }, { "epoch": 1.5464071856287425, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.2031, "step": 1033 }, { "epoch": 1.5479041916167664, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2031, "step": 1034 }, { "epoch": 1.5494011976047903, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2081, "step": 1035 }, { "epoch": 1.5508982035928143, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2075, "step": 1036 }, { "epoch": 1.5523952095808382, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2064, "step": 1037 }, { "epoch": 1.5538922155688621, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2106, "step": 1038 }, { "epoch": 1.555389221556886, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2055, "step": 1039 }, { "epoch": 1.55688622754491, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.2047, "step": 1040 }, { "epoch": 1.5583832335329342, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.2051, "step": 1041 }, { "epoch": 1.5598802395209581, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2072, "step": 1042 }, { "epoch": 1.561377245508982, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.2055, "step": 1043 }, { "epoch": 1.562874251497006, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2028, "step": 1044 }, { "epoch": 1.56437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.204, "step": 1045 }, { "epoch": 1.5658682634730539, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2117, "step": 1046 }, { "epoch": 1.5673652694610778, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2005, "step": 1047 }, { "epoch": 1.568862275449102, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.2032, "step": 1048 }, { "epoch": 1.5703592814371259, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2109, "step": 1049 }, { "epoch": 1.5718562874251498, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2036, "step": 1050 }, { "epoch": 1.5733532934131738, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1996, "step": 1051 }, { "epoch": 1.5748502994011977, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2076, "step": 1052 }, { "epoch": 1.5763473053892216, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2064, "step": 1053 }, { "epoch": 1.5778443113772456, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2028, "step": 1054 }, { "epoch": 1.5793413173652695, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.2081, "step": 1055 }, { "epoch": 1.5808383233532934, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1996, "step": 1056 }, { "epoch": 1.5823353293413174, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2042, "step": 1057 }, { "epoch": 1.5838323353293413, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1969, "step": 1058 }, { "epoch": 1.5853293413173652, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2051, "step": 1059 }, { "epoch": 1.5868263473053892, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2038, "step": 1060 }, { "epoch": 1.5883233532934131, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2081, "step": 1061 }, { "epoch": 1.589820359281437, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.2055, "step": 1062 }, { "epoch": 1.591317365269461, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.202, "step": 1063 }, { "epoch": 1.592814371257485, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2008, "step": 1064 }, { "epoch": 1.5943113772455089, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.2024, "step": 1065 }, { "epoch": 1.5958083832335328, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.2078, "step": 1066 }, { "epoch": 1.5973053892215567, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.2053, "step": 1067 }, { "epoch": 1.5988023952095807, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2041, "step": 1068 }, { "epoch": 1.6002994011976048, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2056, "step": 1069 }, { "epoch": 1.6017964071856288, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2009, "step": 1070 }, { "epoch": 1.6032934131736527, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.211, "step": 1071 }, { "epoch": 1.6047904191616766, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2129, "step": 1072 }, { "epoch": 1.6062874251497006, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2025, "step": 1073 }, { "epoch": 1.6077844311377245, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2019, "step": 1074 }, { "epoch": 1.6092814371257484, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.2059, "step": 1075 }, { "epoch": 1.6107784431137726, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.2036, "step": 1076 }, { "epoch": 1.6122754491017965, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2065, "step": 1077 }, { "epoch": 1.6137724550898205, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2114, "step": 1078 }, { "epoch": 1.6152694610778444, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2037, "step": 1079 }, { "epoch": 1.6167664670658684, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2015, "step": 1080 }, { "epoch": 1.6182634730538923, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.199, "step": 1081 }, { "epoch": 1.6197604790419162, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2123, "step": 1082 }, { "epoch": 1.6212574850299402, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.2005, "step": 1083 }, { "epoch": 1.622754491017964, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1993, "step": 1084 }, { "epoch": 1.624251497005988, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.204, "step": 1085 }, { "epoch": 1.625748502994012, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2042, "step": 1086 }, { "epoch": 1.627245508982036, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1993, "step": 1087 }, { "epoch": 1.6287425149700598, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.2052, "step": 1088 }, { "epoch": 1.6302395209580838, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2036, "step": 1089 }, { "epoch": 1.6317365269461077, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1921, "step": 1090 }, { "epoch": 1.6332335329341316, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2045, "step": 1091 }, { "epoch": 1.6347305389221556, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.2008, "step": 1092 }, { "epoch": 1.6362275449101795, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.204, "step": 1093 }, { "epoch": 1.6377245508982035, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1952, "step": 1094 }, { "epoch": 1.6392215568862274, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.201, "step": 1095 }, { "epoch": 1.6407185628742516, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1998, "step": 1096 }, { "epoch": 1.6422155688622755, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2058, "step": 1097 }, { "epoch": 1.6437125748502994, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.207, "step": 1098 }, { "epoch": 1.6452095808383234, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2026, "step": 1099 }, { "epoch": 1.6467065868263473, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1981, "step": 1100 }, { "epoch": 1.6482035928143712, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2022, "step": 1101 }, { "epoch": 1.6497005988023952, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.2003, "step": 1102 }, { "epoch": 1.6511976047904193, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1976, "step": 1103 }, { "epoch": 1.6526946107784433, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.2066, "step": 1104 }, { "epoch": 1.6541916167664672, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2024, "step": 1105 }, { "epoch": 1.6556886227544911, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.208, "step": 1106 }, { "epoch": 1.657185628742515, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1998, "step": 1107 }, { "epoch": 1.658682634730539, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2023, "step": 1108 }, { "epoch": 1.660179640718563, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.2033, "step": 1109 }, { "epoch": 1.6616766467065869, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1999, "step": 1110 }, { "epoch": 1.6631736526946108, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.2061, "step": 1111 }, { "epoch": 1.6646706586826348, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2035, "step": 1112 }, { "epoch": 1.6661676646706587, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1901, "step": 1113 }, { "epoch": 1.6676646706586826, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1957, "step": 1114 }, { "epoch": 1.6691616766467066, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1969, "step": 1115 }, { "epoch": 1.6706586826347305, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2045, "step": 1116 }, { "epoch": 1.6721556886227544, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2036, "step": 1117 }, { "epoch": 1.6736526946107784, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.2022, "step": 1118 }, { "epoch": 1.6751497005988023, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2028, "step": 1119 }, { "epoch": 1.6766467065868262, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2063, "step": 1120 }, { "epoch": 1.6781437125748502, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.203, "step": 1121 }, { "epoch": 1.6796407185628741, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2046, "step": 1122 }, { "epoch": 1.681137724550898, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2002, "step": 1123 }, { "epoch": 1.6826347305389222, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2078, "step": 1124 }, { "epoch": 1.6841317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2104, "step": 1125 }, { "epoch": 1.68562874251497, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.2006, "step": 1126 }, { "epoch": 1.687125748502994, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.207, "step": 1127 }, { "epoch": 1.688622754491018, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2039, "step": 1128 }, { "epoch": 1.6901197604790419, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.2034, "step": 1129 }, { "epoch": 1.6916167664670658, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2091, "step": 1130 }, { "epoch": 1.69311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1965, "step": 1131 }, { "epoch": 1.694610778443114, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2039, "step": 1132 }, { "epoch": 1.6961077844311379, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1979, "step": 1133 }, { "epoch": 1.6976047904191618, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.2006, "step": 1134 }, { "epoch": 1.6991017964071857, "grad_norm": 0.236328125, "learning_rate": 0.0008, "loss": 1.1998, "step": 1135 }, { "epoch": 1.7005988023952097, "grad_norm": 0.482421875, "learning_rate": 0.0008, "loss": 1.2086, "step": 1136 }, { "epoch": 1.7020958083832336, "grad_norm": 1.328125, "learning_rate": 0.0008, "loss": 1.2406, "step": 1137 }, { "epoch": 1.7035928143712575, "grad_norm": 1.28125, "learning_rate": 0.0008, "loss": 1.2445, "step": 1138 }, { "epoch": 1.7050898203592815, "grad_norm": 0.32421875, "learning_rate": 0.0008, "loss": 1.2307, "step": 1139 }, { "epoch": 1.7065868263473054, "grad_norm": 0.50390625, "learning_rate": 0.0008, "loss": 1.2324, "step": 1140 }, { "epoch": 1.7080838323353293, "grad_norm": 0.400390625, "learning_rate": 0.0008, "loss": 1.2266, "step": 1141 }, { "epoch": 1.7095808383233533, "grad_norm": 0.31640625, "learning_rate": 0.0008, "loss": 1.2242, "step": 1142 }, { "epoch": 1.7110778443113772, "grad_norm": 0.30078125, "learning_rate": 0.0008, "loss": 1.2131, "step": 1143 }, { "epoch": 1.7125748502994012, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.2149, "step": 1144 }, { "epoch": 1.714071856287425, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.2205, "step": 1145 }, { "epoch": 1.715568862275449, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.2154, "step": 1146 }, { "epoch": 1.717065868263473, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.2186, "step": 1147 }, { "epoch": 1.718562874251497, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.2107, "step": 1148 }, { "epoch": 1.7200598802395208, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.2122, "step": 1149 }, { "epoch": 1.7215568862275448, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.2155, "step": 1150 }, { "epoch": 1.7230538922155687, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.2003, "step": 1151 }, { "epoch": 1.7245508982035929, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.2131, "step": 1152 }, { "epoch": 1.7260479041916168, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.2108, "step": 1153 }, { "epoch": 1.7275449101796407, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2048, "step": 1154 }, { "epoch": 1.7290419161676647, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.2034, "step": 1155 }, { "epoch": 1.7305389221556886, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2099, "step": 1156 }, { "epoch": 1.7320359281437125, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2067, "step": 1157 }, { "epoch": 1.7335329341317365, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1989, "step": 1158 }, { "epoch": 1.7350299401197606, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2037, "step": 1159 }, { "epoch": 1.7365269461077846, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1999, "step": 1160 }, { "epoch": 1.7380239520958085, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.2002, "step": 1161 }, { "epoch": 1.7395209580838324, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.2082, "step": 1162 }, { "epoch": 1.7410179640718564, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2056, "step": 1163 }, { "epoch": 1.7425149700598803, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.204, "step": 1164 }, { "epoch": 1.7440119760479043, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.2007, "step": 1165 }, { "epoch": 1.7455089820359282, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.2006, "step": 1166 }, { "epoch": 1.7470059880239521, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2037, "step": 1167 }, { "epoch": 1.748502994011976, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1902, "step": 1168 }, { "epoch": 1.75, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2009, "step": 1169 }, { "epoch": 1.751497005988024, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1987, "step": 1170 }, { "epoch": 1.7529940119760479, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.2076, "step": 1171 }, { "epoch": 1.7544910179640718, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2014, "step": 1172 }, { "epoch": 1.7559880239520957, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1927, "step": 1173 }, { "epoch": 1.7574850299401197, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2002, "step": 1174 }, { "epoch": 1.7589820359281436, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.2018, "step": 1175 }, { "epoch": 1.7604790419161676, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1954, "step": 1176 }, { "epoch": 1.7619760479041915, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1957, "step": 1177 }, { "epoch": 1.7634730538922154, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.201, "step": 1178 }, { "epoch": 1.7649700598802394, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1982, "step": 1179 }, { "epoch": 1.7664670658682635, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1993, "step": 1180 }, { "epoch": 1.7679640718562875, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1968, "step": 1181 }, { "epoch": 1.7694610778443114, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.198, "step": 1182 }, { "epoch": 1.7709580838323353, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1968, "step": 1183 }, { "epoch": 1.7724550898203593, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.2056, "step": 1184 }, { "epoch": 1.7739520958083832, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1906, "step": 1185 }, { "epoch": 1.7754491017964071, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.206, "step": 1186 }, { "epoch": 1.7769461077844313, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1957, "step": 1187 }, { "epoch": 1.7784431137724552, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1938, "step": 1188 }, { "epoch": 1.7799401197604792, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1938, "step": 1189 }, { "epoch": 1.781437125748503, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.2022, "step": 1190 }, { "epoch": 1.782934131736527, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2001, "step": 1191 }, { "epoch": 1.784431137724551, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2002, "step": 1192 }, { "epoch": 1.785928143712575, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2004, "step": 1193 }, { "epoch": 1.7874251497005988, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1993, "step": 1194 }, { "epoch": 1.7889221556886228, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2006, "step": 1195 }, { "epoch": 1.7904191616766467, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2053, "step": 1196 }, { "epoch": 1.7919161676646707, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2009, "step": 1197 }, { "epoch": 1.7934131736526946, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1989, "step": 1198 }, { "epoch": 1.7949101796407185, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.199, "step": 1199 }, { "epoch": 1.7964071856287425, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1943, "step": 1200 }, { "epoch": 1.7979041916167664, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1942, "step": 1201 }, { "epoch": 1.7994011976047903, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1922, "step": 1202 }, { "epoch": 1.8008982035928143, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1993, "step": 1203 }, { "epoch": 1.8023952095808382, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.197, "step": 1204 }, { "epoch": 1.8038922155688621, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.2087, "step": 1205 }, { "epoch": 1.805389221556886, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1982, "step": 1206 }, { "epoch": 1.80688622754491, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1941, "step": 1207 }, { "epoch": 1.8083832335329342, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1965, "step": 1208 }, { "epoch": 1.8098802395209581, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2009, "step": 1209 }, { "epoch": 1.811377245508982, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1945, "step": 1210 }, { "epoch": 1.812874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1946, "step": 1211 }, { "epoch": 1.81437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1955, "step": 1212 }, { "epoch": 1.8158682634730539, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2027, "step": 1213 }, { "epoch": 1.8173652694610778, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.2017, "step": 1214 }, { "epoch": 1.818862275449102, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2033, "step": 1215 }, { "epoch": 1.8203592814371259, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2018, "step": 1216 }, { "epoch": 1.8218562874251498, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.2031, "step": 1217 }, { "epoch": 1.8233532934131738, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1977, "step": 1218 }, { "epoch": 1.8248502994011977, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1978, "step": 1219 }, { "epoch": 1.8263473053892216, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.2027, "step": 1220 }, { "epoch": 1.8278443113772456, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1979, "step": 1221 }, { "epoch": 1.8293413173652695, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.2007, "step": 1222 }, { "epoch": 1.8308383233532934, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2038, "step": 1223 }, { "epoch": 1.8323353293413174, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.197, "step": 1224 }, { "epoch": 1.8338323353293413, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1969, "step": 1225 }, { "epoch": 1.8353293413173652, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2045, "step": 1226 }, { "epoch": 1.8368263473053892, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1982, "step": 1227 }, { "epoch": 1.8383233532934131, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1956, "step": 1228 }, { "epoch": 1.839820359281437, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1997, "step": 1229 }, { "epoch": 1.841317365269461, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1935, "step": 1230 }, { "epoch": 1.842814371257485, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1979, "step": 1231 }, { "epoch": 1.8443113772455089, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1963, "step": 1232 }, { "epoch": 1.8458083832335328, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2048, "step": 1233 }, { "epoch": 1.8473053892215567, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1986, "step": 1234 }, { "epoch": 1.8488023952095807, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1992, "step": 1235 }, { "epoch": 1.8502994011976048, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1962, "step": 1236 }, { "epoch": 1.8517964071856288, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1967, "step": 1237 }, { "epoch": 1.8532934131736527, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1933, "step": 1238 }, { "epoch": 1.8547904191616766, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1951, "step": 1239 }, { "epoch": 1.8562874251497006, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1929, "step": 1240 }, { "epoch": 1.8577844311377245, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1963, "step": 1241 }, { "epoch": 1.8592814371257484, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1939, "step": 1242 }, { "epoch": 1.8607784431137726, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1955, "step": 1243 }, { "epoch": 1.8622754491017965, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1972, "step": 1244 }, { "epoch": 1.8637724550898205, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2039, "step": 1245 }, { "epoch": 1.8652694610778444, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1959, "step": 1246 }, { "epoch": 1.8667664670658684, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1993, "step": 1247 }, { "epoch": 1.8682634730538923, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1956, "step": 1248 }, { "epoch": 1.8697604790419162, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1985, "step": 1249 }, { "epoch": 1.8712574850299402, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1916, "step": 1250 }, { "epoch": 1.872754491017964, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1957, "step": 1251 }, { "epoch": 1.874251497005988, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1984, "step": 1252 }, { "epoch": 1.875748502994012, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1953, "step": 1253 }, { "epoch": 1.877245508982036, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.19, "step": 1254 }, { "epoch": 1.8787425149700598, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1997, "step": 1255 }, { "epoch": 1.8802395209580838, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1897, "step": 1256 }, { "epoch": 1.8817365269461077, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1943, "step": 1257 }, { "epoch": 1.8832335329341316, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1961, "step": 1258 }, { "epoch": 1.8847305389221556, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1967, "step": 1259 }, { "epoch": 1.8862275449101795, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1975, "step": 1260 }, { "epoch": 1.8877245508982035, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1958, "step": 1261 }, { "epoch": 1.8892215568862274, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1961, "step": 1262 }, { "epoch": 1.8907185628742516, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1865, "step": 1263 }, { "epoch": 1.8922155688622755, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1995, "step": 1264 }, { "epoch": 1.8937125748502994, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.195, "step": 1265 }, { "epoch": 1.8952095808383234, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1946, "step": 1266 }, { "epoch": 1.8967065868263473, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.2047, "step": 1267 }, { "epoch": 1.8982035928143712, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1946, "step": 1268 }, { "epoch": 1.8997005988023952, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1983, "step": 1269 }, { "epoch": 1.9011976047904193, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.197, "step": 1270 }, { "epoch": 1.9026946107784433, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.193, "step": 1271 }, { "epoch": 1.9041916167664672, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1948, "step": 1272 }, { "epoch": 1.9056886227544911, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1926, "step": 1273 }, { "epoch": 1.907185628742515, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1966, "step": 1274 }, { "epoch": 1.908682634730539, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2019, "step": 1275 }, { "epoch": 1.910179640718563, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1955, "step": 1276 }, { "epoch": 1.9116766467065869, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1993, "step": 1277 }, { "epoch": 1.9131736526946108, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1926, "step": 1278 }, { "epoch": 1.9146706586826348, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1943, "step": 1279 }, { "epoch": 1.9161676646706587, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2037, "step": 1280 }, { "epoch": 1.9176646706586826, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.194, "step": 1281 }, { "epoch": 1.9191616766467066, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1899, "step": 1282 }, { "epoch": 1.9206586826347305, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2013, "step": 1283 }, { "epoch": 1.9221556886227544, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1933, "step": 1284 }, { "epoch": 1.9236526946107784, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1984, "step": 1285 }, { "epoch": 1.9251497005988023, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1964, "step": 1286 }, { "epoch": 1.9266467065868262, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2036, "step": 1287 }, { "epoch": 1.9281437125748502, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1956, "step": 1288 }, { "epoch": 1.9296407185628741, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1956, "step": 1289 }, { "epoch": 1.931137724550898, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1928, "step": 1290 }, { "epoch": 1.9326347305389222, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1901, "step": 1291 }, { "epoch": 1.9341317365269461, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2008, "step": 1292 }, { "epoch": 1.93562874251497, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1991, "step": 1293 }, { "epoch": 1.937125748502994, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1963, "step": 1294 }, { "epoch": 1.938622754491018, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1948, "step": 1295 }, { "epoch": 1.9401197604790419, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1986, "step": 1296 }, { "epoch": 1.9416167664670658, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1952, "step": 1297 }, { "epoch": 1.94311377245509, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.195, "step": 1298 }, { "epoch": 1.944610778443114, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2005, "step": 1299 }, { "epoch": 1.9461077844311379, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1994, "step": 1300 }, { "epoch": 1.9476047904191618, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1986, "step": 1301 }, { "epoch": 1.9491017964071857, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1969, "step": 1302 }, { "epoch": 1.9505988023952097, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1933, "step": 1303 }, { "epoch": 1.9520958083832336, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1958, "step": 1304 }, { "epoch": 1.9535928143712575, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1988, "step": 1305 }, { "epoch": 1.9550898203592815, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1983, "step": 1306 }, { "epoch": 1.9565868263473054, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.2006, "step": 1307 }, { "epoch": 1.9580838323353293, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1964, "step": 1308 }, { "epoch": 1.9595808383233533, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1933, "step": 1309 }, { "epoch": 1.9610778443113772, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2014, "step": 1310 }, { "epoch": 1.9625748502994012, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.195, "step": 1311 }, { "epoch": 1.964071856287425, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1968, "step": 1312 }, { "epoch": 1.965568862275449, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1899, "step": 1313 }, { "epoch": 1.967065868263473, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1916, "step": 1314 }, { "epoch": 1.968562874251497, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1935, "step": 1315 }, { "epoch": 1.9700598802395208, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1894, "step": 1316 }, { "epoch": 1.9715568862275448, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.194, "step": 1317 }, { "epoch": 1.9730538922155687, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1992, "step": 1318 }, { "epoch": 1.9745508982035929, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1932, "step": 1319 }, { "epoch": 1.9760479041916168, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1972, "step": 1320 }, { "epoch": 1.9775449101796407, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1878, "step": 1321 }, { "epoch": 1.9790419161676647, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1934, "step": 1322 }, { "epoch": 1.9805389221556886, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1848, "step": 1323 }, { "epoch": 1.9820359281437125, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1922, "step": 1324 }, { "epoch": 1.9835329341317365, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2001, "step": 1325 }, { "epoch": 1.9850299401197606, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1906, "step": 1326 }, { "epoch": 1.9865269461077846, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1969, "step": 1327 }, { "epoch": 1.9880239520958085, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1859, "step": 1328 }, { "epoch": 1.9895209580838324, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1982, "step": 1329 }, { "epoch": 1.9910179640718564, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1932, "step": 1330 }, { "epoch": 1.9925149700598803, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1922, "step": 1331 }, { "epoch": 1.9940119760479043, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1961, "step": 1332 }, { "epoch": 1.9955089820359282, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.195, "step": 1333 }, { "epoch": 1.9970059880239521, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1998, "step": 1334 }, { "epoch": 1.998502994011976, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.2005, "step": 1335 }, { "epoch": 2.0, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.2005, "step": 1336 }, { "epoch": 2.001497005988024, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1958, "step": 1337 }, { "epoch": 2.002994011976048, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1954, "step": 1338 }, { "epoch": 2.004491017964072, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1946, "step": 1339 }, { "epoch": 2.0059880239520957, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.199, "step": 1340 }, { "epoch": 2.0074850299401197, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1871, "step": 1341 }, { "epoch": 2.0089820359281436, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1966, "step": 1342 }, { "epoch": 2.0104790419161676, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1905, "step": 1343 }, { "epoch": 2.0119760479041915, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.199, "step": 1344 }, { "epoch": 2.0134730538922154, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1907, "step": 1345 }, { "epoch": 2.0149700598802394, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2011, "step": 1346 }, { "epoch": 2.0164670658682633, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1897, "step": 1347 }, { "epoch": 2.0179640718562872, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1983, "step": 1348 }, { "epoch": 2.019461077844311, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1901, "step": 1349 }, { "epoch": 2.020958083832335, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1962, "step": 1350 }, { "epoch": 2.0224550898203595, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.194, "step": 1351 }, { "epoch": 2.0239520958083834, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1909, "step": 1352 }, { "epoch": 2.0254491017964074, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2032, "step": 1353 }, { "epoch": 2.0269461077844313, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1925, "step": 1354 }, { "epoch": 2.0284431137724552, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1884, "step": 1355 }, { "epoch": 2.029940119760479, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1977, "step": 1356 }, { "epoch": 2.031437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1943, "step": 1357 }, { "epoch": 2.032934131736527, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1926, "step": 1358 }, { "epoch": 2.034431137724551, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1863, "step": 1359 }, { "epoch": 2.035928143712575, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1936, "step": 1360 }, { "epoch": 2.037425149700599, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1934, "step": 1361 }, { "epoch": 2.038922155688623, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1982, "step": 1362 }, { "epoch": 2.0404191616766467, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1959, "step": 1363 }, { "epoch": 2.0419161676646707, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1966, "step": 1364 }, { "epoch": 2.0434131736526946, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1923, "step": 1365 }, { "epoch": 2.0449101796407185, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.194, "step": 1366 }, { "epoch": 2.0464071856287425, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1953, "step": 1367 }, { "epoch": 2.0479041916167664, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.193, "step": 1368 }, { "epoch": 2.0494011976047903, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.194, "step": 1369 }, { "epoch": 2.0508982035928143, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1945, "step": 1370 }, { "epoch": 2.052395209580838, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1969, "step": 1371 }, { "epoch": 2.053892215568862, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1943, "step": 1372 }, { "epoch": 2.055389221556886, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1939, "step": 1373 }, { "epoch": 2.05688622754491, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1909, "step": 1374 }, { "epoch": 2.058383233532934, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1855, "step": 1375 }, { "epoch": 2.059880239520958, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1896, "step": 1376 }, { "epoch": 2.061377245508982, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.188, "step": 1377 }, { "epoch": 2.062874251497006, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1959, "step": 1378 }, { "epoch": 2.06437125748503, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1956, "step": 1379 }, { "epoch": 2.065868263473054, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1839, "step": 1380 }, { "epoch": 2.067365269461078, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1939, "step": 1381 }, { "epoch": 2.068862275449102, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1906, "step": 1382 }, { "epoch": 2.070359281437126, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1945, "step": 1383 }, { "epoch": 2.07185628742515, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1945, "step": 1384 }, { "epoch": 2.0733532934131738, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1956, "step": 1385 }, { "epoch": 2.0748502994011977, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1963, "step": 1386 }, { "epoch": 2.0763473053892216, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1867, "step": 1387 }, { "epoch": 2.0778443113772456, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1957, "step": 1388 }, { "epoch": 2.0793413173652695, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1934, "step": 1389 }, { "epoch": 2.0808383233532934, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.193, "step": 1390 }, { "epoch": 2.0823353293413174, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1868, "step": 1391 }, { "epoch": 2.0838323353293413, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1901, "step": 1392 }, { "epoch": 2.0853293413173652, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1895, "step": 1393 }, { "epoch": 2.086826347305389, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1976, "step": 1394 }, { "epoch": 2.088323353293413, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1947, "step": 1395 }, { "epoch": 2.089820359281437, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1936, "step": 1396 }, { "epoch": 2.091317365269461, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1937, "step": 1397 }, { "epoch": 2.092814371257485, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.186, "step": 1398 }, { "epoch": 2.094311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1921, "step": 1399 }, { "epoch": 2.095808383233533, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1882, "step": 1400 }, { "epoch": 2.0973053892215567, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1893, "step": 1401 }, { "epoch": 2.0988023952095807, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1941, "step": 1402 }, { "epoch": 2.1002994011976046, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1909, "step": 1403 }, { "epoch": 2.1017964071856285, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1901, "step": 1404 }, { "epoch": 2.1032934131736525, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1978, "step": 1405 }, { "epoch": 2.1047904191616764, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1931, "step": 1406 }, { "epoch": 2.106287425149701, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1882, "step": 1407 }, { "epoch": 2.1077844311377247, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.186, "step": 1408 }, { "epoch": 2.1092814371257487, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1917, "step": 1409 }, { "epoch": 2.1107784431137726, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1915, "step": 1410 }, { "epoch": 2.1122754491017965, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1943, "step": 1411 }, { "epoch": 2.1137724550898205, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1916, "step": 1412 }, { "epoch": 2.1152694610778444, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1839, "step": 1413 }, { "epoch": 2.1167664670658684, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1864, "step": 1414 }, { "epoch": 2.1182634730538923, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1986, "step": 1415 }, { "epoch": 2.1197604790419162, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1941, "step": 1416 }, { "epoch": 2.12125748502994, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1936, "step": 1417 }, { "epoch": 2.122754491017964, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1944, "step": 1418 }, { "epoch": 2.124251497005988, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1981, "step": 1419 }, { "epoch": 2.125748502994012, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1893, "step": 1420 }, { "epoch": 2.127245508982036, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1952, "step": 1421 }, { "epoch": 2.12874251497006, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1851, "step": 1422 }, { "epoch": 2.1302395209580838, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1933, "step": 1423 }, { "epoch": 2.1317365269461077, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1958, "step": 1424 }, { "epoch": 2.1332335329341316, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1903, "step": 1425 }, { "epoch": 2.1347305389221556, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.191, "step": 1426 }, { "epoch": 2.1362275449101795, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1894, "step": 1427 }, { "epoch": 2.1377245508982035, "grad_norm": 0.228515625, "learning_rate": 0.0008, "loss": 1.1844, "step": 1428 }, { "epoch": 2.1392215568862274, "grad_norm": 0.419921875, "learning_rate": 0.0008, "loss": 1.1981, "step": 1429 }, { "epoch": 2.1407185628742513, "grad_norm": 0.99609375, "learning_rate": 0.0008, "loss": 1.205, "step": 1430 }, { "epoch": 2.1422155688622753, "grad_norm": 1.875, "learning_rate": 0.0008, "loss": 1.2759, "step": 1431 }, { "epoch": 2.143712574850299, "grad_norm": 0.490234375, "learning_rate": 0.0008, "loss": 1.2266, "step": 1432 }, { "epoch": 2.1452095808383236, "grad_norm": 1.2109375, "learning_rate": 0.0008, "loss": 1.2652, "step": 1433 }, { "epoch": 2.1467065868263475, "grad_norm": 0.423828125, "learning_rate": 0.0008, "loss": 1.2499, "step": 1434 }, { "epoch": 2.1482035928143715, "grad_norm": 0.6171875, "learning_rate": 0.0008, "loss": 1.2517, "step": 1435 }, { "epoch": 2.1497005988023954, "grad_norm": 0.388671875, "learning_rate": 0.0008, "loss": 1.2465, "step": 1436 }, { "epoch": 2.1511976047904193, "grad_norm": 0.341796875, "learning_rate": 0.0008, "loss": 1.2361, "step": 1437 }, { "epoch": 2.1526946107784433, "grad_norm": 0.365234375, "learning_rate": 0.0008, "loss": 1.2366, "step": 1438 }, { "epoch": 2.154191616766467, "grad_norm": 0.322265625, "learning_rate": 0.0008, "loss": 1.2336, "step": 1439 }, { "epoch": 2.155688622754491, "grad_norm": 0.248046875, "learning_rate": 0.0008, "loss": 1.23, "step": 1440 }, { "epoch": 2.157185628742515, "grad_norm": 0.2490234375, "learning_rate": 0.0008, "loss": 1.2179, "step": 1441 }, { "epoch": 2.158682634730539, "grad_norm": 0.265625, "learning_rate": 0.0008, "loss": 1.2181, "step": 1442 }, { "epoch": 2.160179640718563, "grad_norm": 0.1943359375, "learning_rate": 0.0008, "loss": 1.2198, "step": 1443 }, { "epoch": 2.161676646706587, "grad_norm": 0.193359375, "learning_rate": 0.0008, "loss": 1.2163, "step": 1444 }, { "epoch": 2.163173652694611, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.212, "step": 1445 }, { "epoch": 2.1646706586826348, "grad_norm": 0.1630859375, "learning_rate": 0.0008, "loss": 1.2124, "step": 1446 }, { "epoch": 2.1661676646706587, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.2083, "step": 1447 }, { "epoch": 2.1676646706586826, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.2074, "step": 1448 }, { "epoch": 2.1691616766467066, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.2028, "step": 1449 }, { "epoch": 2.1706586826347305, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.2135, "step": 1450 }, { "epoch": 2.1721556886227544, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.2023, "step": 1451 }, { "epoch": 2.1736526946107784, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2095, "step": 1452 }, { "epoch": 2.1751497005988023, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.209, "step": 1453 }, { "epoch": 2.1766467065868262, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2018, "step": 1454 }, { "epoch": 2.17814371257485, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1958, "step": 1455 }, { "epoch": 2.179640718562874, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1998, "step": 1456 }, { "epoch": 2.181137724550898, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2, "step": 1457 }, { "epoch": 2.182634730538922, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2027, "step": 1458 }, { "epoch": 2.184131736526946, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1967, "step": 1459 }, { "epoch": 2.18562874251497, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2036, "step": 1460 }, { "epoch": 2.187125748502994, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1984, "step": 1461 }, { "epoch": 2.1886227544910177, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2036, "step": 1462 }, { "epoch": 2.190119760479042, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.205, "step": 1463 }, { "epoch": 2.191616766467066, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1921, "step": 1464 }, { "epoch": 2.19311377245509, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.199, "step": 1465 }, { "epoch": 2.194610778443114, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1914, "step": 1466 }, { "epoch": 2.196107784431138, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1985, "step": 1467 }, { "epoch": 2.197604790419162, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2006, "step": 1468 }, { "epoch": 2.1991017964071857, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1987, "step": 1469 }, { "epoch": 2.2005988023952097, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1985, "step": 1470 }, { "epoch": 2.2020958083832336, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1966, "step": 1471 }, { "epoch": 2.2035928143712575, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1943, "step": 1472 }, { "epoch": 2.2050898203592815, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2011, "step": 1473 }, { "epoch": 2.2065868263473054, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1975, "step": 1474 }, { "epoch": 2.2080838323353293, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2034, "step": 1475 }, { "epoch": 2.2095808383233533, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1873, "step": 1476 }, { "epoch": 2.211077844311377, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2029, "step": 1477 }, { "epoch": 2.212574850299401, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1927, "step": 1478 }, { "epoch": 2.214071856287425, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.188, "step": 1479 }, { "epoch": 2.215568862275449, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1931, "step": 1480 }, { "epoch": 2.217065868263473, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1884, "step": 1481 }, { "epoch": 2.218562874251497, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1996, "step": 1482 }, { "epoch": 2.220059880239521, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1932, "step": 1483 }, { "epoch": 2.2215568862275448, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.197, "step": 1484 }, { "epoch": 2.2230538922155687, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1909, "step": 1485 }, { "epoch": 2.2245508982035926, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1947, "step": 1486 }, { "epoch": 2.2260479041916166, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1912, "step": 1487 }, { "epoch": 2.2275449101796405, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1913, "step": 1488 }, { "epoch": 2.229041916167665, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1919, "step": 1489 }, { "epoch": 2.230538922155689, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.187, "step": 1490 }, { "epoch": 2.2320359281437128, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1928, "step": 1491 }, { "epoch": 2.2335329341317367, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1785, "step": 1492 }, { "epoch": 2.2350299401197606, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1893, "step": 1493 }, { "epoch": 2.2365269461077846, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1801, "step": 1494 }, { "epoch": 2.2380239520958085, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1915, "step": 1495 }, { "epoch": 2.2395209580838324, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1852, "step": 1496 }, { "epoch": 2.2410179640718564, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1882, "step": 1497 }, { "epoch": 2.2425149700598803, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1948, "step": 1498 }, { "epoch": 2.2440119760479043, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1855, "step": 1499 }, { "epoch": 2.245508982035928, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1905, "step": 1500 }, { "epoch": 2.247005988023952, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1941, "step": 1501 }, { "epoch": 2.248502994011976, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1944, "step": 1502 }, { "epoch": 2.25, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1907, "step": 1503 }, { "epoch": 2.251497005988024, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1924, "step": 1504 }, { "epoch": 2.252994011976048, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.189, "step": 1505 }, { "epoch": 2.254491017964072, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1919, "step": 1506 }, { "epoch": 2.2559880239520957, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1871, "step": 1507 }, { "epoch": 2.2574850299401197, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.196, "step": 1508 }, { "epoch": 2.2589820359281436, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1979, "step": 1509 }, { "epoch": 2.2604790419161676, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1822, "step": 1510 }, { "epoch": 2.2619760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1863, "step": 1511 }, { "epoch": 2.2634730538922154, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1856, "step": 1512 }, { "epoch": 2.2649700598802394, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1891, "step": 1513 }, { "epoch": 2.2664670658682633, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1848, "step": 1514 }, { "epoch": 2.2679640718562872, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1944, "step": 1515 }, { "epoch": 2.269461077844311, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1957, "step": 1516 }, { "epoch": 2.270958083832335, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1876, "step": 1517 }, { "epoch": 2.272455089820359, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.19, "step": 1518 }, { "epoch": 2.2739520958083834, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1802, "step": 1519 }, { "epoch": 2.2754491017964074, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1838, "step": 1520 }, { "epoch": 2.2769461077844313, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1959, "step": 1521 }, { "epoch": 2.2784431137724552, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1993, "step": 1522 }, { "epoch": 2.279940119760479, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1901, "step": 1523 }, { "epoch": 2.281437125748503, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1815, "step": 1524 }, { "epoch": 2.282934131736527, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1943, "step": 1525 }, { "epoch": 2.284431137724551, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.2008, "step": 1526 }, { "epoch": 2.285928143712575, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1897, "step": 1527 }, { "epoch": 2.287425149700599, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.192, "step": 1528 }, { "epoch": 2.288922155688623, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.192, "step": 1529 }, { "epoch": 2.2904191616766467, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1904, "step": 1530 }, { "epoch": 2.2919161676646707, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1796, "step": 1531 }, { "epoch": 2.2934131736526946, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1839, "step": 1532 }, { "epoch": 2.2949101796407185, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1858, "step": 1533 }, { "epoch": 2.2964071856287425, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1899, "step": 1534 }, { "epoch": 2.2979041916167664, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2006, "step": 1535 }, { "epoch": 2.2994011976047903, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1873, "step": 1536 }, { "epoch": 2.3008982035928143, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.187, "step": 1537 }, { "epoch": 2.302395209580838, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1934, "step": 1538 }, { "epoch": 2.303892215568862, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1892, "step": 1539 }, { "epoch": 2.305389221556886, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1885, "step": 1540 }, { "epoch": 2.30688622754491, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.197, "step": 1541 }, { "epoch": 2.308383233532934, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1912, "step": 1542 }, { "epoch": 2.309880239520958, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1864, "step": 1543 }, { "epoch": 2.3113772455089823, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1952, "step": 1544 }, { "epoch": 2.312874251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1903, "step": 1545 }, { "epoch": 2.31437125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1918, "step": 1546 }, { "epoch": 2.315868263473054, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1836, "step": 1547 }, { "epoch": 2.317365269461078, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.187, "step": 1548 }, { "epoch": 2.318862275449102, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1908, "step": 1549 }, { "epoch": 2.320359281437126, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1937, "step": 1550 }, { "epoch": 2.32185628742515, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1874, "step": 1551 }, { "epoch": 2.3233532934131738, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1552 }, { "epoch": 2.3248502994011977, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1883, "step": 1553 }, { "epoch": 2.3263473053892216, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1901, "step": 1554 }, { "epoch": 2.3278443113772456, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1942, "step": 1555 }, { "epoch": 2.3293413173652695, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1838, "step": 1556 }, { "epoch": 2.3308383233532934, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.187, "step": 1557 }, { "epoch": 2.3323353293413174, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1858, "step": 1558 }, { "epoch": 2.3338323353293413, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1824, "step": 1559 }, { "epoch": 2.3353293413173652, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1842, "step": 1560 }, { "epoch": 2.336826347305389, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1931, "step": 1561 }, { "epoch": 2.338323353293413, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1898, "step": 1562 }, { "epoch": 2.339820359281437, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.185, "step": 1563 }, { "epoch": 2.341317365269461, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1816, "step": 1564 }, { "epoch": 2.342814371257485, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1845, "step": 1565 }, { "epoch": 2.344311377245509, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1906, "step": 1566 }, { "epoch": 2.345808383233533, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1877, "step": 1567 }, { "epoch": 2.3473053892215567, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1931, "step": 1568 }, { "epoch": 2.3488023952095807, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1833, "step": 1569 }, { "epoch": 2.3502994011976046, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1893, "step": 1570 }, { "epoch": 2.3517964071856285, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1978, "step": 1571 }, { "epoch": 2.3532934131736525, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1813, "step": 1572 }, { "epoch": 2.3547904191616764, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1875, "step": 1573 }, { "epoch": 2.3562874251497004, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1831, "step": 1574 }, { "epoch": 2.3577844311377247, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1852, "step": 1575 }, { "epoch": 2.3592814371257487, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1816, "step": 1576 }, { "epoch": 2.3607784431137726, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1909, "step": 1577 }, { "epoch": 2.3622754491017965, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1883, "step": 1578 }, { "epoch": 2.3637724550898205, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1891, "step": 1579 }, { "epoch": 2.3652694610778444, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.189, "step": 1580 }, { "epoch": 2.3667664670658684, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1867, "step": 1581 }, { "epoch": 2.3682634730538923, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1893, "step": 1582 }, { "epoch": 2.3697604790419162, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1884, "step": 1583 }, { "epoch": 2.37125748502994, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.182, "step": 1584 }, { "epoch": 2.372754491017964, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.187, "step": 1585 }, { "epoch": 2.374251497005988, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1887, "step": 1586 }, { "epoch": 2.375748502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1878, "step": 1587 }, { "epoch": 2.377245508982036, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1938, "step": 1588 }, { "epoch": 2.37874251497006, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1822, "step": 1589 }, { "epoch": 2.3802395209580838, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1834, "step": 1590 }, { "epoch": 2.3817365269461077, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1851, "step": 1591 }, { "epoch": 2.3832335329341316, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1846, "step": 1592 }, { "epoch": 2.3847305389221556, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1839, "step": 1593 }, { "epoch": 2.3862275449101795, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.182, "step": 1594 }, { "epoch": 2.3877245508982035, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1845, "step": 1595 }, { "epoch": 2.3892215568862274, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1937, "step": 1596 }, { "epoch": 2.3907185628742513, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1867, "step": 1597 }, { "epoch": 2.3922155688622753, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1821, "step": 1598 }, { "epoch": 2.3937125748502996, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1908, "step": 1599 }, { "epoch": 2.3952095808383236, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1937, "step": 1600 }, { "epoch": 2.3967065868263475, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1865, "step": 1601 }, { "epoch": 2.3982035928143715, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1889, "step": 1602 }, { "epoch": 2.3997005988023954, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1921, "step": 1603 }, { "epoch": 2.4011976047904193, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1867, "step": 1604 }, { "epoch": 2.4026946107784433, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1819, "step": 1605 }, { "epoch": 2.404191616766467, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1875, "step": 1606 }, { "epoch": 2.405688622754491, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1847, "step": 1607 }, { "epoch": 2.407185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1846, "step": 1608 }, { "epoch": 2.408682634730539, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1948, "step": 1609 }, { "epoch": 2.410179640718563, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.186, "step": 1610 }, { "epoch": 2.411676646706587, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1886, "step": 1611 }, { "epoch": 2.413173652694611, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1887, "step": 1612 }, { "epoch": 2.4146706586826348, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1831, "step": 1613 }, { "epoch": 2.4161676646706587, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.183, "step": 1614 }, { "epoch": 2.4176646706586826, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1863, "step": 1615 }, { "epoch": 2.4191616766467066, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.187, "step": 1616 }, { "epoch": 2.4206586826347305, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1896, "step": 1617 }, { "epoch": 2.4221556886227544, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1889, "step": 1618 }, { "epoch": 2.4236526946107784, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1882, "step": 1619 }, { "epoch": 2.4251497005988023, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1835, "step": 1620 }, { "epoch": 2.4266467065868262, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1882, "step": 1621 }, { "epoch": 2.42814371257485, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1873, "step": 1622 }, { "epoch": 2.429640718562874, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1948, "step": 1623 }, { "epoch": 2.431137724550898, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1879, "step": 1624 }, { "epoch": 2.432634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1909, "step": 1625 }, { "epoch": 2.434131736526946, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.178, "step": 1626 }, { "epoch": 2.43562874251497, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1953, "step": 1627 }, { "epoch": 2.437125748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1878, "step": 1628 }, { "epoch": 2.4386227544910177, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1837, "step": 1629 }, { "epoch": 2.440119760479042, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1779, "step": 1630 }, { "epoch": 2.441616766467066, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1887, "step": 1631 }, { "epoch": 2.44311377245509, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1849, "step": 1632 }, { "epoch": 2.444610778443114, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1887, "step": 1633 }, { "epoch": 2.446107784431138, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1871, "step": 1634 }, { "epoch": 2.447604790419162, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1879, "step": 1635 }, { "epoch": 2.4491017964071857, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1636 }, { "epoch": 2.4505988023952097, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1824, "step": 1637 }, { "epoch": 2.4520958083832336, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1778, "step": 1638 }, { "epoch": 2.4535928143712575, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1872, "step": 1639 }, { "epoch": 2.4550898203592815, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1913, "step": 1640 }, { "epoch": 2.4565868263473054, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1838, "step": 1641 }, { "epoch": 2.4580838323353293, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1818, "step": 1642 }, { "epoch": 2.4595808383233533, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1771, "step": 1643 }, { "epoch": 2.461077844311377, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1915, "step": 1644 }, { "epoch": 2.462574850299401, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1867, "step": 1645 }, { "epoch": 2.464071856287425, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1905, "step": 1646 }, { "epoch": 2.465568862275449, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1843, "step": 1647 }, { "epoch": 2.467065868263473, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1892, "step": 1648 }, { "epoch": 2.468562874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.185, "step": 1649 }, { "epoch": 2.470059880239521, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1863, "step": 1650 }, { "epoch": 2.4715568862275448, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1839, "step": 1651 }, { "epoch": 2.4730538922155687, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1917, "step": 1652 }, { "epoch": 2.4745508982035926, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1821, "step": 1653 }, { "epoch": 2.4760479041916166, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1858, "step": 1654 }, { "epoch": 2.477544910179641, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1864, "step": 1655 }, { "epoch": 2.479041916167665, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1803, "step": 1656 }, { "epoch": 2.480538922155689, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1925, "step": 1657 }, { "epoch": 2.4820359281437128, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1813, "step": 1658 }, { "epoch": 2.4835329341317367, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1904, "step": 1659 }, { "epoch": 2.4850299401197606, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1896, "step": 1660 }, { "epoch": 2.4865269461077846, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.188, "step": 1661 }, { "epoch": 2.4880239520958085, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1813, "step": 1662 }, { "epoch": 2.4895209580838324, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1888, "step": 1663 }, { "epoch": 2.4910179640718564, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1751, "step": 1664 }, { "epoch": 2.4925149700598803, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1868, "step": 1665 }, { "epoch": 2.4940119760479043, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1874, "step": 1666 }, { "epoch": 2.495508982035928, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1818, "step": 1667 }, { "epoch": 2.497005988023952, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1792, "step": 1668 }, { "epoch": 2.498502994011976, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1833, "step": 1669 }, { "epoch": 2.5, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1899, "step": 1670 }, { "epoch": 2.501497005988024, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1855, "step": 1671 }, { "epoch": 2.502994011976048, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1812, "step": 1672 }, { "epoch": 2.504491017964072, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1922, "step": 1673 }, { "epoch": 2.5059880239520957, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.187, "step": 1674 }, { "epoch": 2.5074850299401197, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1827, "step": 1675 }, { "epoch": 2.5089820359281436, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1886, "step": 1676 }, { "epoch": 2.5104790419161676, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1913, "step": 1677 }, { "epoch": 2.5119760479041915, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1848, "step": 1678 }, { "epoch": 2.5134730538922154, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1834, "step": 1679 }, { "epoch": 2.5149700598802394, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1853, "step": 1680 }, { "epoch": 2.5164670658682633, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.187, "step": 1681 }, { "epoch": 2.5179640718562872, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.177, "step": 1682 }, { "epoch": 2.519461077844311, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1822, "step": 1683 }, { "epoch": 2.520958083832335, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1879, "step": 1684 }, { "epoch": 2.522455089820359, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1877, "step": 1685 }, { "epoch": 2.523952095808383, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1873, "step": 1686 }, { "epoch": 2.525449101796407, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1849, "step": 1687 }, { "epoch": 2.5269461077844313, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1831, "step": 1688 }, { "epoch": 2.5284431137724552, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1909, "step": 1689 }, { "epoch": 2.529940119760479, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.18, "step": 1690 }, { "epoch": 2.531437125748503, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1869, "step": 1691 }, { "epoch": 2.532934131736527, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.19, "step": 1692 }, { "epoch": 2.534431137724551, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1832, "step": 1693 }, { "epoch": 2.535928143712575, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1862, "step": 1694 }, { "epoch": 2.537425149700599, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1863, "step": 1695 }, { "epoch": 2.538922155688623, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.192, "step": 1696 }, { "epoch": 2.5404191616766467, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1827, "step": 1697 }, { "epoch": 2.5419161676646707, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1887, "step": 1698 }, { "epoch": 2.5434131736526946, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1739, "step": 1699 }, { "epoch": 2.5449101796407185, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1846, "step": 1700 }, { "epoch": 2.5464071856287425, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1784, "step": 1701 }, { "epoch": 2.5479041916167664, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.187, "step": 1702 }, { "epoch": 2.5494011976047903, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1772, "step": 1703 }, { "epoch": 2.5508982035928143, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1907, "step": 1704 }, { "epoch": 2.552395209580838, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1876, "step": 1705 }, { "epoch": 2.553892215568862, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1839, "step": 1706 }, { "epoch": 2.555389221556886, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1796, "step": 1707 }, { "epoch": 2.55688622754491, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1888, "step": 1708 }, { "epoch": 2.5583832335329344, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1837, "step": 1709 }, { "epoch": 2.5598802395209583, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1887, "step": 1710 }, { "epoch": 2.5613772455089823, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1904, "step": 1711 }, { "epoch": 2.562874251497006, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1858, "step": 1712 }, { "epoch": 2.56437125748503, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1877, "step": 1713 }, { "epoch": 2.565868263473054, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1824, "step": 1714 }, { "epoch": 2.567365269461078, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1892, "step": 1715 }, { "epoch": 2.568862275449102, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1901, "step": 1716 }, { "epoch": 2.570359281437126, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1904, "step": 1717 }, { "epoch": 2.57185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1868, "step": 1718 }, { "epoch": 2.5733532934131738, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1862, "step": 1719 }, { "epoch": 2.5748502994011977, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1782, "step": 1720 }, { "epoch": 2.5763473053892216, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1721 }, { "epoch": 2.5778443113772456, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.186, "step": 1722 }, { "epoch": 2.5793413173652695, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1846, "step": 1723 }, { "epoch": 2.5808383233532934, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1858, "step": 1724 }, { "epoch": 2.5823353293413174, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1868, "step": 1725 }, { "epoch": 2.5838323353293413, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1804, "step": 1726 }, { "epoch": 2.5853293413173652, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.182, "step": 1727 }, { "epoch": 2.586826347305389, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1889, "step": 1728 }, { "epoch": 2.588323353293413, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1837, "step": 1729 }, { "epoch": 2.589820359281437, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1867, "step": 1730 }, { "epoch": 2.591317365269461, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1822, "step": 1731 }, { "epoch": 2.592814371257485, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.181, "step": 1732 }, { "epoch": 2.594311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1816, "step": 1733 }, { "epoch": 2.595808383233533, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.191, "step": 1734 }, { "epoch": 2.5973053892215567, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1893, "step": 1735 }, { "epoch": 2.5988023952095807, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1843, "step": 1736 }, { "epoch": 2.6002994011976046, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1866, "step": 1737 }, { "epoch": 2.6017964071856285, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1834, "step": 1738 }, { "epoch": 2.6032934131736525, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1893, "step": 1739 }, { "epoch": 2.6047904191616764, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1811, "step": 1740 }, { "epoch": 2.6062874251497004, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.189, "step": 1741 }, { "epoch": 2.6077844311377243, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1893, "step": 1742 }, { "epoch": 2.6092814371257482, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1838, "step": 1743 }, { "epoch": 2.6107784431137726, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.181, "step": 1744 }, { "epoch": 2.6122754491017965, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1785, "step": 1745 }, { "epoch": 2.6137724550898205, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1845, "step": 1746 }, { "epoch": 2.6152694610778444, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1825, "step": 1747 }, { "epoch": 2.6167664670658684, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1812, "step": 1748 }, { "epoch": 2.6182634730538923, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1811, "step": 1749 }, { "epoch": 2.6197604790419162, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1755, "step": 1750 }, { "epoch": 2.62125748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1808, "step": 1751 }, { "epoch": 2.622754491017964, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1755, "step": 1752 }, { "epoch": 2.624251497005988, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1854, "step": 1753 }, { "epoch": 2.625748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1842, "step": 1754 }, { "epoch": 2.627245508982036, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.187, "step": 1755 }, { "epoch": 2.62874251497006, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1825, "step": 1756 }, { "epoch": 2.6302395209580838, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1924, "step": 1757 }, { "epoch": 2.6317365269461077, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1828, "step": 1758 }, { "epoch": 2.6332335329341316, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1905, "step": 1759 }, { "epoch": 2.6347305389221556, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1835, "step": 1760 }, { "epoch": 2.6362275449101795, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1801, "step": 1761 }, { "epoch": 2.6377245508982035, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1871, "step": 1762 }, { "epoch": 2.6392215568862274, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1849, "step": 1763 }, { "epoch": 2.6407185628742518, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1822, "step": 1764 }, { "epoch": 2.6422155688622757, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1874, "step": 1765 }, { "epoch": 2.6437125748502996, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1794, "step": 1766 }, { "epoch": 2.6452095808383236, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1843, "step": 1767 }, { "epoch": 2.6467065868263475, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1862, "step": 1768 }, { "epoch": 2.6482035928143715, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1827, "step": 1769 }, { "epoch": 2.6497005988023954, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.184, "step": 1770 }, { "epoch": 2.6511976047904193, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1893, "step": 1771 }, { "epoch": 2.6526946107784433, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1823, "step": 1772 }, { "epoch": 2.654191616766467, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1814, "step": 1773 }, { "epoch": 2.655688622754491, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.185, "step": 1774 }, { "epoch": 2.657185628742515, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1882, "step": 1775 }, { "epoch": 2.658682634730539, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1881, "step": 1776 }, { "epoch": 2.660179640718563, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1912, "step": 1777 }, { "epoch": 2.661676646706587, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1795, "step": 1778 }, { "epoch": 2.663173652694611, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1891, "step": 1779 }, { "epoch": 2.6646706586826348, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1780 }, { "epoch": 2.6661676646706587, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1862, "step": 1781 }, { "epoch": 2.6676646706586826, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1875, "step": 1782 }, { "epoch": 2.6691616766467066, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1809, "step": 1783 }, { "epoch": 2.6706586826347305, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1803, "step": 1784 }, { "epoch": 2.6721556886227544, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1839, "step": 1785 }, { "epoch": 2.6736526946107784, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1915, "step": 1786 }, { "epoch": 2.6751497005988023, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1808, "step": 1787 }, { "epoch": 2.6766467065868262, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1849, "step": 1788 }, { "epoch": 2.67814371257485, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1855, "step": 1789 }, { "epoch": 2.679640718562874, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1898, "step": 1790 }, { "epoch": 2.681137724550898, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1809, "step": 1791 }, { "epoch": 2.682634730538922, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1851, "step": 1792 }, { "epoch": 2.684131736526946, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1938, "step": 1793 }, { "epoch": 2.68562874251497, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1843, "step": 1794 }, { "epoch": 2.687125748502994, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1805, "step": 1795 }, { "epoch": 2.6886227544910177, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1851, "step": 1796 }, { "epoch": 2.6901197604790417, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1797, "step": 1797 }, { "epoch": 2.6916167664670656, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1814, "step": 1798 }, { "epoch": 2.69311377245509, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1902, "step": 1799 }, { "epoch": 2.694610778443114, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.179, "step": 1800 }, { "epoch": 2.696107784431138, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1783, "step": 1801 }, { "epoch": 2.697604790419162, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1787, "step": 1802 }, { "epoch": 2.6991017964071857, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1894, "step": 1803 }, { "epoch": 2.7005988023952097, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1855, "step": 1804 }, { "epoch": 2.7020958083832336, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1845, "step": 1805 }, { "epoch": 2.7035928143712575, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1827, "step": 1806 }, { "epoch": 2.7050898203592815, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1903, "step": 1807 }, { "epoch": 2.7065868263473054, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1753, "step": 1808 }, { "epoch": 2.7080838323353293, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.187, "step": 1809 }, { "epoch": 2.7095808383233533, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1876, "step": 1810 }, { "epoch": 2.711077844311377, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1766, "step": 1811 }, { "epoch": 2.712574850299401, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1816, "step": 1812 }, { "epoch": 2.714071856287425, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1825, "step": 1813 }, { "epoch": 2.715568862275449, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1904, "step": 1814 }, { "epoch": 2.717065868263473, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1805, "step": 1815 }, { "epoch": 2.718562874251497, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1896, "step": 1816 }, { "epoch": 2.720059880239521, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1796, "step": 1817 }, { "epoch": 2.7215568862275448, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1883, "step": 1818 }, { "epoch": 2.7230538922155687, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1784, "step": 1819 }, { "epoch": 2.724550898203593, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1865, "step": 1820 }, { "epoch": 2.726047904191617, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1855, "step": 1821 }, { "epoch": 2.727544910179641, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.185, "step": 1822 }, { "epoch": 2.729041916167665, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.185, "step": 1823 }, { "epoch": 2.730538922155689, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1836, "step": 1824 }, { "epoch": 2.7320359281437128, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1859, "step": 1825 }, { "epoch": 2.7335329341317367, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1846, "step": 1826 }, { "epoch": 2.7350299401197606, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1845, "step": 1827 }, { "epoch": 2.7365269461077846, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1817, "step": 1828 }, { "epoch": 2.7380239520958085, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1821, "step": 1829 }, { "epoch": 2.7395209580838324, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.183, "step": 1830 }, { "epoch": 2.7410179640718564, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1794, "step": 1831 }, { "epoch": 2.7425149700598803, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1798, "step": 1832 }, { "epoch": 2.7440119760479043, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1797, "step": 1833 }, { "epoch": 2.745508982035928, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1787, "step": 1834 }, { "epoch": 2.747005988023952, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1872, "step": 1835 }, { "epoch": 2.748502994011976, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1893, "step": 1836 }, { "epoch": 2.75, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1863, "step": 1837 }, { "epoch": 2.751497005988024, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1857, "step": 1838 }, { "epoch": 2.752994011976048, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1812, "step": 1839 }, { "epoch": 2.754491017964072, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1798, "step": 1840 }, { "epoch": 2.7559880239520957, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1804, "step": 1841 }, { "epoch": 2.7574850299401197, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1831, "step": 1842 }, { "epoch": 2.7589820359281436, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1801, "step": 1843 }, { "epoch": 2.7604790419161676, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1858, "step": 1844 }, { "epoch": 2.7619760479041915, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1807, "step": 1845 }, { "epoch": 2.7634730538922154, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1874, "step": 1846 }, { "epoch": 2.7649700598802394, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1818, "step": 1847 }, { "epoch": 2.7664670658682633, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1781, "step": 1848 }, { "epoch": 2.7679640718562872, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1867, "step": 1849 }, { "epoch": 2.769461077844311, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1813, "step": 1850 }, { "epoch": 2.770958083832335, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1781, "step": 1851 }, { "epoch": 2.772455089820359, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1795, "step": 1852 }, { "epoch": 2.773952095808383, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1887, "step": 1853 }, { "epoch": 2.775449101796407, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.18, "step": 1854 }, { "epoch": 2.7769461077844313, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1788, "step": 1855 }, { "epoch": 2.7784431137724552, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1769, "step": 1856 }, { "epoch": 2.779940119760479, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1805, "step": 1857 }, { "epoch": 2.781437125748503, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1799, "step": 1858 }, { "epoch": 2.782934131736527, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1787, "step": 1859 }, { "epoch": 2.784431137724551, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1841, "step": 1860 }, { "epoch": 2.785928143712575, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1733, "step": 1861 }, { "epoch": 2.787425149700599, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1865, "step": 1862 }, { "epoch": 2.788922155688623, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1776, "step": 1863 }, { "epoch": 2.7904191616766467, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1818, "step": 1864 }, { "epoch": 2.7919161676646707, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1833, "step": 1865 }, { "epoch": 2.7934131736526946, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1837, "step": 1866 }, { "epoch": 2.7949101796407185, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1839, "step": 1867 }, { "epoch": 2.7964071856287425, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1821, "step": 1868 }, { "epoch": 2.7979041916167664, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1764, "step": 1869 }, { "epoch": 2.7994011976047903, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1805, "step": 1870 }, { "epoch": 2.8008982035928143, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1782, "step": 1871 }, { "epoch": 2.802395209580838, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1757, "step": 1872 }, { "epoch": 2.803892215568862, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.19, "step": 1873 }, { "epoch": 2.805389221556886, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1822, "step": 1874 }, { "epoch": 2.80688622754491, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1875 }, { "epoch": 2.8083832335329344, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1769, "step": 1876 }, { "epoch": 2.8098802395209583, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1805, "step": 1877 }, { "epoch": 2.8113772455089823, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1895, "step": 1878 }, { "epoch": 2.812874251497006, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1847, "step": 1879 }, { "epoch": 2.81437125748503, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1880 }, { "epoch": 2.815868263473054, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.182, "step": 1881 }, { "epoch": 2.817365269461078, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1909, "step": 1882 }, { "epoch": 2.818862275449102, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1779, "step": 1883 }, { "epoch": 2.820359281437126, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1803, "step": 1884 }, { "epoch": 2.82185628742515, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1854, "step": 1885 }, { "epoch": 2.8233532934131738, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.1795, "step": 1886 }, { "epoch": 2.8248502994011977, "grad_norm": 0.201171875, "learning_rate": 0.0008, "loss": 1.1766, "step": 1887 }, { "epoch": 2.8263473053892216, "grad_norm": 0.322265625, "learning_rate": 0.0008, "loss": 1.1869, "step": 1888 }, { "epoch": 2.8278443113772456, "grad_norm": 0.71875, "learning_rate": 0.0008, "loss": 1.1896, "step": 1889 }, { "epoch": 2.8293413173652695, "grad_norm": 1.6328125, "learning_rate": 0.0008, "loss": 1.2451, "step": 1890 }, { "epoch": 2.8308383233532934, "grad_norm": 0.31640625, "learning_rate": 0.0008, "loss": 1.1999, "step": 1891 }, { "epoch": 2.8323353293413174, "grad_norm": 0.70703125, "learning_rate": 0.0008, "loss": 1.2046, "step": 1892 }, { "epoch": 2.8338323353293413, "grad_norm": 0.6015625, "learning_rate": 0.0008, "loss": 1.2139, "step": 1893 }, { "epoch": 2.8353293413173652, "grad_norm": 0.390625, "learning_rate": 0.0008, "loss": 1.2034, "step": 1894 }, { "epoch": 2.836826347305389, "grad_norm": 0.80078125, "learning_rate": 0.0008, "loss": 1.2113, "step": 1895 }, { "epoch": 2.838323353293413, "grad_norm": 0.6484375, "learning_rate": 0.0008, "loss": 1.2191, "step": 1896 }, { "epoch": 2.839820359281437, "grad_norm": 0.494140625, "learning_rate": 0.0008, "loss": 1.2157, "step": 1897 }, { "epoch": 2.841317365269461, "grad_norm": 0.44140625, "learning_rate": 0.0008, "loss": 1.2156, "step": 1898 }, { "epoch": 2.842814371257485, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.2043, "step": 1899 }, { "epoch": 2.844311377245509, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.2027, "step": 1900 }, { "epoch": 2.845808383233533, "grad_norm": 0.2255859375, "learning_rate": 0.0008, "loss": 1.2037, "step": 1901 }, { "epoch": 2.8473053892215567, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.1945, "step": 1902 }, { "epoch": 2.8488023952095807, "grad_norm": 0.171875, "learning_rate": 0.0008, "loss": 1.1947, "step": 1903 }, { "epoch": 2.8502994011976046, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1927, "step": 1904 }, { "epoch": 2.8517964071856285, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1911, "step": 1905 }, { "epoch": 2.8532934131736525, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1888, "step": 1906 }, { "epoch": 2.8547904191616764, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1883, "step": 1907 }, { "epoch": 2.8562874251497004, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1916, "step": 1908 }, { "epoch": 2.8577844311377243, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1928, "step": 1909 }, { "epoch": 2.8592814371257482, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1925, "step": 1910 }, { "epoch": 2.8607784431137726, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1893, "step": 1911 }, { "epoch": 2.8622754491017965, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1854, "step": 1912 }, { "epoch": 2.8637724550898205, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1877, "step": 1913 }, { "epoch": 2.8652694610778444, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1842, "step": 1914 }, { "epoch": 2.8667664670658684, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1872, "step": 1915 }, { "epoch": 2.8682634730538923, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1879, "step": 1916 }, { "epoch": 2.8697604790419162, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1901, "step": 1917 }, { "epoch": 2.87125748502994, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1848, "step": 1918 }, { "epoch": 2.872754491017964, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1843, "step": 1919 }, { "epoch": 2.874251497005988, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1872, "step": 1920 }, { "epoch": 2.875748502994012, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1794, "step": 1921 }, { "epoch": 2.877245508982036, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1844, "step": 1922 }, { "epoch": 2.87874251497006, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1907, "step": 1923 }, { "epoch": 2.8802395209580838, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1878, "step": 1924 }, { "epoch": 2.8817365269461077, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1843, "step": 1925 }, { "epoch": 2.8832335329341316, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.178, "step": 1926 }, { "epoch": 2.8847305389221556, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1822, "step": 1927 }, { "epoch": 2.8862275449101795, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1834, "step": 1928 }, { "epoch": 2.8877245508982035, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.186, "step": 1929 }, { "epoch": 2.8892215568862274, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1909, "step": 1930 }, { "epoch": 2.8907185628742518, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1883, "step": 1931 }, { "epoch": 2.8922155688622757, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1848, "step": 1932 }, { "epoch": 2.8937125748502996, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.184, "step": 1933 }, { "epoch": 2.8952095808383236, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1867, "step": 1934 }, { "epoch": 2.8967065868263475, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1754, "step": 1935 }, { "epoch": 2.8982035928143715, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1803, "step": 1936 }, { "epoch": 2.8997005988023954, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1808, "step": 1937 }, { "epoch": 2.9011976047904193, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1799, "step": 1938 }, { "epoch": 2.9026946107784433, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.181, "step": 1939 }, { "epoch": 2.904191616766467, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1814, "step": 1940 }, { "epoch": 2.905688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1832, "step": 1941 }, { "epoch": 2.907185628742515, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1739, "step": 1942 }, { "epoch": 2.908682634730539, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1809, "step": 1943 }, { "epoch": 2.910179640718563, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1867, "step": 1944 }, { "epoch": 2.911676646706587, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1907, "step": 1945 }, { "epoch": 2.913173652694611, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1828, "step": 1946 }, { "epoch": 2.9146706586826348, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1797, "step": 1947 }, { "epoch": 2.9161676646706587, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1848, "step": 1948 }, { "epoch": 2.9176646706586826, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1761, "step": 1949 }, { "epoch": 2.9191616766467066, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1902, "step": 1950 }, { "epoch": 2.9206586826347305, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1825, "step": 1951 }, { "epoch": 2.9221556886227544, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1784, "step": 1952 }, { "epoch": 2.9236526946107784, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1792, "step": 1953 }, { "epoch": 2.9251497005988023, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.175, "step": 1954 }, { "epoch": 2.9266467065868262, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1842, "step": 1955 }, { "epoch": 2.92814371257485, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1811, "step": 1956 }, { "epoch": 2.929640718562874, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1818, "step": 1957 }, { "epoch": 2.931137724550898, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1782, "step": 1958 }, { "epoch": 2.932634730538922, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1768, "step": 1959 }, { "epoch": 2.934131736526946, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1846, "step": 1960 }, { "epoch": 2.93562874251497, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1858, "step": 1961 }, { "epoch": 2.937125748502994, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1895, "step": 1962 }, { "epoch": 2.9386227544910177, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1838, "step": 1963 }, { "epoch": 2.9401197604790417, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1748, "step": 1964 }, { "epoch": 2.9416167664670656, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1831, "step": 1965 }, { "epoch": 2.94311377245509, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1788, "step": 1966 }, { "epoch": 2.944610778443114, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1804, "step": 1967 }, { "epoch": 2.946107784431138, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1756, "step": 1968 }, { "epoch": 2.947604790419162, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1852, "step": 1969 }, { "epoch": 2.9491017964071857, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1794, "step": 1970 }, { "epoch": 2.9505988023952097, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1787, "step": 1971 }, { "epoch": 2.9520958083832336, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1844, "step": 1972 }, { "epoch": 2.9535928143712575, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.177, "step": 1973 }, { "epoch": 2.9550898203592815, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1811, "step": 1974 }, { "epoch": 2.9565868263473054, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1809, "step": 1975 }, { "epoch": 2.9580838323353293, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.183, "step": 1976 }, { "epoch": 2.9595808383233533, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1787, "step": 1977 }, { "epoch": 2.961077844311377, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1835, "step": 1978 }, { "epoch": 2.962574850299401, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1846, "step": 1979 }, { "epoch": 2.964071856287425, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1834, "step": 1980 }, { "epoch": 2.965568862275449, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1786, "step": 1981 }, { "epoch": 2.967065868263473, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1827, "step": 1982 }, { "epoch": 2.968562874251497, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1846, "step": 1983 }, { "epoch": 2.970059880239521, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1791, "step": 1984 }, { "epoch": 2.9715568862275448, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.179, "step": 1985 }, { "epoch": 2.9730538922155687, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1833, "step": 1986 }, { "epoch": 2.974550898203593, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1856, "step": 1987 }, { "epoch": 2.976047904191617, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1822, "step": 1988 }, { "epoch": 2.977544910179641, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1882, "step": 1989 }, { "epoch": 2.979041916167665, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.184, "step": 1990 }, { "epoch": 2.980538922155689, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1791, "step": 1991 }, { "epoch": 2.9820359281437128, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1781, "step": 1992 }, { "epoch": 2.9835329341317367, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1777, "step": 1993 }, { "epoch": 2.9850299401197606, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1788, "step": 1994 }, { "epoch": 2.9865269461077846, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.179, "step": 1995 }, { "epoch": 2.9880239520958085, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1746, "step": 1996 }, { "epoch": 2.9895209580838324, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1774, "step": 1997 }, { "epoch": 2.9910179640718564, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1793, "step": 1998 }, { "epoch": 2.9925149700598803, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1913, "step": 1999 }, { "epoch": 2.9940119760479043, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1788, "step": 2000 }, { "epoch": 2.995508982035928, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1812, "step": 2001 }, { "epoch": 2.997005988023952, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1774, "step": 2002 }, { "epoch": 2.998502994011976, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1888, "step": 2003 }, { "epoch": 3.0, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1848, "step": 2004 }, { "epoch": 3.001497005988024, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1775, "step": 2005 }, { "epoch": 3.002994011976048, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1796, "step": 2006 }, { "epoch": 3.004491017964072, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1843, "step": 2007 }, { "epoch": 3.0059880239520957, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1879, "step": 2008 }, { "epoch": 3.0074850299401197, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1827, "step": 2009 }, { "epoch": 3.0089820359281436, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1779, "step": 2010 }, { "epoch": 3.0104790419161676, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1726, "step": 2011 }, { "epoch": 3.0119760479041915, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1801, "step": 2012 }, { "epoch": 3.0134730538922154, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1811, "step": 2013 }, { "epoch": 3.0149700598802394, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1767, "step": 2014 }, { "epoch": 3.0164670658682633, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1843, "step": 2015 }, { "epoch": 3.0179640718562872, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1747, "step": 2016 }, { "epoch": 3.019461077844311, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1788, "step": 2017 }, { "epoch": 3.020958083832335, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1785, "step": 2018 }, { "epoch": 3.0224550898203595, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.171, "step": 2019 }, { "epoch": 3.0239520958083834, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1845, "step": 2020 }, { "epoch": 3.0254491017964074, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1837, "step": 2021 }, { "epoch": 3.0269461077844313, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1806, "step": 2022 }, { "epoch": 3.0284431137724552, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1809, "step": 2023 }, { "epoch": 3.029940119760479, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1848, "step": 2024 }, { "epoch": 3.031437125748503, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1782, "step": 2025 }, { "epoch": 3.032934131736527, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1786, "step": 2026 }, { "epoch": 3.034431137724551, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1764, "step": 2027 }, { "epoch": 3.035928143712575, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1857, "step": 2028 }, { "epoch": 3.037425149700599, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1756, "step": 2029 }, { "epoch": 3.038922155688623, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1794, "step": 2030 }, { "epoch": 3.0404191616766467, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.177, "step": 2031 }, { "epoch": 3.0419161676646707, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1807, "step": 2032 }, { "epoch": 3.0434131736526946, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1807, "step": 2033 }, { "epoch": 3.0449101796407185, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1773, "step": 2034 }, { "epoch": 3.0464071856287425, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1882, "step": 2035 }, { "epoch": 3.0479041916167664, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1763, "step": 2036 }, { "epoch": 3.0494011976047903, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1784, "step": 2037 }, { "epoch": 3.0508982035928143, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1741, "step": 2038 }, { "epoch": 3.052395209580838, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1798, "step": 2039 }, { "epoch": 3.053892215568862, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1907, "step": 2040 }, { "epoch": 3.055389221556886, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1814, "step": 2041 }, { "epoch": 3.05688622754491, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1848, "step": 2042 }, { "epoch": 3.058383233532934, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1797, "step": 2043 }, { "epoch": 3.059880239520958, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1734, "step": 2044 }, { "epoch": 3.061377245508982, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1807, "step": 2045 }, { "epoch": 3.062874251497006, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1824, "step": 2046 }, { "epoch": 3.06437125748503, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1758, "step": 2047 }, { "epoch": 3.065868263473054, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1851, "step": 2048 }, { "epoch": 3.067365269461078, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1798, "step": 2049 }, { "epoch": 3.068862275449102, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1795, "step": 2050 }, { "epoch": 3.070359281437126, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1779, "step": 2051 }, { "epoch": 3.07185628742515, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1715, "step": 2052 }, { "epoch": 3.0733532934131738, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1796, "step": 2053 }, { "epoch": 3.0748502994011977, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1826, "step": 2054 }, { "epoch": 3.0763473053892216, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1817, "step": 2055 }, { "epoch": 3.0778443113772456, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1742, "step": 2056 }, { "epoch": 3.0793413173652695, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1801, "step": 2057 }, { "epoch": 3.0808383233532934, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1813, "step": 2058 }, { "epoch": 3.0823353293413174, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1811, "step": 2059 }, { "epoch": 3.0838323353293413, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1775, "step": 2060 }, { "epoch": 3.0853293413173652, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1868, "step": 2061 }, { "epoch": 3.086826347305389, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1824, "step": 2062 }, { "epoch": 3.088323353293413, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1796, "step": 2063 }, { "epoch": 3.089820359281437, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1829, "step": 2064 }, { "epoch": 3.091317365269461, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1758, "step": 2065 }, { "epoch": 3.092814371257485, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1802, "step": 2066 }, { "epoch": 3.094311377245509, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1823, "step": 2067 }, { "epoch": 3.095808383233533, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1808, "step": 2068 }, { "epoch": 3.0973053892215567, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1794, "step": 2069 }, { "epoch": 3.0988023952095807, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1798, "step": 2070 }, { "epoch": 3.1002994011976046, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1805, "step": 2071 }, { "epoch": 3.1017964071856285, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1762, "step": 2072 }, { "epoch": 3.1032934131736525, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1714, "step": 2073 }, { "epoch": 3.1047904191616764, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1789, "step": 2074 }, { "epoch": 3.106287425149701, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1812, "step": 2075 }, { "epoch": 3.1077844311377247, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1792, "step": 2076 }, { "epoch": 3.1092814371257487, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.182, "step": 2077 }, { "epoch": 3.1107784431137726, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1872, "step": 2078 }, { "epoch": 3.1122754491017965, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1803, "step": 2079 }, { "epoch": 3.1137724550898205, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1795, "step": 2080 }, { "epoch": 3.1152694610778444, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1806, "step": 2081 }, { "epoch": 3.1167664670658684, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1792, "step": 2082 }, { "epoch": 3.1182634730538923, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1825, "step": 2083 }, { "epoch": 3.1197604790419162, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1732, "step": 2084 }, { "epoch": 3.12125748502994, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.182, "step": 2085 }, { "epoch": 3.122754491017964, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1848, "step": 2086 }, { "epoch": 3.124251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1771, "step": 2087 }, { "epoch": 3.125748502994012, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1805, "step": 2088 }, { "epoch": 3.127245508982036, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1764, "step": 2089 }, { "epoch": 3.12874251497006, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1797, "step": 2090 }, { "epoch": 3.1302395209580838, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1759, "step": 2091 }, { "epoch": 3.1317365269461077, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1763, "step": 2092 }, { "epoch": 3.1332335329341316, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1781, "step": 2093 }, { "epoch": 3.1347305389221556, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1765, "step": 2094 }, { "epoch": 3.1362275449101795, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1793, "step": 2095 }, { "epoch": 3.1377245508982035, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1753, "step": 2096 }, { "epoch": 3.1392215568862274, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1779, "step": 2097 }, { "epoch": 3.1407185628742513, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1788, "step": 2098 }, { "epoch": 3.1422155688622753, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1756, "step": 2099 }, { "epoch": 3.143712574850299, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1834, "step": 2100 }, { "epoch": 3.1452095808383236, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1811, "step": 2101 }, { "epoch": 3.1467065868263475, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1784, "step": 2102 }, { "epoch": 3.1482035928143715, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1851, "step": 2103 }, { "epoch": 3.1497005988023954, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1792, "step": 2104 }, { "epoch": 3.1511976047904193, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1804, "step": 2105 }, { "epoch": 3.1526946107784433, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1754, "step": 2106 }, { "epoch": 3.154191616766467, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1767, "step": 2107 }, { "epoch": 3.155688622754491, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.176, "step": 2108 }, { "epoch": 3.157185628742515, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1814, "step": 2109 }, { "epoch": 3.158682634730539, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1724, "step": 2110 }, { "epoch": 3.160179640718563, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1801, "step": 2111 }, { "epoch": 3.161676646706587, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1784, "step": 2112 }, { "epoch": 3.163173652694611, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1851, "step": 2113 }, { "epoch": 3.1646706586826348, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1815, "step": 2114 }, { "epoch": 3.1661676646706587, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1734, "step": 2115 }, { "epoch": 3.1676646706586826, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1785, "step": 2116 }, { "epoch": 3.1691616766467066, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1772, "step": 2117 }, { "epoch": 3.1706586826347305, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1814, "step": 2118 }, { "epoch": 3.1721556886227544, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1816, "step": 2119 }, { "epoch": 3.1736526946107784, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1771, "step": 2120 }, { "epoch": 3.1751497005988023, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1777, "step": 2121 }, { "epoch": 3.1766467065868262, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1839, "step": 2122 }, { "epoch": 3.17814371257485, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1765, "step": 2123 }, { "epoch": 3.179640718562874, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1825, "step": 2124 }, { "epoch": 3.181137724550898, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1825, "step": 2125 }, { "epoch": 3.182634730538922, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1772, "step": 2126 }, { "epoch": 3.184131736526946, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1787, "step": 2127 }, { "epoch": 3.18562874251497, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1821, "step": 2128 }, { "epoch": 3.187125748502994, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1792, "step": 2129 }, { "epoch": 3.1886227544910177, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1773, "step": 2130 }, { "epoch": 3.190119760479042, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.178, "step": 2131 }, { "epoch": 3.191616766467066, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1744, "step": 2132 }, { "epoch": 3.19311377245509, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1752, "step": 2133 }, { "epoch": 3.194610778443114, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1762, "step": 2134 }, { "epoch": 3.196107784431138, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1749, "step": 2135 }, { "epoch": 3.197604790419162, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1742, "step": 2136 }, { "epoch": 3.1991017964071857, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1725, "step": 2137 }, { "epoch": 3.2005988023952097, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1854, "step": 2138 }, { "epoch": 3.2020958083832336, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1817, "step": 2139 }, { "epoch": 3.2035928143712575, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1702, "step": 2140 }, { "epoch": 3.2050898203592815, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1785, "step": 2141 }, { "epoch": 3.2065868263473054, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1799, "step": 2142 }, { "epoch": 3.2080838323353293, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1773, "step": 2143 }, { "epoch": 3.2095808383233533, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1805, "step": 2144 }, { "epoch": 3.211077844311377, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1805, "step": 2145 }, { "epoch": 3.212574850299401, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.183, "step": 2146 }, { "epoch": 3.214071856287425, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.181, "step": 2147 }, { "epoch": 3.215568862275449, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1748, "step": 2148 }, { "epoch": 3.217065868263473, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1759, "step": 2149 }, { "epoch": 3.218562874251497, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1768, "step": 2150 }, { "epoch": 3.220059880239521, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1803, "step": 2151 }, { "epoch": 3.2215568862275448, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1827, "step": 2152 }, { "epoch": 3.2230538922155687, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1733, "step": 2153 }, { "epoch": 3.2245508982035926, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1779, "step": 2154 }, { "epoch": 3.2260479041916166, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1827, "step": 2155 }, { "epoch": 3.2275449101796405, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1818, "step": 2156 }, { "epoch": 3.229041916167665, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.178, "step": 2157 }, { "epoch": 3.230538922155689, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1727, "step": 2158 }, { "epoch": 3.2320359281437128, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1732, "step": 2159 }, { "epoch": 3.2335329341317367, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1828, "step": 2160 }, { "epoch": 3.2350299401197606, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1781, "step": 2161 }, { "epoch": 3.2365269461077846, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1762, "step": 2162 }, { "epoch": 3.2380239520958085, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1736, "step": 2163 }, { "epoch": 3.2395209580838324, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1843, "step": 2164 }, { "epoch": 3.2410179640718564, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1785, "step": 2165 }, { "epoch": 3.2425149700598803, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1746, "step": 2166 }, { "epoch": 3.2440119760479043, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1779, "step": 2167 }, { "epoch": 3.245508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1755, "step": 2168 }, { "epoch": 3.247005988023952, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1766, "step": 2169 }, { "epoch": 3.248502994011976, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1796, "step": 2170 }, { "epoch": 3.25, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1746, "step": 2171 }, { "epoch": 3.251497005988024, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1758, "step": 2172 }, { "epoch": 3.252994011976048, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1778, "step": 2173 }, { "epoch": 3.254491017964072, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1766, "step": 2174 }, { "epoch": 3.2559880239520957, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1801, "step": 2175 }, { "epoch": 3.2574850299401197, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1742, "step": 2176 }, { "epoch": 3.2589820359281436, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1758, "step": 2177 }, { "epoch": 3.2604790419161676, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1743, "step": 2178 }, { "epoch": 3.2619760479041915, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1859, "step": 2179 }, { "epoch": 3.2634730538922154, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1767, "step": 2180 }, { "epoch": 3.2649700598802394, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1796, "step": 2181 }, { "epoch": 3.2664670658682633, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1756, "step": 2182 }, { "epoch": 3.2679640718562872, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1789, "step": 2183 }, { "epoch": 3.269461077844311, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1796, "step": 2184 }, { "epoch": 3.270958083832335, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1811, "step": 2185 }, { "epoch": 3.272455089820359, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1732, "step": 2186 }, { "epoch": 3.2739520958083834, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1761, "step": 2187 }, { "epoch": 3.2754491017964074, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1772, "step": 2188 }, { "epoch": 3.2769461077844313, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1784, "step": 2189 }, { "epoch": 3.2784431137724552, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1743, "step": 2190 }, { "epoch": 3.279940119760479, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1662, "step": 2191 }, { "epoch": 3.281437125748503, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1671, "step": 2192 }, { "epoch": 3.282934131736527, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1824, "step": 2193 }, { "epoch": 3.284431137724551, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1787, "step": 2194 }, { "epoch": 3.285928143712575, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1704, "step": 2195 }, { "epoch": 3.287425149700599, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1758, "step": 2196 }, { "epoch": 3.288922155688623, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.173, "step": 2197 }, { "epoch": 3.2904191616766467, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1809, "step": 2198 }, { "epoch": 3.2919161676646707, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1722, "step": 2199 }, { "epoch": 3.2934131736526946, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1825, "step": 2200 }, { "epoch": 3.2949101796407185, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1796, "step": 2201 }, { "epoch": 3.2964071856287425, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1778, "step": 2202 }, { "epoch": 3.2979041916167664, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1735, "step": 2203 }, { "epoch": 3.2994011976047903, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.179, "step": 2204 }, { "epoch": 3.3008982035928143, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1829, "step": 2205 }, { "epoch": 3.302395209580838, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1733, "step": 2206 }, { "epoch": 3.303892215568862, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.177, "step": 2207 }, { "epoch": 3.305389221556886, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1741, "step": 2208 }, { "epoch": 3.30688622754491, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1789, "step": 2209 }, { "epoch": 3.308383233532934, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1835, "step": 2210 }, { "epoch": 3.309880239520958, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1797, "step": 2211 }, { "epoch": 3.3113772455089823, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1758, "step": 2212 }, { "epoch": 3.312874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1719, "step": 2213 }, { "epoch": 3.31437125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1754, "step": 2214 }, { "epoch": 3.315868263473054, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1757, "step": 2215 }, { "epoch": 3.317365269461078, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1843, "step": 2216 }, { "epoch": 3.318862275449102, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1759, "step": 2217 }, { "epoch": 3.320359281437126, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1732, "step": 2218 }, { "epoch": 3.32185628742515, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1723, "step": 2219 }, { "epoch": 3.3233532934131738, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1794, "step": 2220 }, { "epoch": 3.3248502994011977, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1812, "step": 2221 }, { "epoch": 3.3263473053892216, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.18, "step": 2222 }, { "epoch": 3.3278443113772456, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1728, "step": 2223 }, { "epoch": 3.3293413173652695, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1739, "step": 2224 }, { "epoch": 3.3308383233532934, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1791, "step": 2225 }, { "epoch": 3.3323353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1732, "step": 2226 }, { "epoch": 3.3338323353293413, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1792, "step": 2227 }, { "epoch": 3.3353293413173652, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.175, "step": 2228 }, { "epoch": 3.336826347305389, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1789, "step": 2229 }, { "epoch": 3.338323353293413, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1707, "step": 2230 }, { "epoch": 3.339820359281437, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1766, "step": 2231 }, { "epoch": 3.341317365269461, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1738, "step": 2232 }, { "epoch": 3.342814371257485, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1719, "step": 2233 }, { "epoch": 3.344311377245509, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1797, "step": 2234 }, { "epoch": 3.345808383233533, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.178, "step": 2235 }, { "epoch": 3.3473053892215567, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1732, "step": 2236 }, { "epoch": 3.3488023952095807, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1717, "step": 2237 }, { "epoch": 3.3502994011976046, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1818, "step": 2238 }, { "epoch": 3.3517964071856285, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1789, "step": 2239 }, { "epoch": 3.3532934131736525, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1745, "step": 2240 }, { "epoch": 3.3547904191616764, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1743, "step": 2241 }, { "epoch": 3.3562874251497004, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1751, "step": 2242 }, { "epoch": 3.3577844311377247, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1774, "step": 2243 }, { "epoch": 3.3592814371257487, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1812, "step": 2244 }, { "epoch": 3.3607784431137726, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.172, "step": 2245 }, { "epoch": 3.3622754491017965, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1708, "step": 2246 }, { "epoch": 3.3637724550898205, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1784, "step": 2247 }, { "epoch": 3.3652694610778444, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1759, "step": 2248 }, { "epoch": 3.3667664670658684, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1759, "step": 2249 }, { "epoch": 3.3682634730538923, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.183, "step": 2250 }, { "epoch": 3.3697604790419162, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1802, "step": 2251 }, { "epoch": 3.37125748502994, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1726, "step": 2252 }, { "epoch": 3.372754491017964, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1756, "step": 2253 }, { "epoch": 3.374251497005988, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1774, "step": 2254 }, { "epoch": 3.375748502994012, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1728, "step": 2255 }, { "epoch": 3.377245508982036, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1756, "step": 2256 }, { "epoch": 3.37874251497006, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1718, "step": 2257 }, { "epoch": 3.3802395209580838, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1773, "step": 2258 }, { "epoch": 3.3817365269461077, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1834, "step": 2259 }, { "epoch": 3.3832335329341316, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1754, "step": 2260 }, { "epoch": 3.3847305389221556, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1675, "step": 2261 }, { "epoch": 3.3862275449101795, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.166, "step": 2262 }, { "epoch": 3.3877245508982035, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.1785, "step": 2263 }, { "epoch": 3.3892215568862274, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1747, "step": 2264 }, { "epoch": 3.3907185628742513, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1714, "step": 2265 }, { "epoch": 3.3922155688622753, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.181, "step": 2266 }, { "epoch": 3.3937125748502996, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1781, "step": 2267 }, { "epoch": 3.3952095808383236, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1779, "step": 2268 }, { "epoch": 3.3967065868263475, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1732, "step": 2269 }, { "epoch": 3.3982035928143715, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1754, "step": 2270 }, { "epoch": 3.3997005988023954, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1756, "step": 2271 }, { "epoch": 3.4011976047904193, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1793, "step": 2272 }, { "epoch": 3.4026946107784433, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1767, "step": 2273 }, { "epoch": 3.404191616766467, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1695, "step": 2274 }, { "epoch": 3.405688622754491, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1709, "step": 2275 }, { "epoch": 3.407185628742515, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1738, "step": 2276 }, { "epoch": 3.408682634730539, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1665, "step": 2277 }, { "epoch": 3.410179640718563, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1753, "step": 2278 }, { "epoch": 3.411676646706587, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1796, "step": 2279 }, { "epoch": 3.413173652694611, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1796, "step": 2280 }, { "epoch": 3.4146706586826348, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1808, "step": 2281 }, { "epoch": 3.4161676646706587, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1722, "step": 2282 }, { "epoch": 3.4176646706586826, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1758, "step": 2283 }, { "epoch": 3.4191616766467066, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.173, "step": 2284 }, { "epoch": 3.4206586826347305, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1709, "step": 2285 }, { "epoch": 3.4221556886227544, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.178, "step": 2286 }, { "epoch": 3.4236526946107784, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1745, "step": 2287 }, { "epoch": 3.4251497005988023, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1794, "step": 2288 }, { "epoch": 3.4266467065868262, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1739, "step": 2289 }, { "epoch": 3.42814371257485, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1737, "step": 2290 }, { "epoch": 3.429640718562874, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1778, "step": 2291 }, { "epoch": 3.431137724550898, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1774, "step": 2292 }, { "epoch": 3.432634730538922, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1737, "step": 2293 }, { "epoch": 3.434131736526946, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1736, "step": 2294 }, { "epoch": 3.43562874251497, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1722, "step": 2295 }, { "epoch": 3.437125748502994, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1791, "step": 2296 }, { "epoch": 3.4386227544910177, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1789, "step": 2297 }, { "epoch": 3.440119760479042, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1798, "step": 2298 }, { "epoch": 3.441616766467066, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.176, "step": 2299 }, { "epoch": 3.44311377245509, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1806, "step": 2300 }, { "epoch": 3.444610778443114, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1718, "step": 2301 }, { "epoch": 3.446107784431138, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1687, "step": 2302 }, { "epoch": 3.447604790419162, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1742, "step": 2303 }, { "epoch": 3.4491017964071857, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1784, "step": 2304 }, { "epoch": 3.4505988023952097, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1756, "step": 2305 }, { "epoch": 3.4520958083832336, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1715, "step": 2306 }, { "epoch": 3.4535928143712575, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1709, "step": 2307 }, { "epoch": 3.4550898203592815, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1721, "step": 2308 }, { "epoch": 3.4565868263473054, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1842, "step": 2309 }, { "epoch": 3.4580838323353293, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1835, "step": 2310 }, { "epoch": 3.4595808383233533, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1761, "step": 2311 }, { "epoch": 3.461077844311377, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1756, "step": 2312 }, { "epoch": 3.462574850299401, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1753, "step": 2313 }, { "epoch": 3.464071856287425, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1773, "step": 2314 }, { "epoch": 3.465568862275449, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1696, "step": 2315 }, { "epoch": 3.467065868263473, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1746, "step": 2316 }, { "epoch": 3.468562874251497, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1728, "step": 2317 }, { "epoch": 3.470059880239521, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1763, "step": 2318 }, { "epoch": 3.4715568862275448, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1686, "step": 2319 }, { "epoch": 3.4730538922155687, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1728, "step": 2320 }, { "epoch": 3.4745508982035926, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1736, "step": 2321 }, { "epoch": 3.4760479041916166, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1753, "step": 2322 }, { "epoch": 3.477544910179641, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1808, "step": 2323 }, { "epoch": 3.479041916167665, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1715, "step": 2324 }, { "epoch": 3.480538922155689, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1839, "step": 2325 }, { "epoch": 3.4820359281437128, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1747, "step": 2326 }, { "epoch": 3.4835329341317367, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1778, "step": 2327 }, { "epoch": 3.4850299401197606, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1673, "step": 2328 }, { "epoch": 3.4865269461077846, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1832, "step": 2329 }, { "epoch": 3.4880239520958085, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.17, "step": 2330 }, { "epoch": 3.4895209580838324, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1751, "step": 2331 }, { "epoch": 3.4910179640718564, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.172, "step": 2332 }, { "epoch": 3.4925149700598803, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.172, "step": 2333 }, { "epoch": 3.4940119760479043, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1814, "step": 2334 }, { "epoch": 3.495508982035928, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1769, "step": 2335 }, { "epoch": 3.497005988023952, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1684, "step": 2336 }, { "epoch": 3.498502994011976, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1831, "step": 2337 }, { "epoch": 3.5, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1738, "step": 2338 }, { "epoch": 3.501497005988024, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1823, "step": 2339 }, { "epoch": 3.502994011976048, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1778, "step": 2340 }, { "epoch": 3.504491017964072, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1802, "step": 2341 }, { "epoch": 3.5059880239520957, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1745, "step": 2342 }, { "epoch": 3.5074850299401197, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1705, "step": 2343 }, { "epoch": 3.5089820359281436, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1705, "step": 2344 }, { "epoch": 3.5104790419161676, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1762, "step": 2345 }, { "epoch": 3.5119760479041915, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1699, "step": 2346 }, { "epoch": 3.5134730538922154, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1743, "step": 2347 }, { "epoch": 3.5149700598802394, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1783, "step": 2348 }, { "epoch": 3.5164670658682633, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1728, "step": 2349 }, { "epoch": 3.5179640718562872, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1739, "step": 2350 }, { "epoch": 3.519461077844311, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.178, "step": 2351 }, { "epoch": 3.520958083832335, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.184, "step": 2352 }, { "epoch": 3.522455089820359, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1738, "step": 2353 }, { "epoch": 3.523952095808383, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1737, "step": 2354 }, { "epoch": 3.525449101796407, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1819, "step": 2355 }, { "epoch": 3.5269461077844313, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1711, "step": 2356 }, { "epoch": 3.5284431137724552, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1685, "step": 2357 }, { "epoch": 3.529940119760479, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1759, "step": 2358 }, { "epoch": 3.531437125748503, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1716, "step": 2359 }, { "epoch": 3.532934131736527, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1773, "step": 2360 }, { "epoch": 3.534431137724551, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1782, "step": 2361 }, { "epoch": 3.535928143712575, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1748, "step": 2362 }, { "epoch": 3.537425149700599, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1752, "step": 2363 }, { "epoch": 3.538922155688623, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1691, "step": 2364 }, { "epoch": 3.5404191616766467, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1711, "step": 2365 }, { "epoch": 3.5419161676646707, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1825, "step": 2366 }, { "epoch": 3.5434131736526946, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1831, "step": 2367 }, { "epoch": 3.5449101796407185, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1745, "step": 2368 }, { "epoch": 3.5464071856287425, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1772, "step": 2369 }, { "epoch": 3.5479041916167664, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1674, "step": 2370 }, { "epoch": 3.5494011976047903, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1781, "step": 2371 }, { "epoch": 3.5508982035928143, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1764, "step": 2372 }, { "epoch": 3.552395209580838, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1799, "step": 2373 }, { "epoch": 3.553892215568862, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1786, "step": 2374 }, { "epoch": 3.555389221556886, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.178, "step": 2375 }, { "epoch": 3.55688622754491, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1824, "step": 2376 }, { "epoch": 3.5583832335329344, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1739, "step": 2377 }, { "epoch": 3.5598802395209583, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1771, "step": 2378 }, { "epoch": 3.5613772455089823, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1802, "step": 2379 }, { "epoch": 3.562874251497006, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1759, "step": 2380 }, { "epoch": 3.56437125748503, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1756, "step": 2381 }, { "epoch": 3.565868263473054, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.178, "step": 2382 }, { "epoch": 3.567365269461078, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1741, "step": 2383 }, { "epoch": 3.568862275449102, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1812, "step": 2384 }, { "epoch": 3.570359281437126, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.178, "step": 2385 }, { "epoch": 3.57185628742515, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1743, "step": 2386 }, { "epoch": 3.5733532934131738, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1732, "step": 2387 }, { "epoch": 3.5748502994011977, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1734, "step": 2388 }, { "epoch": 3.5763473053892216, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1696, "step": 2389 }, { "epoch": 3.5778443113772456, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1779, "step": 2390 }, { "epoch": 3.5793413173652695, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1778, "step": 2391 }, { "epoch": 3.5808383233532934, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1757, "step": 2392 }, { "epoch": 3.5823353293413174, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.1754, "step": 2393 }, { "epoch": 3.5838323353293413, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1786, "step": 2394 }, { "epoch": 3.5853293413173652, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.1803, "step": 2395 }, { "epoch": 3.586826347305389, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1738, "step": 2396 }, { "epoch": 3.588323353293413, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.165, "step": 2397 }, { "epoch": 3.589820359281437, "grad_norm": 0.17578125, "learning_rate": 0.0008, "loss": 1.1724, "step": 2398 }, { "epoch": 3.591317365269461, "grad_norm": 0.224609375, "learning_rate": 0.0008, "loss": 1.1817, "step": 2399 }, { "epoch": 3.592814371257485, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.1725, "step": 2400 }, { "epoch": 3.594311377245509, "grad_norm": 0.486328125, "learning_rate": 0.0008, "loss": 1.1851, "step": 2401 }, { "epoch": 3.595808383233533, "grad_norm": 1.0, "learning_rate": 0.0008, "loss": 1.1945, "step": 2402 }, { "epoch": 3.5973053892215567, "grad_norm": 1.6484375, "learning_rate": 0.0008, "loss": 1.24, "step": 2403 }, { "epoch": 3.5988023952095807, "grad_norm": 0.37890625, "learning_rate": 0.0008, "loss": 1.1938, "step": 2404 }, { "epoch": 3.6002994011976046, "grad_norm": 1.1484375, "learning_rate": 0.0008, "loss": 1.2344, "step": 2405 }, { "epoch": 3.6017964071856285, "grad_norm": 0.39453125, "learning_rate": 0.0008, "loss": 1.2008, "step": 2406 }, { "epoch": 3.6032934131736525, "grad_norm": 0.546875, "learning_rate": 0.0008, "loss": 1.2244, "step": 2407 }, { "epoch": 3.6047904191616764, "grad_norm": 0.314453125, "learning_rate": 0.0008, "loss": 1.2083, "step": 2408 }, { "epoch": 3.6062874251497004, "grad_norm": 0.24609375, "learning_rate": 0.0008, "loss": 1.198, "step": 2409 }, { "epoch": 3.6077844311377243, "grad_norm": 0.2333984375, "learning_rate": 0.0008, "loss": 1.2014, "step": 2410 }, { "epoch": 3.6092814371257482, "grad_norm": 0.1982421875, "learning_rate": 0.0008, "loss": 1.1989, "step": 2411 }, { "epoch": 3.6107784431137726, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.1992, "step": 2412 }, { "epoch": 3.6122754491017965, "grad_norm": 0.171875, "learning_rate": 0.0008, "loss": 1.1956, "step": 2413 }, { "epoch": 3.6137724550898205, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.1881, "step": 2414 }, { "epoch": 3.6152694610778444, "grad_norm": 0.189453125, "learning_rate": 0.0008, "loss": 1.195, "step": 2415 }, { "epoch": 3.6167664670658684, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1901, "step": 2416 }, { "epoch": 3.6182634730538923, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1778, "step": 2417 }, { "epoch": 3.6197604790419162, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1927, "step": 2418 }, { "epoch": 3.62125748502994, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.184, "step": 2419 }, { "epoch": 3.622754491017964, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1833, "step": 2420 }, { "epoch": 3.624251497005988, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1809, "step": 2421 }, { "epoch": 3.625748502994012, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1795, "step": 2422 }, { "epoch": 3.627245508982036, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1864, "step": 2423 }, { "epoch": 3.62874251497006, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1775, "step": 2424 }, { "epoch": 3.6302395209580838, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.185, "step": 2425 }, { "epoch": 3.6317365269461077, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.185, "step": 2426 }, { "epoch": 3.6332335329341316, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1895, "step": 2427 }, { "epoch": 3.6347305389221556, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1788, "step": 2428 }, { "epoch": 3.6362275449101795, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1795, "step": 2429 }, { "epoch": 3.6377245508982035, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1752, "step": 2430 }, { "epoch": 3.6392215568862274, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1735, "step": 2431 }, { "epoch": 3.6407185628742518, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1855, "step": 2432 }, { "epoch": 3.6422155688622757, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1738, "step": 2433 }, { "epoch": 3.6437125748502996, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.175, "step": 2434 }, { "epoch": 3.6452095808383236, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1782, "step": 2435 }, { "epoch": 3.6467065868263475, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1755, "step": 2436 }, { "epoch": 3.6482035928143715, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1661, "step": 2437 }, { "epoch": 3.6497005988023954, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1763, "step": 2438 }, { "epoch": 3.6511976047904193, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.171, "step": 2439 }, { "epoch": 3.6526946107784433, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1772, "step": 2440 }, { "epoch": 3.654191616766467, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1738, "step": 2441 }, { "epoch": 3.655688622754491, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1767, "step": 2442 }, { "epoch": 3.657185628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1754, "step": 2443 }, { "epoch": 3.658682634730539, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1735, "step": 2444 }, { "epoch": 3.660179640718563, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1733, "step": 2445 }, { "epoch": 3.661676646706587, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1792, "step": 2446 }, { "epoch": 3.663173652694611, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1738, "step": 2447 }, { "epoch": 3.6646706586826348, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1787, "step": 2448 }, { "epoch": 3.6661676646706587, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1723, "step": 2449 }, { "epoch": 3.6676646706586826, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1706, "step": 2450 }, { "epoch": 3.6691616766467066, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1735, "step": 2451 }, { "epoch": 3.6706586826347305, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1801, "step": 2452 }, { "epoch": 3.6721556886227544, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1686, "step": 2453 }, { "epoch": 3.6736526946107784, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1707, "step": 2454 }, { "epoch": 3.6751497005988023, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1685, "step": 2455 }, { "epoch": 3.6766467065868262, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1617, "step": 2456 }, { "epoch": 3.67814371257485, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1713, "step": 2457 }, { "epoch": 3.679640718562874, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.181, "step": 2458 }, { "epoch": 3.681137724550898, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1706, "step": 2459 }, { "epoch": 3.682634730538922, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1726, "step": 2460 }, { "epoch": 3.684131736526946, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1736, "step": 2461 }, { "epoch": 3.68562874251497, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1752, "step": 2462 }, { "epoch": 3.687125748502994, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1707, "step": 2463 }, { "epoch": 3.6886227544910177, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1779, "step": 2464 }, { "epoch": 3.6901197604790417, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1761, "step": 2465 }, { "epoch": 3.6916167664670656, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1791, "step": 2466 }, { "epoch": 3.69311377245509, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1753, "step": 2467 }, { "epoch": 3.694610778443114, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1748, "step": 2468 }, { "epoch": 3.696107784431138, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1829, "step": 2469 }, { "epoch": 3.697604790419162, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1736, "step": 2470 }, { "epoch": 3.6991017964071857, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1752, "step": 2471 }, { "epoch": 3.7005988023952097, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1702, "step": 2472 }, { "epoch": 3.7020958083832336, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.172, "step": 2473 }, { "epoch": 3.7035928143712575, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1714, "step": 2474 }, { "epoch": 3.7050898203592815, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1745, "step": 2475 }, { "epoch": 3.7065868263473054, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1675, "step": 2476 }, { "epoch": 3.7080838323353293, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1662, "step": 2477 }, { "epoch": 3.7095808383233533, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1772, "step": 2478 }, { "epoch": 3.711077844311377, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1769, "step": 2479 }, { "epoch": 3.712574850299401, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1733, "step": 2480 }, { "epoch": 3.714071856287425, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.176, "step": 2481 }, { "epoch": 3.715568862275449, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1835, "step": 2482 }, { "epoch": 3.717065868263473, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.182, "step": 2483 }, { "epoch": 3.718562874251497, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.173, "step": 2484 }, { "epoch": 3.720059880239521, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1762, "step": 2485 }, { "epoch": 3.7215568862275448, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1737, "step": 2486 }, { "epoch": 3.7230538922155687, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1685, "step": 2487 }, { "epoch": 3.724550898203593, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1731, "step": 2488 }, { "epoch": 3.726047904191617, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1713, "step": 2489 }, { "epoch": 3.727544910179641, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.168, "step": 2490 }, { "epoch": 3.729041916167665, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1633, "step": 2491 }, { "epoch": 3.730538922155689, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1759, "step": 2492 }, { "epoch": 3.7320359281437128, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.169, "step": 2493 }, { "epoch": 3.7335329341317367, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1729, "step": 2494 }, { "epoch": 3.7350299401197606, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1702, "step": 2495 }, { "epoch": 3.7365269461077846, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1731, "step": 2496 }, { "epoch": 3.7380239520958085, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1754, "step": 2497 }, { "epoch": 3.7395209580838324, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1674, "step": 2498 }, { "epoch": 3.7410179640718564, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1773, "step": 2499 }, { "epoch": 3.7425149700598803, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1686, "step": 2500 }, { "epoch": 3.7440119760479043, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1743, "step": 2501 }, { "epoch": 3.745508982035928, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1712, "step": 2502 }, { "epoch": 3.747005988023952, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1732, "step": 2503 }, { "epoch": 3.748502994011976, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1741, "step": 2504 }, { "epoch": 3.75, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1766, "step": 2505 }, { "epoch": 3.751497005988024, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1733, "step": 2506 }, { "epoch": 3.752994011976048, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1707, "step": 2507 }, { "epoch": 3.754491017964072, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1724, "step": 2508 }, { "epoch": 3.7559880239520957, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1689, "step": 2509 }, { "epoch": 3.7574850299401197, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1753, "step": 2510 }, { "epoch": 3.7589820359281436, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1746, "step": 2511 }, { "epoch": 3.7604790419161676, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1692, "step": 2512 }, { "epoch": 3.7619760479041915, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1814, "step": 2513 }, { "epoch": 3.7634730538922154, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1742, "step": 2514 }, { "epoch": 3.7649700598802394, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1716, "step": 2515 }, { "epoch": 3.7664670658682633, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1728, "step": 2516 }, { "epoch": 3.7679640718562872, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1756, "step": 2517 }, { "epoch": 3.769461077844311, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1734, "step": 2518 }, { "epoch": 3.770958083832335, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1751, "step": 2519 }, { "epoch": 3.772455089820359, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1694, "step": 2520 }, { "epoch": 3.773952095808383, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1784, "step": 2521 }, { "epoch": 3.775449101796407, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1704, "step": 2522 }, { "epoch": 3.7769461077844313, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1696, "step": 2523 }, { "epoch": 3.7784431137724552, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1694, "step": 2524 }, { "epoch": 3.779940119760479, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1702, "step": 2525 }, { "epoch": 3.781437125748503, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1701, "step": 2526 }, { "epoch": 3.782934131736527, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1714, "step": 2527 }, { "epoch": 3.784431137724551, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1746, "step": 2528 }, { "epoch": 3.785928143712575, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1776, "step": 2529 }, { "epoch": 3.787425149700599, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1674, "step": 2530 }, { "epoch": 3.788922155688623, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1682, "step": 2531 }, { "epoch": 3.7904191616766467, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1761, "step": 2532 }, { "epoch": 3.7919161676646707, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1727, "step": 2533 }, { "epoch": 3.7934131736526946, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1713, "step": 2534 }, { "epoch": 3.7949101796407185, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1706, "step": 2535 }, { "epoch": 3.7964071856287425, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1706, "step": 2536 }, { "epoch": 3.7979041916167664, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1829, "step": 2537 }, { "epoch": 3.7994011976047903, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.169, "step": 2538 }, { "epoch": 3.8008982035928143, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1814, "step": 2539 }, { "epoch": 3.802395209580838, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1663, "step": 2540 }, { "epoch": 3.803892215568862, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1709, "step": 2541 }, { "epoch": 3.805389221556886, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.167, "step": 2542 }, { "epoch": 3.80688622754491, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1723, "step": 2543 }, { "epoch": 3.8083832335329344, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1704, "step": 2544 }, { "epoch": 3.8098802395209583, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1757, "step": 2545 }, { "epoch": 3.8113772455089823, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1716, "step": 2546 }, { "epoch": 3.812874251497006, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.179, "step": 2547 }, { "epoch": 3.81437125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1723, "step": 2548 }, { "epoch": 3.815868263473054, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1703, "step": 2549 }, { "epoch": 3.817365269461078, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1714, "step": 2550 }, { "epoch": 3.818862275449102, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.176, "step": 2551 }, { "epoch": 3.820359281437126, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1764, "step": 2552 }, { "epoch": 3.82185628742515, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.169, "step": 2553 }, { "epoch": 3.8233532934131738, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1763, "step": 2554 }, { "epoch": 3.8248502994011977, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1697, "step": 2555 }, { "epoch": 3.8263473053892216, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1669, "step": 2556 }, { "epoch": 3.8278443113772456, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1722, "step": 2557 }, { "epoch": 3.8293413173652695, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.165, "step": 2558 }, { "epoch": 3.8308383233532934, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1711, "step": 2559 }, { "epoch": 3.8323353293413174, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1667, "step": 2560 }, { "epoch": 3.8338323353293413, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1645, "step": 2561 }, { "epoch": 3.8353293413173652, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1736, "step": 2562 }, { "epoch": 3.836826347305389, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1743, "step": 2563 }, { "epoch": 3.838323353293413, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1659, "step": 2564 }, { "epoch": 3.839820359281437, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1734, "step": 2565 }, { "epoch": 3.841317365269461, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.176, "step": 2566 }, { "epoch": 3.842814371257485, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1731, "step": 2567 }, { "epoch": 3.844311377245509, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1751, "step": 2568 }, { "epoch": 3.845808383233533, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1704, "step": 2569 }, { "epoch": 3.8473053892215567, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1695, "step": 2570 }, { "epoch": 3.8488023952095807, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1724, "step": 2571 }, { "epoch": 3.8502994011976046, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1688, "step": 2572 }, { "epoch": 3.8517964071856285, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.171, "step": 2573 }, { "epoch": 3.8532934131736525, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1785, "step": 2574 }, { "epoch": 3.8547904191616764, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1703, "step": 2575 }, { "epoch": 3.8562874251497004, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.171, "step": 2576 }, { "epoch": 3.8577844311377243, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.172, "step": 2577 }, { "epoch": 3.8592814371257482, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1724, "step": 2578 }, { "epoch": 3.8607784431137726, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1699, "step": 2579 }, { "epoch": 3.8622754491017965, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1742, "step": 2580 }, { "epoch": 3.8637724550898205, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1762, "step": 2581 }, { "epoch": 3.8652694610778444, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1658, "step": 2582 }, { "epoch": 3.8667664670658684, "grad_norm": 0.265625, "learning_rate": 0.0008, "loss": 1.1766, "step": 2583 }, { "epoch": 3.8682634730538923, "grad_norm": 0.384765625, "learning_rate": 0.0008, "loss": 1.1762, "step": 2584 }, { "epoch": 3.8697604790419162, "grad_norm": 0.68359375, "learning_rate": 0.0008, "loss": 1.1842, "step": 2585 }, { "epoch": 3.87125748502994, "grad_norm": 1.140625, "learning_rate": 0.0008, "loss": 1.2016, "step": 2586 }, { "epoch": 3.872754491017964, "grad_norm": 0.91015625, "learning_rate": 0.0008, "loss": 1.195, "step": 2587 }, { "epoch": 3.874251497005988, "grad_norm": 0.328125, "learning_rate": 0.0008, "loss": 1.1825, "step": 2588 }, { "epoch": 3.875748502994012, "grad_norm": 0.48828125, "learning_rate": 0.0008, "loss": 1.195, "step": 2589 }, { "epoch": 3.877245508982036, "grad_norm": 0.2734375, "learning_rate": 0.0008, "loss": 1.1849, "step": 2590 }, { "epoch": 3.87874251497006, "grad_norm": 0.306640625, "learning_rate": 0.0008, "loss": 1.1929, "step": 2591 }, { "epoch": 3.8802395209580838, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.1809, "step": 2592 }, { "epoch": 3.8817365269461077, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.1813, "step": 2593 }, { "epoch": 3.8832335329341316, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.1835, "step": 2594 }, { "epoch": 3.8847305389221556, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.1786, "step": 2595 }, { "epoch": 3.8862275449101795, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1772, "step": 2596 }, { "epoch": 3.8877245508982035, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1751, "step": 2597 }, { "epoch": 3.8892215568862274, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1808, "step": 2598 }, { "epoch": 3.8907185628742518, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1784, "step": 2599 }, { "epoch": 3.8922155688622757, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1724, "step": 2600 }, { "epoch": 3.8937125748502996, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1768, "step": 2601 }, { "epoch": 3.8952095808383236, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1761, "step": 2602 }, { "epoch": 3.8967065868263475, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1711, "step": 2603 }, { "epoch": 3.8982035928143715, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1731, "step": 2604 }, { "epoch": 3.8997005988023954, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1782, "step": 2605 }, { "epoch": 3.9011976047904193, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1786, "step": 2606 }, { "epoch": 3.9026946107784433, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1726, "step": 2607 }, { "epoch": 3.904191616766467, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1703, "step": 2608 }, { "epoch": 3.905688622754491, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1773, "step": 2609 }, { "epoch": 3.907185628742515, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1703, "step": 2610 }, { "epoch": 3.908682634730539, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1656, "step": 2611 }, { "epoch": 3.910179640718563, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1732, "step": 2612 }, { "epoch": 3.911676646706587, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1735, "step": 2613 }, { "epoch": 3.913173652694611, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1686, "step": 2614 }, { "epoch": 3.9146706586826348, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1787, "step": 2615 }, { "epoch": 3.9161676646706587, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1739, "step": 2616 }, { "epoch": 3.9176646706586826, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1768, "step": 2617 }, { "epoch": 3.9191616766467066, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1775, "step": 2618 }, { "epoch": 3.9206586826347305, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1748, "step": 2619 }, { "epoch": 3.9221556886227544, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1675, "step": 2620 }, { "epoch": 3.9236526946107784, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1699, "step": 2621 }, { "epoch": 3.9251497005988023, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1734, "step": 2622 }, { "epoch": 3.9266467065868262, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1738, "step": 2623 }, { "epoch": 3.92814371257485, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1694, "step": 2624 }, { "epoch": 3.929640718562874, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1726, "step": 2625 }, { "epoch": 3.931137724550898, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1719, "step": 2626 }, { "epoch": 3.932634730538922, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1727, "step": 2627 }, { "epoch": 3.934131736526946, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.17, "step": 2628 }, { "epoch": 3.93562874251497, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1735, "step": 2629 }, { "epoch": 3.937125748502994, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1718, "step": 2630 }, { "epoch": 3.9386227544910177, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1733, "step": 2631 }, { "epoch": 3.9401197604790417, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1728, "step": 2632 }, { "epoch": 3.9416167664670656, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1714, "step": 2633 }, { "epoch": 3.94311377245509, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1749, "step": 2634 }, { "epoch": 3.944610778443114, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1653, "step": 2635 }, { "epoch": 3.946107784431138, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1714, "step": 2636 }, { "epoch": 3.947604790419162, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1722, "step": 2637 }, { "epoch": 3.9491017964071857, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1649, "step": 2638 }, { "epoch": 3.9505988023952097, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1731, "step": 2639 }, { "epoch": 3.9520958083832336, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1622, "step": 2640 }, { "epoch": 3.9535928143712575, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1654, "step": 2641 }, { "epoch": 3.9550898203592815, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1689, "step": 2642 }, { "epoch": 3.9565868263473054, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.169, "step": 2643 }, { "epoch": 3.9580838323353293, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1737, "step": 2644 }, { "epoch": 3.9595808383233533, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.175, "step": 2645 }, { "epoch": 3.961077844311377, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1711, "step": 2646 }, { "epoch": 3.962574850299401, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1717, "step": 2647 }, { "epoch": 3.964071856287425, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1686, "step": 2648 }, { "epoch": 3.965568862275449, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1654, "step": 2649 }, { "epoch": 3.967065868263473, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.178, "step": 2650 }, { "epoch": 3.968562874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1751, "step": 2651 }, { "epoch": 3.970059880239521, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1692, "step": 2652 }, { "epoch": 3.9715568862275448, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1743, "step": 2653 }, { "epoch": 3.9730538922155687, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1795, "step": 2654 }, { "epoch": 3.974550898203593, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1774, "step": 2655 }, { "epoch": 3.976047904191617, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1705, "step": 2656 }, { "epoch": 3.977544910179641, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1787, "step": 2657 }, { "epoch": 3.979041916167665, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1726, "step": 2658 }, { "epoch": 3.980538922155689, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.165, "step": 2659 }, { "epoch": 3.9820359281437128, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1711, "step": 2660 }, { "epoch": 3.9835329341317367, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1704, "step": 2661 }, { "epoch": 3.9850299401197606, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.17, "step": 2662 }, { "epoch": 3.9865269461077846, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1733, "step": 2663 }, { "epoch": 3.9880239520958085, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.169, "step": 2664 }, { "epoch": 3.9895209580838324, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1696, "step": 2665 }, { "epoch": 3.9910179640718564, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.18, "step": 2666 }, { "epoch": 3.9925149700598803, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1663, "step": 2667 }, { "epoch": 3.9940119760479043, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1693, "step": 2668 }, { "epoch": 3.995508982035928, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1683, "step": 2669 }, { "epoch": 3.997005988023952, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1745, "step": 2670 }, { "epoch": 3.998502994011976, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1649, "step": 2671 }, { "epoch": 4.0, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1702, "step": 2672 }, { "epoch": 4.001497005988024, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1743, "step": 2673 }, { "epoch": 4.002994011976048, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1709, "step": 2674 }, { "epoch": 4.004491017964072, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1735, "step": 2675 }, { "epoch": 4.005988023952096, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1728, "step": 2676 }, { "epoch": 4.00748502994012, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1692, "step": 2677 }, { "epoch": 4.008982035928144, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.171, "step": 2678 }, { "epoch": 4.0104790419161676, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1659, "step": 2679 }, { "epoch": 4.0119760479041915, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1722, "step": 2680 }, { "epoch": 4.013473053892215, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1771, "step": 2681 }, { "epoch": 4.014970059880239, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1743, "step": 2682 }, { "epoch": 4.016467065868263, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.176, "step": 2683 }, { "epoch": 4.017964071856287, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1657, "step": 2684 }, { "epoch": 4.019461077844311, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1695, "step": 2685 }, { "epoch": 4.020958083832335, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1688, "step": 2686 }, { "epoch": 4.022455089820359, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1737, "step": 2687 }, { "epoch": 4.023952095808383, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1599, "step": 2688 }, { "epoch": 4.025449101796407, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.17, "step": 2689 }, { "epoch": 4.026946107784431, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1746, "step": 2690 }, { "epoch": 4.028443113772455, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1708, "step": 2691 }, { "epoch": 4.029940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1705, "step": 2692 }, { "epoch": 4.031437125748503, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1654, "step": 2693 }, { "epoch": 4.032934131736527, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1732, "step": 2694 }, { "epoch": 4.0344311377245505, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1728, "step": 2695 }, { "epoch": 4.0359281437125745, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1683, "step": 2696 }, { "epoch": 4.037425149700598, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1724, "step": 2697 }, { "epoch": 4.038922155688622, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1634, "step": 2698 }, { "epoch": 4.040419161676646, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1735, "step": 2699 }, { "epoch": 4.04191616766467, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1717, "step": 2700 }, { "epoch": 4.043413173652695, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1666, "step": 2701 }, { "epoch": 4.044910179640719, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1659, "step": 2702 }, { "epoch": 4.046407185628743, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1679, "step": 2703 }, { "epoch": 4.047904191616767, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1803, "step": 2704 }, { "epoch": 4.049401197604791, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1744, "step": 2705 }, { "epoch": 4.050898203592815, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.172, "step": 2706 }, { "epoch": 4.052395209580839, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1713, "step": 2707 }, { "epoch": 4.053892215568863, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1728, "step": 2708 }, { "epoch": 4.0553892215568865, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1743, "step": 2709 }, { "epoch": 4.0568862275449105, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1618, "step": 2710 }, { "epoch": 4.058383233532934, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1694, "step": 2711 }, { "epoch": 4.059880239520958, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1634, "step": 2712 }, { "epoch": 4.061377245508982, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1679, "step": 2713 }, { "epoch": 4.062874251497006, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1737, "step": 2714 }, { "epoch": 4.06437125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1684, "step": 2715 }, { "epoch": 4.065868263473054, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1723, "step": 2716 }, { "epoch": 4.067365269461078, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.173, "step": 2717 }, { "epoch": 4.068862275449102, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1679, "step": 2718 }, { "epoch": 4.070359281437126, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1663, "step": 2719 }, { "epoch": 4.07185628742515, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1788, "step": 2720 }, { "epoch": 4.073353293413174, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1678, "step": 2721 }, { "epoch": 4.074850299401198, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1627, "step": 2722 }, { "epoch": 4.076347305389222, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1725, "step": 2723 }, { "epoch": 4.077844311377246, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1735, "step": 2724 }, { "epoch": 4.0793413173652695, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1627, "step": 2725 }, { "epoch": 4.080838323353293, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1658, "step": 2726 }, { "epoch": 4.082335329341317, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1705, "step": 2727 }, { "epoch": 4.083832335329341, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1732, "step": 2728 }, { "epoch": 4.085329341317365, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1729, "step": 2729 }, { "epoch": 4.086826347305389, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1642, "step": 2730 }, { "epoch": 4.088323353293413, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1716, "step": 2731 }, { "epoch": 4.089820359281437, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1649, "step": 2732 }, { "epoch": 4.091317365269461, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1698, "step": 2733 }, { "epoch": 4.092814371257485, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.17, "step": 2734 }, { "epoch": 4.094311377245509, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1727, "step": 2735 }, { "epoch": 4.095808383233533, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.17, "step": 2736 }, { "epoch": 4.097305389221557, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.164, "step": 2737 }, { "epoch": 4.098802395209581, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1721, "step": 2738 }, { "epoch": 4.100299401197605, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1679, "step": 2739 }, { "epoch": 4.1017964071856285, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1601, "step": 2740 }, { "epoch": 4.1032934131736525, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1654, "step": 2741 }, { "epoch": 4.104790419161676, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1728, "step": 2742 }, { "epoch": 4.1062874251497, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1683, "step": 2743 }, { "epoch": 4.107784431137724, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.169, "step": 2744 }, { "epoch": 4.109281437125748, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1686, "step": 2745 }, { "epoch": 4.110778443113772, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1742, "step": 2746 }, { "epoch": 4.112275449101796, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1683, "step": 2747 }, { "epoch": 4.11377245508982, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1677, "step": 2748 }, { "epoch": 4.115269461077844, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1623, "step": 2749 }, { "epoch": 4.116766467065868, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1598, "step": 2750 }, { "epoch": 4.118263473053892, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1741, "step": 2751 }, { "epoch": 4.119760479041916, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1774, "step": 2752 }, { "epoch": 4.12125748502994, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1661, "step": 2753 }, { "epoch": 4.122754491017964, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1627, "step": 2754 }, { "epoch": 4.124251497005988, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1728, "step": 2755 }, { "epoch": 4.125748502994012, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1683, "step": 2756 }, { "epoch": 4.127245508982036, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.168, "step": 2757 }, { "epoch": 4.12874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1746, "step": 2758 }, { "epoch": 4.130239520958084, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1751, "step": 2759 }, { "epoch": 4.131736526946108, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1747, "step": 2760 }, { "epoch": 4.133233532934132, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.162, "step": 2761 }, { "epoch": 4.134730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1757, "step": 2762 }, { "epoch": 4.13622754491018, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1781, "step": 2763 }, { "epoch": 4.137724550898204, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1706, "step": 2764 }, { "epoch": 4.139221556886228, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1634, "step": 2765 }, { "epoch": 4.140718562874252, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1711, "step": 2766 }, { "epoch": 4.142215568862276, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1717, "step": 2767 }, { "epoch": 4.1437125748503, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.175, "step": 2768 }, { "epoch": 4.145209580838324, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.172, "step": 2769 }, { "epoch": 4.1467065868263475, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.169, "step": 2770 }, { "epoch": 4.1482035928143715, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1679, "step": 2771 }, { "epoch": 4.149700598802395, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1721, "step": 2772 }, { "epoch": 4.151197604790419, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1694, "step": 2773 }, { "epoch": 4.152694610778443, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1631, "step": 2774 }, { "epoch": 4.154191616766467, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1742, "step": 2775 }, { "epoch": 4.155688622754491, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1761, "step": 2776 }, { "epoch": 4.157185628742515, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1733, "step": 2777 }, { "epoch": 4.158682634730539, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1678, "step": 2778 }, { "epoch": 4.160179640718563, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1673, "step": 2779 }, { "epoch": 4.161676646706587, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1646, "step": 2780 }, { "epoch": 4.163173652694611, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1773, "step": 2781 }, { "epoch": 4.164670658682635, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1657, "step": 2782 }, { "epoch": 4.166167664670659, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.166, "step": 2783 }, { "epoch": 4.167664670658683, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1711, "step": 2784 }, { "epoch": 4.169161676646707, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.17, "step": 2785 }, { "epoch": 4.1706586826347305, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.167, "step": 2786 }, { "epoch": 4.172155688622754, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1712, "step": 2787 }, { "epoch": 4.173652694610778, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1751, "step": 2788 }, { "epoch": 4.175149700598802, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1672, "step": 2789 }, { "epoch": 4.176646706586826, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1659, "step": 2790 }, { "epoch": 4.17814371257485, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1686, "step": 2791 }, { "epoch": 4.179640718562874, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1619, "step": 2792 }, { "epoch": 4.181137724550898, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1705, "step": 2793 }, { "epoch": 4.182634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1692, "step": 2794 }, { "epoch": 4.184131736526946, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1704, "step": 2795 }, { "epoch": 4.18562874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1636, "step": 2796 }, { "epoch": 4.187125748502994, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.167, "step": 2797 }, { "epoch": 4.188622754491018, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1683, "step": 2798 }, { "epoch": 4.190119760479042, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1659, "step": 2799 }, { "epoch": 4.191616766467066, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1709, "step": 2800 }, { "epoch": 4.1931137724550895, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1673, "step": 2801 }, { "epoch": 4.1946107784431135, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1694, "step": 2802 }, { "epoch": 4.196107784431137, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1635, "step": 2803 }, { "epoch": 4.197604790419161, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1705, "step": 2804 }, { "epoch": 4.199101796407185, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1666, "step": 2805 }, { "epoch": 4.200598802395209, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1708, "step": 2806 }, { "epoch": 4.202095808383233, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1731, "step": 2807 }, { "epoch": 4.203592814371257, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1685, "step": 2808 }, { "epoch": 4.205089820359281, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1714, "step": 2809 }, { "epoch": 4.206586826347305, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1667, "step": 2810 }, { "epoch": 4.20808383233533, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1683, "step": 2811 }, { "epoch": 4.209580838323353, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1689, "step": 2812 }, { "epoch": 4.211077844311378, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1737, "step": 2813 }, { "epoch": 4.212574850299402, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1765, "step": 2814 }, { "epoch": 4.2140718562874255, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.163, "step": 2815 }, { "epoch": 4.2155688622754495, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1668, "step": 2816 }, { "epoch": 4.217065868263473, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1711, "step": 2817 }, { "epoch": 4.218562874251497, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1665, "step": 2818 }, { "epoch": 4.220059880239521, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1682, "step": 2819 }, { "epoch": 4.221556886227545, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1746, "step": 2820 }, { "epoch": 4.223053892215569, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1689, "step": 2821 }, { "epoch": 4.224550898203593, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1718, "step": 2822 }, { "epoch": 4.226047904191617, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1764, "step": 2823 }, { "epoch": 4.227544910179641, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1692, "step": 2824 }, { "epoch": 4.229041916167665, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1677, "step": 2825 }, { "epoch": 4.230538922155689, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1702, "step": 2826 }, { "epoch": 4.232035928143713, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1735, "step": 2827 }, { "epoch": 4.233532934131737, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1707, "step": 2828 }, { "epoch": 4.235029940119761, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1731, "step": 2829 }, { "epoch": 4.236526946107785, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1787, "step": 2830 }, { "epoch": 4.2380239520958085, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1683, "step": 2831 }, { "epoch": 4.2395209580838324, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1771, "step": 2832 }, { "epoch": 4.241017964071856, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1733, "step": 2833 }, { "epoch": 4.24251497005988, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1766, "step": 2834 }, { "epoch": 4.244011976047904, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1673, "step": 2835 }, { "epoch": 4.245508982035928, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1706, "step": 2836 }, { "epoch": 4.247005988023952, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1692, "step": 2837 }, { "epoch": 4.248502994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1663, "step": 2838 }, { "epoch": 4.25, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1647, "step": 2839 }, { "epoch": 4.251497005988024, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1691, "step": 2840 }, { "epoch": 4.252994011976048, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1724, "step": 2841 }, { "epoch": 4.254491017964072, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1672, "step": 2842 }, { "epoch": 4.255988023952096, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1668, "step": 2843 }, { "epoch": 4.25748502994012, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1674, "step": 2844 }, { "epoch": 4.258982035928144, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1707, "step": 2845 }, { "epoch": 4.2604790419161676, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1698, "step": 2846 }, { "epoch": 4.2619760479041915, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1739, "step": 2847 }, { "epoch": 4.263473053892215, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1635, "step": 2848 }, { "epoch": 4.264970059880239, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1747, "step": 2849 }, { "epoch": 4.266467065868263, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1627, "step": 2850 }, { "epoch": 4.267964071856287, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.171, "step": 2851 }, { "epoch": 4.269461077844311, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1747, "step": 2852 }, { "epoch": 4.270958083832335, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1705, "step": 2853 }, { "epoch": 4.272455089820359, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1663, "step": 2854 }, { "epoch": 4.273952095808383, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1695, "step": 2855 }, { "epoch": 4.275449101796407, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1771, "step": 2856 }, { "epoch": 4.276946107784431, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1617, "step": 2857 }, { "epoch": 4.278443113772455, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1652, "step": 2858 }, { "epoch": 4.279940119760479, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1666, "step": 2859 }, { "epoch": 4.281437125748503, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1716, "step": 2860 }, { "epoch": 4.282934131736527, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1644, "step": 2861 }, { "epoch": 4.2844311377245505, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1652, "step": 2862 }, { "epoch": 4.2859281437125745, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1697, "step": 2863 }, { "epoch": 4.287425149700598, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1657, "step": 2864 }, { "epoch": 4.288922155688622, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1701, "step": 2865 }, { "epoch": 4.290419161676647, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1689, "step": 2866 }, { "epoch": 4.29191616766467, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1747, "step": 2867 }, { "epoch": 4.293413173652695, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1674, "step": 2868 }, { "epoch": 4.294910179640719, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1676, "step": 2869 }, { "epoch": 4.296407185628743, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1756, "step": 2870 }, { "epoch": 4.297904191616767, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1726, "step": 2871 }, { "epoch": 4.299401197604791, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.174, "step": 2872 }, { "epoch": 4.300898203592815, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1635, "step": 2873 }, { "epoch": 4.302395209580839, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1755, "step": 2874 }, { "epoch": 4.303892215568863, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.166, "step": 2875 }, { "epoch": 4.3053892215568865, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1719, "step": 2876 }, { "epoch": 4.3068862275449105, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1654, "step": 2877 }, { "epoch": 4.308383233532934, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.162, "step": 2878 }, { "epoch": 4.309880239520958, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1643, "step": 2879 }, { "epoch": 4.311377245508982, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1743, "step": 2880 }, { "epoch": 4.312874251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1688, "step": 2881 }, { "epoch": 4.31437125748503, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1735, "step": 2882 }, { "epoch": 4.315868263473054, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1629, "step": 2883 }, { "epoch": 4.317365269461078, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1697, "step": 2884 }, { "epoch": 4.318862275449102, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1617, "step": 2885 }, { "epoch": 4.320359281437126, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1727, "step": 2886 }, { "epoch": 4.32185628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1721, "step": 2887 }, { "epoch": 4.323353293413174, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1629, "step": 2888 }, { "epoch": 4.324850299401198, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1695, "step": 2889 }, { "epoch": 4.326347305389222, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.162, "step": 2890 }, { "epoch": 4.327844311377246, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1658, "step": 2891 }, { "epoch": 4.3293413173652695, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1608, "step": 2892 }, { "epoch": 4.330838323353293, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1706, "step": 2893 }, { "epoch": 4.332335329341317, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.1666, "step": 2894 }, { "epoch": 4.333832335329341, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1694, "step": 2895 }, { "epoch": 4.335329341317365, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1771, "step": 2896 }, { "epoch": 4.336826347305389, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1733, "step": 2897 }, { "epoch": 4.338323353293413, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1666, "step": 2898 }, { "epoch": 4.339820359281437, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1686, "step": 2899 }, { "epoch": 4.341317365269461, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1629, "step": 2900 }, { "epoch": 4.342814371257485, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1683, "step": 2901 }, { "epoch": 4.344311377245509, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1612, "step": 2902 }, { "epoch": 4.345808383233533, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1692, "step": 2903 }, { "epoch": 4.347305389221557, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1706, "step": 2904 }, { "epoch": 4.348802395209581, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1669, "step": 2905 }, { "epoch": 4.350299401197605, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1667, "step": 2906 }, { "epoch": 4.3517964071856285, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1667, "step": 2907 }, { "epoch": 4.3532934131736525, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1596, "step": 2908 }, { "epoch": 4.354790419161676, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1713, "step": 2909 }, { "epoch": 4.3562874251497, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1635, "step": 2910 }, { "epoch": 4.357784431137724, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1675, "step": 2911 }, { "epoch": 4.359281437125748, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.17, "step": 2912 }, { "epoch": 4.360778443113772, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1723, "step": 2913 }, { "epoch": 4.362275449101796, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1695, "step": 2914 }, { "epoch": 4.36377245508982, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.169, "step": 2915 }, { "epoch": 4.365269461077844, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1658, "step": 2916 }, { "epoch": 4.366766467065868, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1719, "step": 2917 }, { "epoch": 4.368263473053892, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.176, "step": 2918 }, { "epoch": 4.369760479041916, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.168, "step": 2919 }, { "epoch": 4.37125748502994, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1657, "step": 2920 }, { "epoch": 4.3727544910179645, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1696, "step": 2921 }, { "epoch": 4.374251497005988, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1687, "step": 2922 }, { "epoch": 4.375748502994012, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1703, "step": 2923 }, { "epoch": 4.3772455089820355, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1687, "step": 2924 }, { "epoch": 4.37874251497006, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1685, "step": 2925 }, { "epoch": 4.380239520958084, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1703, "step": 2926 }, { "epoch": 4.381736526946108, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1663, "step": 2927 }, { "epoch": 4.383233532934132, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1633, "step": 2928 }, { "epoch": 4.384730538922156, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1702, "step": 2929 }, { "epoch": 4.38622754491018, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1697, "step": 2930 }, { "epoch": 4.387724550898204, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1746, "step": 2931 }, { "epoch": 4.389221556886228, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1702, "step": 2932 }, { "epoch": 4.390718562874252, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1657, "step": 2933 }, { "epoch": 4.392215568862276, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1638, "step": 2934 }, { "epoch": 4.3937125748503, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1696, "step": 2935 }, { "epoch": 4.395209580838324, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1714, "step": 2936 }, { "epoch": 4.3967065868263475, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1649, "step": 2937 }, { "epoch": 4.3982035928143715, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.169, "step": 2938 }, { "epoch": 4.399700598802395, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1694, "step": 2939 }, { "epoch": 4.401197604790419, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1719, "step": 2940 }, { "epoch": 4.402694610778443, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1716, "step": 2941 }, { "epoch": 4.404191616766467, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1696, "step": 2942 }, { "epoch": 4.405688622754491, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1636, "step": 2943 }, { "epoch": 4.407185628742515, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1643, "step": 2944 }, { "epoch": 4.408682634730539, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1651, "step": 2945 }, { "epoch": 4.410179640718563, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1728, "step": 2946 }, { "epoch": 4.411676646706587, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1691, "step": 2947 }, { "epoch": 4.413173652694611, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.1686, "step": 2948 }, { "epoch": 4.414670658682635, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1662, "step": 2949 }, { "epoch": 4.416167664670659, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1645, "step": 2950 }, { "epoch": 4.417664670658683, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1752, "step": 2951 }, { "epoch": 4.419161676646707, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1684, "step": 2952 }, { "epoch": 4.4206586826347305, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1658, "step": 2953 }, { "epoch": 4.422155688622754, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1729, "step": 2954 }, { "epoch": 4.423652694610778, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1759, "step": 2955 }, { "epoch": 4.425149700598802, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1629, "step": 2956 }, { "epoch": 4.426646706586826, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1652, "step": 2957 }, { "epoch": 4.42814371257485, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1695, "step": 2958 }, { "epoch": 4.429640718562874, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1694, "step": 2959 }, { "epoch": 4.431137724550898, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1651, "step": 2960 }, { "epoch": 4.432634730538922, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1707, "step": 2961 }, { "epoch": 4.434131736526946, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1767, "step": 2962 }, { "epoch": 4.43562874251497, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1626, "step": 2963 }, { "epoch": 4.437125748502994, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1639, "step": 2964 }, { "epoch": 4.438622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1654, "step": 2965 }, { "epoch": 4.440119760479042, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1645, "step": 2966 }, { "epoch": 4.441616766467066, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1643, "step": 2967 }, { "epoch": 4.4431137724550895, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1705, "step": 2968 }, { "epoch": 4.4446107784431135, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1651, "step": 2969 }, { "epoch": 4.446107784431137, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1666, "step": 2970 }, { "epoch": 4.447604790419161, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1667, "step": 2971 }, { "epoch": 4.449101796407185, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1727, "step": 2972 }, { "epoch": 4.450598802395209, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.173, "step": 2973 }, { "epoch": 4.452095808383233, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.172, "step": 2974 }, { "epoch": 4.453592814371257, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1671, "step": 2975 }, { "epoch": 4.455089820359281, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1698, "step": 2976 }, { "epoch": 4.456586826347305, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1648, "step": 2977 }, { "epoch": 4.45808383233533, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.168, "step": 2978 }, { "epoch": 4.459580838323353, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1632, "step": 2979 }, { "epoch": 4.461077844311378, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1646, "step": 2980 }, { "epoch": 4.462574850299402, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1554, "step": 2981 }, { "epoch": 4.4640718562874255, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.164, "step": 2982 }, { "epoch": 4.4655688622754495, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1702, "step": 2983 }, { "epoch": 4.467065868263473, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1629, "step": 2984 }, { "epoch": 4.468562874251497, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1664, "step": 2985 }, { "epoch": 4.470059880239521, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.167, "step": 2986 }, { "epoch": 4.471556886227545, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1611, "step": 2987 }, { "epoch": 4.473053892215569, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1733, "step": 2988 }, { "epoch": 4.474550898203593, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1699, "step": 2989 }, { "epoch": 4.476047904191617, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1655, "step": 2990 }, { "epoch": 4.477544910179641, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1698, "step": 2991 }, { "epoch": 4.479041916167665, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1669, "step": 2992 }, { "epoch": 4.480538922155689, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1706, "step": 2993 }, { "epoch": 4.482035928143713, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1751, "step": 2994 }, { "epoch": 4.483532934131737, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1688, "step": 2995 }, { "epoch": 4.485029940119761, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1696, "step": 2996 }, { "epoch": 4.486526946107785, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1613, "step": 2997 }, { "epoch": 4.4880239520958085, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1698, "step": 2998 }, { "epoch": 4.4895209580838324, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1695, "step": 2999 }, { "epoch": 4.491017964071856, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1731, "step": 3000 }, { "epoch": 4.49251497005988, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1676, "step": 3001 }, { "epoch": 4.494011976047904, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1759, "step": 3002 }, { "epoch": 4.495508982035928, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1696, "step": 3003 }, { "epoch": 4.497005988023952, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1733, "step": 3004 }, { "epoch": 4.498502994011976, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1715, "step": 3005 }, { "epoch": 4.5, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1671, "step": 3006 }, { "epoch": 4.501497005988024, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1656, "step": 3007 }, { "epoch": 4.502994011976048, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1629, "step": 3008 }, { "epoch": 4.504491017964072, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1709, "step": 3009 }, { "epoch": 4.505988023952096, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1643, "step": 3010 }, { "epoch": 4.50748502994012, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3011 }, { "epoch": 4.508982035928144, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1735, "step": 3012 }, { "epoch": 4.5104790419161676, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1668, "step": 3013 }, { "epoch": 4.5119760479041915, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1678, "step": 3014 }, { "epoch": 4.513473053892215, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3015 }, { "epoch": 4.514970059880239, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.163, "step": 3016 }, { "epoch": 4.516467065868263, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1639, "step": 3017 }, { "epoch": 4.517964071856287, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1697, "step": 3018 }, { "epoch": 4.519461077844311, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3019 }, { "epoch": 4.520958083832335, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1649, "step": 3020 }, { "epoch": 4.522455089820359, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.175, "step": 3021 }, { "epoch": 4.523952095808383, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1683, "step": 3022 }, { "epoch": 4.525449101796407, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1725, "step": 3023 }, { "epoch": 4.526946107784431, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1617, "step": 3024 }, { "epoch": 4.528443113772455, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1688, "step": 3025 }, { "epoch": 4.529940119760479, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.166, "step": 3026 }, { "epoch": 4.531437125748503, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1679, "step": 3027 }, { "epoch": 4.532934131736527, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.173, "step": 3028 }, { "epoch": 4.5344311377245505, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1705, "step": 3029 }, { "epoch": 4.5359281437125745, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.167, "step": 3030 }, { "epoch": 4.537425149700599, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1682, "step": 3031 }, { "epoch": 4.538922155688622, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1636, "step": 3032 }, { "epoch": 4.540419161676647, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1608, "step": 3033 }, { "epoch": 4.54191616766467, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1652, "step": 3034 }, { "epoch": 4.543413173652695, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1697, "step": 3035 }, { "epoch": 4.544910179640718, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1692, "step": 3036 }, { "epoch": 4.546407185628743, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.169, "step": 3037 }, { "epoch": 4.547904191616767, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1739, "step": 3038 }, { "epoch": 4.549401197604791, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1746, "step": 3039 }, { "epoch": 4.550898203592815, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1607, "step": 3040 }, { "epoch": 4.552395209580839, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1612, "step": 3041 }, { "epoch": 4.553892215568863, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1604, "step": 3042 }, { "epoch": 4.5553892215568865, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1662, "step": 3043 }, { "epoch": 4.5568862275449105, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1652, "step": 3044 }, { "epoch": 4.558383233532934, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1689, "step": 3045 }, { "epoch": 4.559880239520958, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.174, "step": 3046 }, { "epoch": 4.561377245508982, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1716, "step": 3047 }, { "epoch": 4.562874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1642, "step": 3048 }, { "epoch": 4.56437125748503, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1681, "step": 3049 }, { "epoch": 4.565868263473054, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1729, "step": 3050 }, { "epoch": 4.567365269461078, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1577, "step": 3051 }, { "epoch": 4.568862275449102, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1674, "step": 3052 }, { "epoch": 4.570359281437126, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1718, "step": 3053 }, { "epoch": 4.57185628742515, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1728, "step": 3054 }, { "epoch": 4.573353293413174, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1677, "step": 3055 }, { "epoch": 4.574850299401198, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1677, "step": 3056 }, { "epoch": 4.576347305389222, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1714, "step": 3057 }, { "epoch": 4.577844311377246, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.173, "step": 3058 }, { "epoch": 4.5793413173652695, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1668, "step": 3059 }, { "epoch": 4.580838323353293, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1657, "step": 3060 }, { "epoch": 4.582335329341317, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1677, "step": 3061 }, { "epoch": 4.583832335329341, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1621, "step": 3062 }, { "epoch": 4.585329341317365, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1739, "step": 3063 }, { "epoch": 4.586826347305389, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1659, "step": 3064 }, { "epoch": 4.588323353293413, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1748, "step": 3065 }, { "epoch": 4.589820359281437, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1651, "step": 3066 }, { "epoch": 4.591317365269461, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.167, "step": 3067 }, { "epoch": 4.592814371257485, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1605, "step": 3068 }, { "epoch": 4.594311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1631, "step": 3069 }, { "epoch": 4.595808383233533, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1609, "step": 3070 }, { "epoch": 4.597305389221557, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1661, "step": 3071 }, { "epoch": 4.598802395209581, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1749, "step": 3072 }, { "epoch": 4.600299401197605, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1701, "step": 3073 }, { "epoch": 4.6017964071856285, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1682, "step": 3074 }, { "epoch": 4.6032934131736525, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1558, "step": 3075 }, { "epoch": 4.604790419161676, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1672, "step": 3076 }, { "epoch": 4.6062874251497, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1664, "step": 3077 }, { "epoch": 4.607784431137724, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1641, "step": 3078 }, { "epoch": 4.609281437125748, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1697, "step": 3079 }, { "epoch": 4.610778443113772, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1636, "step": 3080 }, { "epoch": 4.612275449101796, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1635, "step": 3081 }, { "epoch": 4.61377245508982, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1706, "step": 3082 }, { "epoch": 4.615269461077844, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1737, "step": 3083 }, { "epoch": 4.616766467065868, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1653, "step": 3084 }, { "epoch": 4.618263473053892, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1644, "step": 3085 }, { "epoch": 4.619760479041916, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1644, "step": 3086 }, { "epoch": 4.62125748502994, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1737, "step": 3087 }, { "epoch": 4.6227544910179645, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1691, "step": 3088 }, { "epoch": 4.624251497005988, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1714, "step": 3089 }, { "epoch": 4.625748502994012, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1664, "step": 3090 }, { "epoch": 4.6272455089820355, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1725, "step": 3091 }, { "epoch": 4.62874251497006, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1666, "step": 3092 }, { "epoch": 4.630239520958084, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1665, "step": 3093 }, { "epoch": 4.631736526946108, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.171, "step": 3094 }, { "epoch": 4.633233532934132, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1646, "step": 3095 }, { "epoch": 4.634730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1675, "step": 3096 }, { "epoch": 4.63622754491018, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1668, "step": 3097 }, { "epoch": 4.637724550898204, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1624, "step": 3098 }, { "epoch": 4.639221556886228, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1661, "step": 3099 }, { "epoch": 4.640718562874252, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1693, "step": 3100 }, { "epoch": 4.642215568862276, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1591, "step": 3101 }, { "epoch": 4.6437125748503, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3102 }, { "epoch": 4.645209580838324, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1666, "step": 3103 }, { "epoch": 4.6467065868263475, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1674, "step": 3104 }, { "epoch": 4.6482035928143715, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1692, "step": 3105 }, { "epoch": 4.649700598802395, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1646, "step": 3106 }, { "epoch": 4.651197604790419, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.162, "step": 3107 }, { "epoch": 4.652694610778443, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1639, "step": 3108 }, { "epoch": 4.654191616766467, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1649, "step": 3109 }, { "epoch": 4.655688622754491, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1653, "step": 3110 }, { "epoch": 4.657185628742515, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1706, "step": 3111 }, { "epoch": 4.658682634730539, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1681, "step": 3112 }, { "epoch": 4.660179640718563, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1627, "step": 3113 }, { "epoch": 4.661676646706587, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1665, "step": 3114 }, { "epoch": 4.663173652694611, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1681, "step": 3115 }, { "epoch": 4.664670658682635, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.162, "step": 3116 }, { "epoch": 4.666167664670659, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1684, "step": 3117 }, { "epoch": 4.667664670658683, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1635, "step": 3118 }, { "epoch": 4.669161676646707, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1626, "step": 3119 }, { "epoch": 4.6706586826347305, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1644, "step": 3120 }, { "epoch": 4.672155688622754, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1724, "step": 3121 }, { "epoch": 4.673652694610778, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1717, "step": 3122 }, { "epoch": 4.675149700598802, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1676, "step": 3123 }, { "epoch": 4.676646706586826, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1738, "step": 3124 }, { "epoch": 4.67814371257485, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1696, "step": 3125 }, { "epoch": 4.679640718562874, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.169, "step": 3126 }, { "epoch": 4.681137724550898, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3127 }, { "epoch": 4.682634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1707, "step": 3128 }, { "epoch": 4.684131736526946, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1588, "step": 3129 }, { "epoch": 4.68562874251497, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.168, "step": 3130 }, { "epoch": 4.687125748502994, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1681, "step": 3131 }, { "epoch": 4.688622754491018, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1706, "step": 3132 }, { "epoch": 4.690119760479042, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1674, "step": 3133 }, { "epoch": 4.691616766467066, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1641, "step": 3134 }, { "epoch": 4.6931137724550895, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3135 }, { "epoch": 4.6946107784431135, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3136 }, { "epoch": 4.696107784431137, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1751, "step": 3137 }, { "epoch": 4.697604790419161, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1675, "step": 3138 }, { "epoch": 4.699101796407185, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1548, "step": 3139 }, { "epoch": 4.700598802395209, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1629, "step": 3140 }, { "epoch": 4.702095808383233, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1635, "step": 3141 }, { "epoch": 4.703592814371257, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3142 }, { "epoch": 4.705089820359282, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1619, "step": 3143 }, { "epoch": 4.706586826347305, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1636, "step": 3144 }, { "epoch": 4.70808383233533, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1755, "step": 3145 }, { "epoch": 4.709580838323353, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1597, "step": 3146 }, { "epoch": 4.711077844311378, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1711, "step": 3147 }, { "epoch": 4.712574850299401, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1664, "step": 3148 }, { "epoch": 4.7140718562874255, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1692, "step": 3149 }, { "epoch": 4.7155688622754495, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.161, "step": 3150 }, { "epoch": 4.717065868263473, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1669, "step": 3151 }, { "epoch": 4.718562874251497, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1654, "step": 3152 }, { "epoch": 4.720059880239521, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1579, "step": 3153 }, { "epoch": 4.721556886227545, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3154 }, { "epoch": 4.723053892215569, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1585, "step": 3155 }, { "epoch": 4.724550898203593, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.165, "step": 3156 }, { "epoch": 4.726047904191617, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1652, "step": 3157 }, { "epoch": 4.727544910179641, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1679, "step": 3158 }, { "epoch": 4.729041916167665, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1629, "step": 3159 }, { "epoch": 4.730538922155689, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1618, "step": 3160 }, { "epoch": 4.732035928143713, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1634, "step": 3161 }, { "epoch": 4.733532934131737, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1709, "step": 3162 }, { "epoch": 4.735029940119761, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1709, "step": 3163 }, { "epoch": 4.736526946107785, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3164 }, { "epoch": 4.7380239520958085, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1702, "step": 3165 }, { "epoch": 4.7395209580838324, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.168, "step": 3166 }, { "epoch": 4.741017964071856, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1602, "step": 3167 }, { "epoch": 4.74251497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1724, "step": 3168 }, { "epoch": 4.744011976047904, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1743, "step": 3169 }, { "epoch": 4.745508982035928, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1657, "step": 3170 }, { "epoch": 4.747005988023952, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1656, "step": 3171 }, { "epoch": 4.748502994011976, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3172 }, { "epoch": 4.75, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1656, "step": 3173 }, { "epoch": 4.751497005988024, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1684, "step": 3174 }, { "epoch": 4.752994011976048, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1633, "step": 3175 }, { "epoch": 4.754491017964072, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1675, "step": 3176 }, { "epoch": 4.755988023952096, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.168, "step": 3177 }, { "epoch": 4.75748502994012, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1658, "step": 3178 }, { "epoch": 4.758982035928144, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1692, "step": 3179 }, { "epoch": 4.7604790419161676, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1705, "step": 3180 }, { "epoch": 4.7619760479041915, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1725, "step": 3181 }, { "epoch": 4.763473053892215, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1713, "step": 3182 }, { "epoch": 4.764970059880239, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1662, "step": 3183 }, { "epoch": 4.766467065868263, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1702, "step": 3184 }, { "epoch": 4.767964071856287, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1666, "step": 3185 }, { "epoch": 4.769461077844311, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1624, "step": 3186 }, { "epoch": 4.770958083832335, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1699, "step": 3187 }, { "epoch": 4.772455089820359, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1688, "step": 3188 }, { "epoch": 4.773952095808383, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.164, "step": 3189 }, { "epoch": 4.775449101796407, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1676, "step": 3190 }, { "epoch": 4.776946107784431, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1743, "step": 3191 }, { "epoch": 4.778443113772455, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1705, "step": 3192 }, { "epoch": 4.779940119760479, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1645, "step": 3193 }, { "epoch": 4.781437125748503, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1635, "step": 3194 }, { "epoch": 4.782934131736527, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1635, "step": 3195 }, { "epoch": 4.7844311377245505, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1697, "step": 3196 }, { "epoch": 4.7859281437125745, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1662, "step": 3197 }, { "epoch": 4.787425149700599, "grad_norm": 0.1630859375, "learning_rate": 0.0008, "loss": 1.1712, "step": 3198 }, { "epoch": 4.788922155688622, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1702, "step": 3199 }, { "epoch": 4.790419161676647, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1625, "step": 3200 }, { "epoch": 4.79191616766467, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1623, "step": 3201 }, { "epoch": 4.793413173652695, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1633, "step": 3202 }, { "epoch": 4.794910179640718, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.159, "step": 3203 }, { "epoch": 4.796407185628743, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1644, "step": 3204 }, { "epoch": 4.797904191616767, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1658, "step": 3205 }, { "epoch": 4.799401197604791, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1697, "step": 3206 }, { "epoch": 4.800898203592815, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.165, "step": 3207 }, { "epoch": 4.802395209580839, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1577, "step": 3208 }, { "epoch": 4.803892215568863, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.167, "step": 3209 }, { "epoch": 4.8053892215568865, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1696, "step": 3210 }, { "epoch": 4.8068862275449105, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1668, "step": 3211 }, { "epoch": 4.808383233532934, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1681, "step": 3212 }, { "epoch": 4.809880239520958, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1683, "step": 3213 }, { "epoch": 4.811377245508982, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1688, "step": 3214 }, { "epoch": 4.812874251497006, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3215 }, { "epoch": 4.81437125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1653, "step": 3216 }, { "epoch": 4.815868263473054, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1656, "step": 3217 }, { "epoch": 4.817365269461078, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1673, "step": 3218 }, { "epoch": 4.818862275449102, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1707, "step": 3219 }, { "epoch": 4.820359281437126, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1685, "step": 3220 }, { "epoch": 4.82185628742515, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1656, "step": 3221 }, { "epoch": 4.823353293413174, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1634, "step": 3222 }, { "epoch": 4.824850299401198, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1673, "step": 3223 }, { "epoch": 4.826347305389222, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.161, "step": 3224 }, { "epoch": 4.827844311377246, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.162, "step": 3225 }, { "epoch": 4.8293413173652695, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1656, "step": 3226 }, { "epoch": 4.830838323353293, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1593, "step": 3227 }, { "epoch": 4.832335329341317, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1696, "step": 3228 }, { "epoch": 4.833832335329341, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1686, "step": 3229 }, { "epoch": 4.835329341317365, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1619, "step": 3230 }, { "epoch": 4.836826347305389, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.162, "step": 3231 }, { "epoch": 4.838323353293413, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1608, "step": 3232 }, { "epoch": 4.839820359281437, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1682, "step": 3233 }, { "epoch": 4.841317365269461, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1605, "step": 3234 }, { "epoch": 4.842814371257485, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1587, "step": 3235 }, { "epoch": 4.844311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1689, "step": 3236 }, { "epoch": 4.845808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1647, "step": 3237 }, { "epoch": 4.847305389221557, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1666, "step": 3238 }, { "epoch": 4.848802395209581, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1671, "step": 3239 }, { "epoch": 4.850299401197605, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.165, "step": 3240 }, { "epoch": 4.8517964071856285, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1635, "step": 3241 }, { "epoch": 4.8532934131736525, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1659, "step": 3242 }, { "epoch": 4.854790419161676, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1631, "step": 3243 }, { "epoch": 4.8562874251497, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1643, "step": 3244 }, { "epoch": 4.857784431137724, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1673, "step": 3245 }, { "epoch": 4.859281437125748, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.156, "step": 3246 }, { "epoch": 4.860778443113772, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.161, "step": 3247 }, { "epoch": 4.862275449101796, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1672, "step": 3248 }, { "epoch": 4.86377245508982, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1671, "step": 3249 }, { "epoch": 4.865269461077844, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1615, "step": 3250 }, { "epoch": 4.866766467065868, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1596, "step": 3251 }, { "epoch": 4.868263473053892, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1671, "step": 3252 }, { "epoch": 4.869760479041916, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1658, "step": 3253 }, { "epoch": 4.87125748502994, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1654, "step": 3254 }, { "epoch": 4.8727544910179645, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1701, "step": 3255 }, { "epoch": 4.874251497005988, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.173, "step": 3256 }, { "epoch": 4.875748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1671, "step": 3257 }, { "epoch": 4.8772455089820355, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3258 }, { "epoch": 4.87874251497006, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.166, "step": 3259 }, { "epoch": 4.880239520958084, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1652, "step": 3260 }, { "epoch": 4.881736526946108, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.168, "step": 3261 }, { "epoch": 4.883233532934132, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1606, "step": 3262 }, { "epoch": 4.884730538922156, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1633, "step": 3263 }, { "epoch": 4.88622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3264 }, { "epoch": 4.887724550898204, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1683, "step": 3265 }, { "epoch": 4.889221556886228, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1594, "step": 3266 }, { "epoch": 4.890718562874252, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1672, "step": 3267 }, { "epoch": 4.892215568862276, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1696, "step": 3268 }, { "epoch": 4.8937125748503, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1656, "step": 3269 }, { "epoch": 4.895209580838324, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1642, "step": 3270 }, { "epoch": 4.8967065868263475, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1682, "step": 3271 }, { "epoch": 4.8982035928143715, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1652, "step": 3272 }, { "epoch": 4.899700598802395, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.159, "step": 3273 }, { "epoch": 4.901197604790419, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1628, "step": 3274 }, { "epoch": 4.902694610778443, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1656, "step": 3275 }, { "epoch": 4.904191616766467, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.162, "step": 3276 }, { "epoch": 4.905688622754491, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1629, "step": 3277 }, { "epoch": 4.907185628742515, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1627, "step": 3278 }, { "epoch": 4.908682634730539, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1672, "step": 3279 }, { "epoch": 4.910179640718563, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1626, "step": 3280 }, { "epoch": 4.911676646706587, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1627, "step": 3281 }, { "epoch": 4.913173652694611, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1646, "step": 3282 }, { "epoch": 4.914670658682635, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1653, "step": 3283 }, { "epoch": 4.916167664670659, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1558, "step": 3284 }, { "epoch": 4.917664670658683, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1644, "step": 3285 }, { "epoch": 4.919161676646707, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.174, "step": 3286 }, { "epoch": 4.9206586826347305, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1694, "step": 3287 }, { "epoch": 4.922155688622754, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1636, "step": 3288 }, { "epoch": 4.923652694610778, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1747, "step": 3289 }, { "epoch": 4.925149700598802, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1673, "step": 3290 }, { "epoch": 4.926646706586826, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1656, "step": 3291 }, { "epoch": 4.92814371257485, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.16, "step": 3292 }, { "epoch": 4.929640718562874, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1713, "step": 3293 }, { "epoch": 4.931137724550898, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1664, "step": 3294 }, { "epoch": 4.932634730538922, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1691, "step": 3295 }, { "epoch": 4.934131736526946, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1655, "step": 3296 }, { "epoch": 4.93562874251497, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3297 }, { "epoch": 4.937125748502994, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1657, "step": 3298 }, { "epoch": 4.938622754491018, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1668, "step": 3299 }, { "epoch": 4.940119760479042, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1704, "step": 3300 }, { "epoch": 4.941616766467066, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.166, "step": 3301 }, { "epoch": 4.9431137724550895, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1714, "step": 3302 }, { "epoch": 4.9446107784431135, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1577, "step": 3303 }, { "epoch": 4.946107784431137, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1643, "step": 3304 }, { "epoch": 4.947604790419161, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1711, "step": 3305 }, { "epoch": 4.949101796407185, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.158, "step": 3306 }, { "epoch": 4.950598802395209, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3307 }, { "epoch": 4.952095808383233, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1611, "step": 3308 }, { "epoch": 4.953592814371257, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.173, "step": 3309 }, { "epoch": 4.955089820359282, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1589, "step": 3310 }, { "epoch": 4.956586826347305, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.164, "step": 3311 }, { "epoch": 4.95808383233533, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1672, "step": 3312 }, { "epoch": 4.959580838323353, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1626, "step": 3313 }, { "epoch": 4.961077844311378, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1711, "step": 3314 }, { "epoch": 4.962574850299401, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1571, "step": 3315 }, { "epoch": 4.9640718562874255, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.169, "step": 3316 }, { "epoch": 4.9655688622754495, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.17, "step": 3317 }, { "epoch": 4.967065868263473, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1691, "step": 3318 }, { "epoch": 4.968562874251497, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1671, "step": 3319 }, { "epoch": 4.970059880239521, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1712, "step": 3320 }, { "epoch": 4.971556886227545, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1657, "step": 3321 }, { "epoch": 4.973053892215569, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1502, "step": 3322 }, { "epoch": 4.974550898203593, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1623, "step": 3323 }, { "epoch": 4.976047904191617, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1674, "step": 3324 }, { "epoch": 4.977544910179641, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1745, "step": 3325 }, { "epoch": 4.979041916167665, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1629, "step": 3326 }, { "epoch": 4.980538922155689, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1611, "step": 3327 }, { "epoch": 4.982035928143713, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.165, "step": 3328 }, { "epoch": 4.983532934131737, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1593, "step": 3329 }, { "epoch": 4.985029940119761, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.166, "step": 3330 }, { "epoch": 4.986526946107785, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1601, "step": 3331 }, { "epoch": 4.9880239520958085, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1646, "step": 3332 }, { "epoch": 4.9895209580838324, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1647, "step": 3333 }, { "epoch": 4.991017964071856, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.166, "step": 3334 }, { "epoch": 4.99251497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1666, "step": 3335 }, { "epoch": 4.994011976047904, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.164, "step": 3336 }, { "epoch": 4.995508982035928, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1696, "step": 3337 }, { "epoch": 4.997005988023952, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1686, "step": 3338 }, { "epoch": 4.998502994011976, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1641, "step": 3339 }, { "epoch": 5.0, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1668, "step": 3340 }, { "epoch": 5.001497005988024, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1582, "step": 3341 }, { "epoch": 5.002994011976048, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1601, "step": 3342 }, { "epoch": 5.004491017964072, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1642, "step": 3343 }, { "epoch": 5.005988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1676, "step": 3344 }, { "epoch": 5.00748502994012, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1717, "step": 3345 }, { "epoch": 5.008982035928144, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1622, "step": 3346 }, { "epoch": 5.0104790419161676, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1643, "step": 3347 }, { "epoch": 5.0119760479041915, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1692, "step": 3348 }, { "epoch": 5.013473053892215, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1638, "step": 3349 }, { "epoch": 5.014970059880239, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1645, "step": 3350 }, { "epoch": 5.016467065868263, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1701, "step": 3351 }, { "epoch": 5.017964071856287, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.157, "step": 3352 }, { "epoch": 5.019461077844311, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1619, "step": 3353 }, { "epoch": 5.020958083832335, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3354 }, { "epoch": 5.022455089820359, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1573, "step": 3355 }, { "epoch": 5.023952095808383, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1762, "step": 3356 }, { "epoch": 5.025449101796407, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1698, "step": 3357 }, { "epoch": 5.026946107784431, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.161, "step": 3358 }, { "epoch": 5.028443113772455, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1639, "step": 3359 }, { "epoch": 5.029940119760479, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.168, "step": 3360 }, { "epoch": 5.031437125748503, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1661, "step": 3361 }, { "epoch": 5.032934131736527, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1606, "step": 3362 }, { "epoch": 5.0344311377245505, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1598, "step": 3363 }, { "epoch": 5.0359281437125745, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3364 }, { "epoch": 5.037425149700598, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1632, "step": 3365 }, { "epoch": 5.038922155688622, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1572, "step": 3366 }, { "epoch": 5.040419161676646, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1713, "step": 3367 }, { "epoch": 5.04191616766467, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.166, "step": 3368 }, { "epoch": 5.043413173652695, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1691, "step": 3369 }, { "epoch": 5.044910179640719, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1717, "step": 3370 }, { "epoch": 5.046407185628743, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1659, "step": 3371 }, { "epoch": 5.047904191616767, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1641, "step": 3372 }, { "epoch": 5.049401197604791, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1662, "step": 3373 }, { "epoch": 5.050898203592815, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1693, "step": 3374 }, { "epoch": 5.052395209580839, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1679, "step": 3375 }, { "epoch": 5.053892215568863, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1643, "step": 3376 }, { "epoch": 5.0553892215568865, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3377 }, { "epoch": 5.0568862275449105, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1619, "step": 3378 }, { "epoch": 5.058383233532934, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3379 }, { "epoch": 5.059880239520958, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1594, "step": 3380 }, { "epoch": 5.061377245508982, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1635, "step": 3381 }, { "epoch": 5.062874251497006, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1653, "step": 3382 }, { "epoch": 5.06437125748503, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1686, "step": 3383 }, { "epoch": 5.065868263473054, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1659, "step": 3384 }, { "epoch": 5.067365269461078, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1645, "step": 3385 }, { "epoch": 5.068862275449102, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1551, "step": 3386 }, { "epoch": 5.070359281437126, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1613, "step": 3387 }, { "epoch": 5.07185628742515, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1648, "step": 3388 }, { "epoch": 5.073353293413174, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1601, "step": 3389 }, { "epoch": 5.074850299401198, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1708, "step": 3390 }, { "epoch": 5.076347305389222, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1598, "step": 3391 }, { "epoch": 5.077844311377246, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1688, "step": 3392 }, { "epoch": 5.0793413173652695, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1632, "step": 3393 }, { "epoch": 5.080838323353293, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3394 }, { "epoch": 5.082335329341317, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.159, "step": 3395 }, { "epoch": 5.083832335329341, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.164, "step": 3396 }, { "epoch": 5.085329341317365, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1611, "step": 3397 }, { "epoch": 5.086826347305389, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1651, "step": 3398 }, { "epoch": 5.088323353293413, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1558, "step": 3399 }, { "epoch": 5.089820359281437, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1615, "step": 3400 }, { "epoch": 5.091317365269461, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1693, "step": 3401 }, { "epoch": 5.092814371257485, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.168, "step": 3402 }, { "epoch": 5.094311377245509, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1744, "step": 3403 }, { "epoch": 5.095808383233533, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1562, "step": 3404 }, { "epoch": 5.097305389221557, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1656, "step": 3405 }, { "epoch": 5.098802395209581, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1655, "step": 3406 }, { "epoch": 5.100299401197605, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1549, "step": 3407 }, { "epoch": 5.1017964071856285, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.168, "step": 3408 }, { "epoch": 5.1032934131736525, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1604, "step": 3409 }, { "epoch": 5.104790419161676, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1647, "step": 3410 }, { "epoch": 5.1062874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1709, "step": 3411 }, { "epoch": 5.107784431137724, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1618, "step": 3412 }, { "epoch": 5.109281437125748, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1612, "step": 3413 }, { "epoch": 5.110778443113772, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1657, "step": 3414 }, { "epoch": 5.112275449101796, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1679, "step": 3415 }, { "epoch": 5.11377245508982, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.165, "step": 3416 }, { "epoch": 5.115269461077844, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1633, "step": 3417 }, { "epoch": 5.116766467065868, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1663, "step": 3418 }, { "epoch": 5.118263473053892, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1645, "step": 3419 }, { "epoch": 5.119760479041916, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1636, "step": 3420 }, { "epoch": 5.12125748502994, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1633, "step": 3421 }, { "epoch": 5.122754491017964, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1624, "step": 3422 }, { "epoch": 5.124251497005988, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1629, "step": 3423 }, { "epoch": 5.125748502994012, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1685, "step": 3424 }, { "epoch": 5.127245508982036, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1576, "step": 3425 }, { "epoch": 5.12874251497006, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1643, "step": 3426 }, { "epoch": 5.130239520958084, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1566, "step": 3427 }, { "epoch": 5.131736526946108, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1652, "step": 3428 }, { "epoch": 5.133233532934132, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1653, "step": 3429 }, { "epoch": 5.134730538922156, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1596, "step": 3430 }, { "epoch": 5.13622754491018, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1574, "step": 3431 }, { "epoch": 5.137724550898204, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1614, "step": 3432 }, { "epoch": 5.139221556886228, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1617, "step": 3433 }, { "epoch": 5.140718562874252, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1699, "step": 3434 }, { "epoch": 5.142215568862276, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1679, "step": 3435 }, { "epoch": 5.1437125748503, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1617, "step": 3436 }, { "epoch": 5.145209580838324, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1602, "step": 3437 }, { "epoch": 5.1467065868263475, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1669, "step": 3438 }, { "epoch": 5.1482035928143715, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1605, "step": 3439 }, { "epoch": 5.149700598802395, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1646, "step": 3440 }, { "epoch": 5.151197604790419, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3441 }, { "epoch": 5.152694610778443, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1648, "step": 3442 }, { "epoch": 5.154191616766467, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1627, "step": 3443 }, { "epoch": 5.155688622754491, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1548, "step": 3444 }, { "epoch": 5.157185628742515, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1664, "step": 3445 }, { "epoch": 5.158682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1631, "step": 3446 }, { "epoch": 5.160179640718563, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1651, "step": 3447 }, { "epoch": 5.161676646706587, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1661, "step": 3448 }, { "epoch": 5.163173652694611, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1638, "step": 3449 }, { "epoch": 5.164670658682635, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1748, "step": 3450 }, { "epoch": 5.166167664670659, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1658, "step": 3451 }, { "epoch": 5.167664670658683, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1654, "step": 3452 }, { "epoch": 5.169161676646707, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1674, "step": 3453 }, { "epoch": 5.1706586826347305, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1601, "step": 3454 }, { "epoch": 5.172155688622754, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1613, "step": 3455 }, { "epoch": 5.173652694610778, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.162, "step": 3456 }, { "epoch": 5.175149700598802, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.169, "step": 3457 }, { "epoch": 5.176646706586826, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1645, "step": 3458 }, { "epoch": 5.17814371257485, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.17, "step": 3459 }, { "epoch": 5.179640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1683, "step": 3460 }, { "epoch": 5.181137724550898, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1598, "step": 3461 }, { "epoch": 5.182634730538922, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1662, "step": 3462 }, { "epoch": 5.184131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1665, "step": 3463 }, { "epoch": 5.18562874251497, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1683, "step": 3464 }, { "epoch": 5.187125748502994, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1614, "step": 3465 }, { "epoch": 5.188622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1681, "step": 3466 }, { "epoch": 5.190119760479042, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1698, "step": 3467 }, { "epoch": 5.191616766467066, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1688, "step": 3468 }, { "epoch": 5.1931137724550895, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1646, "step": 3469 }, { "epoch": 5.1946107784431135, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1642, "step": 3470 }, { "epoch": 5.196107784431137, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1675, "step": 3471 }, { "epoch": 5.197604790419161, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1649, "step": 3472 }, { "epoch": 5.199101796407185, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1636, "step": 3473 }, { "epoch": 5.200598802395209, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.161, "step": 3474 }, { "epoch": 5.202095808383233, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1595, "step": 3475 }, { "epoch": 5.203592814371257, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1689, "step": 3476 }, { "epoch": 5.205089820359281, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1621, "step": 3477 }, { "epoch": 5.206586826347305, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1589, "step": 3478 }, { "epoch": 5.20808383233533, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1658, "step": 3479 }, { "epoch": 5.209580838323353, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1533, "step": 3480 }, { "epoch": 5.211077844311378, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1597, "step": 3481 }, { "epoch": 5.212574850299402, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1648, "step": 3482 }, { "epoch": 5.2140718562874255, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1689, "step": 3483 }, { "epoch": 5.2155688622754495, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1699, "step": 3484 }, { "epoch": 5.217065868263473, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1702, "step": 3485 }, { "epoch": 5.218562874251497, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.161, "step": 3486 }, { "epoch": 5.220059880239521, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.164, "step": 3487 }, { "epoch": 5.221556886227545, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1633, "step": 3488 }, { "epoch": 5.223053892215569, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1556, "step": 3489 }, { "epoch": 5.224550898203593, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1647, "step": 3490 }, { "epoch": 5.226047904191617, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1664, "step": 3491 }, { "epoch": 5.227544910179641, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1613, "step": 3492 }, { "epoch": 5.229041916167665, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1663, "step": 3493 }, { "epoch": 5.230538922155689, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1717, "step": 3494 }, { "epoch": 5.232035928143713, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1624, "step": 3495 }, { "epoch": 5.233532934131737, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1604, "step": 3496 }, { "epoch": 5.235029940119761, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1654, "step": 3497 }, { "epoch": 5.236526946107785, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1693, "step": 3498 }, { "epoch": 5.2380239520958085, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1629, "step": 3499 }, { "epoch": 5.2395209580838324, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1571, "step": 3500 }, { "epoch": 5.241017964071856, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1609, "step": 3501 }, { "epoch": 5.24251497005988, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1646, "step": 3502 }, { "epoch": 5.244011976047904, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.168, "step": 3503 }, { "epoch": 5.245508982035928, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1593, "step": 3504 }, { "epoch": 5.247005988023952, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1621, "step": 3505 }, { "epoch": 5.248502994011976, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1685, "step": 3506 }, { "epoch": 5.25, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1639, "step": 3507 }, { "epoch": 5.251497005988024, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.162, "step": 3508 }, { "epoch": 5.252994011976048, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1606, "step": 3509 }, { "epoch": 5.254491017964072, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1686, "step": 3510 }, { "epoch": 5.255988023952096, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1604, "step": 3511 }, { "epoch": 5.25748502994012, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1664, "step": 3512 }, { "epoch": 5.258982035928144, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1625, "step": 3513 }, { "epoch": 5.2604790419161676, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1613, "step": 3514 }, { "epoch": 5.2619760479041915, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1609, "step": 3515 }, { "epoch": 5.263473053892215, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.162, "step": 3516 }, { "epoch": 5.264970059880239, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1599, "step": 3517 }, { "epoch": 5.266467065868263, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1589, "step": 3518 }, { "epoch": 5.267964071856287, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1591, "step": 3519 }, { "epoch": 5.269461077844311, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.16, "step": 3520 }, { "epoch": 5.270958083832335, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1616, "step": 3521 }, { "epoch": 5.272455089820359, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1593, "step": 3522 }, { "epoch": 5.273952095808383, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1685, "step": 3523 }, { "epoch": 5.275449101796407, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1647, "step": 3524 }, { "epoch": 5.276946107784431, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.167, "step": 3525 }, { "epoch": 5.278443113772455, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1673, "step": 3526 }, { "epoch": 5.279940119760479, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.167, "step": 3527 }, { "epoch": 5.281437125748503, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3528 }, { "epoch": 5.282934131736527, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1609, "step": 3529 }, { "epoch": 5.2844311377245505, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1625, "step": 3530 }, { "epoch": 5.2859281437125745, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1706, "step": 3531 }, { "epoch": 5.287425149700598, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1591, "step": 3532 }, { "epoch": 5.288922155688622, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1649, "step": 3533 }, { "epoch": 5.290419161676647, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1608, "step": 3534 }, { "epoch": 5.29191616766467, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1651, "step": 3535 }, { "epoch": 5.293413173652695, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1627, "step": 3536 }, { "epoch": 5.294910179640719, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1631, "step": 3537 }, { "epoch": 5.296407185628743, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1578, "step": 3538 }, { "epoch": 5.297904191616767, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1618, "step": 3539 }, { "epoch": 5.299401197604791, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1603, "step": 3540 }, { "epoch": 5.300898203592815, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1582, "step": 3541 }, { "epoch": 5.302395209580839, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1542, "step": 3542 }, { "epoch": 5.303892215568863, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1653, "step": 3543 }, { "epoch": 5.3053892215568865, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1629, "step": 3544 }, { "epoch": 5.3068862275449105, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1613, "step": 3545 }, { "epoch": 5.308383233532934, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1638, "step": 3546 }, { "epoch": 5.309880239520958, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1701, "step": 3547 }, { "epoch": 5.311377245508982, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1565, "step": 3548 }, { "epoch": 5.312874251497006, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3549 }, { "epoch": 5.31437125748503, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1564, "step": 3550 }, { "epoch": 5.315868263473054, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1608, "step": 3551 }, { "epoch": 5.317365269461078, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1611, "step": 3552 }, { "epoch": 5.318862275449102, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.164, "step": 3553 }, { "epoch": 5.320359281437126, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.151, "step": 3554 }, { "epoch": 5.32185628742515, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.161, "step": 3555 }, { "epoch": 5.323353293413174, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1612, "step": 3556 }, { "epoch": 5.324850299401198, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1654, "step": 3557 }, { "epoch": 5.326347305389222, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1617, "step": 3558 }, { "epoch": 5.327844311377246, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1703, "step": 3559 }, { "epoch": 5.3293413173652695, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.165, "step": 3560 }, { "epoch": 5.330838323353293, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1632, "step": 3561 }, { "epoch": 5.332335329341317, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1688, "step": 3562 }, { "epoch": 5.333832335329341, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1617, "step": 3563 }, { "epoch": 5.335329341317365, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1612, "step": 3564 }, { "epoch": 5.336826347305389, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1638, "step": 3565 }, { "epoch": 5.338323353293413, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1617, "step": 3566 }, { "epoch": 5.339820359281437, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1609, "step": 3567 }, { "epoch": 5.341317365269461, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1656, "step": 3568 }, { "epoch": 5.342814371257485, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1664, "step": 3569 }, { "epoch": 5.344311377245509, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1604, "step": 3570 }, { "epoch": 5.345808383233533, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1661, "step": 3571 }, { "epoch": 5.347305389221557, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1591, "step": 3572 }, { "epoch": 5.348802395209581, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1593, "step": 3573 }, { "epoch": 5.350299401197605, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1634, "step": 3574 }, { "epoch": 5.3517964071856285, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1578, "step": 3575 }, { "epoch": 5.3532934131736525, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1638, "step": 3576 }, { "epoch": 5.354790419161676, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1669, "step": 3577 }, { "epoch": 5.3562874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1682, "step": 3578 }, { "epoch": 5.357784431137724, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1599, "step": 3579 }, { "epoch": 5.359281437125748, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3580 }, { "epoch": 5.360778443113772, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1601, "step": 3581 }, { "epoch": 5.362275449101796, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1634, "step": 3582 }, { "epoch": 5.36377245508982, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1597, "step": 3583 }, { "epoch": 5.365269461077844, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.164, "step": 3584 }, { "epoch": 5.366766467065868, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1637, "step": 3585 }, { "epoch": 5.368263473053892, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1577, "step": 3586 }, { "epoch": 5.369760479041916, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3587 }, { "epoch": 5.37125748502994, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.16, "step": 3588 }, { "epoch": 5.3727544910179645, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1668, "step": 3589 }, { "epoch": 5.374251497005988, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1652, "step": 3590 }, { "epoch": 5.375748502994012, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1658, "step": 3591 }, { "epoch": 5.3772455089820355, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1648, "step": 3592 }, { "epoch": 5.37874251497006, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1591, "step": 3593 }, { "epoch": 5.380239520958084, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3594 }, { "epoch": 5.381736526946108, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1676, "step": 3595 }, { "epoch": 5.383233532934132, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.166, "step": 3596 }, { "epoch": 5.384730538922156, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1707, "step": 3597 }, { "epoch": 5.38622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1614, "step": 3598 }, { "epoch": 5.387724550898204, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1671, "step": 3599 }, { "epoch": 5.389221556886228, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1648, "step": 3600 }, { "epoch": 5.390718562874252, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1592, "step": 3601 }, { "epoch": 5.392215568862276, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1697, "step": 3602 }, { "epoch": 5.3937125748503, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1602, "step": 3603 }, { "epoch": 5.395209580838324, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1581, "step": 3604 }, { "epoch": 5.3967065868263475, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1601, "step": 3605 }, { "epoch": 5.3982035928143715, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1603, "step": 3606 }, { "epoch": 5.399700598802395, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1604, "step": 3607 }, { "epoch": 5.401197604790419, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1654, "step": 3608 }, { "epoch": 5.402694610778443, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1605, "step": 3609 }, { "epoch": 5.404191616766467, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1627, "step": 3610 }, { "epoch": 5.405688622754491, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.161, "step": 3611 }, { "epoch": 5.407185628742515, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1615, "step": 3612 }, { "epoch": 5.408682634730539, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1618, "step": 3613 }, { "epoch": 5.410179640718563, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1614, "step": 3614 }, { "epoch": 5.411676646706587, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1614, "step": 3615 }, { "epoch": 5.413173652694611, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1651, "step": 3616 }, { "epoch": 5.414670658682635, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1678, "step": 3617 }, { "epoch": 5.416167664670659, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1681, "step": 3618 }, { "epoch": 5.417664670658683, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1568, "step": 3619 }, { "epoch": 5.419161676646707, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.1649, "step": 3620 }, { "epoch": 5.4206586826347305, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1609, "step": 3621 }, { "epoch": 5.422155688622754, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1628, "step": 3622 }, { "epoch": 5.423652694610778, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1583, "step": 3623 }, { "epoch": 5.425149700598802, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1661, "step": 3624 }, { "epoch": 5.426646706586826, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1665, "step": 3625 }, { "epoch": 5.42814371257485, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1563, "step": 3626 }, { "epoch": 5.429640718562874, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1615, "step": 3627 }, { "epoch": 5.431137724550898, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1607, "step": 3628 }, { "epoch": 5.432634730538922, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1594, "step": 3629 }, { "epoch": 5.434131736526946, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1639, "step": 3630 }, { "epoch": 5.43562874251497, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.164, "step": 3631 }, { "epoch": 5.437125748502994, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.1699, "step": 3632 }, { "epoch": 5.438622754491018, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1591, "step": 3633 }, { "epoch": 5.440119760479042, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1636, "step": 3634 }, { "epoch": 5.441616766467066, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1656, "step": 3635 }, { "epoch": 5.4431137724550895, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1653, "step": 3636 }, { "epoch": 5.4446107784431135, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1597, "step": 3637 }, { "epoch": 5.446107784431137, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1642, "step": 3638 }, { "epoch": 5.447604790419161, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1644, "step": 3639 }, { "epoch": 5.449101796407185, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1583, "step": 3640 }, { "epoch": 5.450598802395209, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1602, "step": 3641 }, { "epoch": 5.452095808383233, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1621, "step": 3642 }, { "epoch": 5.453592814371257, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1619, "step": 3643 }, { "epoch": 5.455089820359281, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1694, "step": 3644 }, { "epoch": 5.456586826347305, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.155, "step": 3645 }, { "epoch": 5.45808383233533, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1615, "step": 3646 }, { "epoch": 5.459580838323353, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1636, "step": 3647 }, { "epoch": 5.461077844311378, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1646, "step": 3648 }, { "epoch": 5.462574850299402, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1626, "step": 3649 }, { "epoch": 5.4640718562874255, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3650 }, { "epoch": 5.4655688622754495, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.1595, "step": 3651 }, { "epoch": 5.467065868263473, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1606, "step": 3652 }, { "epoch": 5.468562874251497, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1686, "step": 3653 }, { "epoch": 5.470059880239521, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1621, "step": 3654 }, { "epoch": 5.471556886227545, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1684, "step": 3655 }, { "epoch": 5.473053892215569, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1673, "step": 3656 }, { "epoch": 5.474550898203593, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1642, "step": 3657 }, { "epoch": 5.476047904191617, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1574, "step": 3658 }, { "epoch": 5.477544910179641, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.168, "step": 3659 }, { "epoch": 5.479041916167665, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1696, "step": 3660 }, { "epoch": 5.480538922155689, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1573, "step": 3661 }, { "epoch": 5.482035928143713, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1641, "step": 3662 }, { "epoch": 5.483532934131737, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1705, "step": 3663 }, { "epoch": 5.485029940119761, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1591, "step": 3664 }, { "epoch": 5.486526946107785, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3665 }, { "epoch": 5.4880239520958085, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1506, "step": 3666 }, { "epoch": 5.4895209580838324, "grad_norm": 0.1572265625, "learning_rate": 0.0008, "loss": 1.1628, "step": 3667 }, { "epoch": 5.491017964071856, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1562, "step": 3668 }, { "epoch": 5.49251497005988, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1635, "step": 3669 }, { "epoch": 5.494011976047904, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1597, "step": 3670 }, { "epoch": 5.495508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1623, "step": 3671 }, { "epoch": 5.497005988023952, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1664, "step": 3672 }, { "epoch": 5.498502994011976, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.168, "step": 3673 }, { "epoch": 5.5, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1535, "step": 3674 }, { "epoch": 5.501497005988024, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1642, "step": 3675 }, { "epoch": 5.502994011976048, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1645, "step": 3676 }, { "epoch": 5.504491017964072, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1649, "step": 3677 }, { "epoch": 5.505988023952096, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1598, "step": 3678 }, { "epoch": 5.50748502994012, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1593, "step": 3679 }, { "epoch": 5.508982035928144, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1621, "step": 3680 }, { "epoch": 5.5104790419161676, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1632, "step": 3681 }, { "epoch": 5.5119760479041915, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.163, "step": 3682 }, { "epoch": 5.513473053892215, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1634, "step": 3683 }, { "epoch": 5.514970059880239, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1671, "step": 3684 }, { "epoch": 5.516467065868263, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1605, "step": 3685 }, { "epoch": 5.517964071856287, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1554, "step": 3686 }, { "epoch": 5.519461077844311, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1644, "step": 3687 }, { "epoch": 5.520958083832335, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1652, "step": 3688 }, { "epoch": 5.522455089820359, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1689, "step": 3689 }, { "epoch": 5.523952095808383, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1651, "step": 3690 }, { "epoch": 5.525449101796407, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1607, "step": 3691 }, { "epoch": 5.526946107784431, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1604, "step": 3692 }, { "epoch": 5.528443113772455, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1669, "step": 3693 }, { "epoch": 5.529940119760479, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1605, "step": 3694 }, { "epoch": 5.531437125748503, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1616, "step": 3695 }, { "epoch": 5.532934131736527, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1616, "step": 3696 }, { "epoch": 5.5344311377245505, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1579, "step": 3697 }, { "epoch": 5.5359281437125745, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1638, "step": 3698 }, { "epoch": 5.537425149700599, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1664, "step": 3699 }, { "epoch": 5.538922155688622, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1693, "step": 3700 }, { "epoch": 5.540419161676647, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1677, "step": 3701 }, { "epoch": 5.54191616766467, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1621, "step": 3702 }, { "epoch": 5.543413173652695, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1655, "step": 3703 }, { "epoch": 5.544910179640718, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1678, "step": 3704 }, { "epoch": 5.546407185628743, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1631, "step": 3705 }, { "epoch": 5.547904191616767, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1542, "step": 3706 }, { "epoch": 5.549401197604791, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1527, "step": 3707 }, { "epoch": 5.550898203592815, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.1576, "step": 3708 }, { "epoch": 5.552395209580839, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1701, "step": 3709 }, { "epoch": 5.553892215568863, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.164, "step": 3710 }, { "epoch": 5.5553892215568865, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1592, "step": 3711 }, { "epoch": 5.5568862275449105, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1642, "step": 3712 }, { "epoch": 5.558383233532934, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.158, "step": 3713 }, { "epoch": 5.559880239520958, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1669, "step": 3714 }, { "epoch": 5.561377245508982, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1589, "step": 3715 }, { "epoch": 5.562874251497006, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1643, "step": 3716 }, { "epoch": 5.56437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3717 }, { "epoch": 5.565868263473054, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1606, "step": 3718 }, { "epoch": 5.567365269461078, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3719 }, { "epoch": 5.568862275449102, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1637, "step": 3720 }, { "epoch": 5.570359281437126, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1626, "step": 3721 }, { "epoch": 5.57185628742515, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1546, "step": 3722 }, { "epoch": 5.573353293413174, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1584, "step": 3723 }, { "epoch": 5.574850299401198, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1656, "step": 3724 }, { "epoch": 5.576347305389222, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1576, "step": 3725 }, { "epoch": 5.577844311377246, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1649, "step": 3726 }, { "epoch": 5.5793413173652695, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1622, "step": 3727 }, { "epoch": 5.580838323353293, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1682, "step": 3728 }, { "epoch": 5.582335329341317, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1683, "step": 3729 }, { "epoch": 5.583832335329341, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1581, "step": 3730 }, { "epoch": 5.585329341317365, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1588, "step": 3731 }, { "epoch": 5.586826347305389, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1562, "step": 3732 }, { "epoch": 5.588323353293413, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1583, "step": 3733 }, { "epoch": 5.589820359281437, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1668, "step": 3734 }, { "epoch": 5.591317365269461, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.156, "step": 3735 }, { "epoch": 5.592814371257485, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1615, "step": 3736 }, { "epoch": 5.594311377245509, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1632, "step": 3737 }, { "epoch": 5.595808383233533, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1571, "step": 3738 }, { "epoch": 5.597305389221557, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.156, "step": 3739 }, { "epoch": 5.598802395209581, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1601, "step": 3740 }, { "epoch": 5.600299401197605, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3741 }, { "epoch": 5.6017964071856285, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.164, "step": 3742 }, { "epoch": 5.6032934131736525, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1562, "step": 3743 }, { "epoch": 5.604790419161676, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1632, "step": 3744 }, { "epoch": 5.6062874251497, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1627, "step": 3745 }, { "epoch": 5.607784431137724, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1606, "step": 3746 }, { "epoch": 5.609281437125748, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1651, "step": 3747 }, { "epoch": 5.610778443113772, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.165, "step": 3748 }, { "epoch": 5.612275449101796, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1647, "step": 3749 }, { "epoch": 5.61377245508982, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1639, "step": 3750 }, { "epoch": 5.615269461077844, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1606, "step": 3751 }, { "epoch": 5.616766467065868, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1589, "step": 3752 }, { "epoch": 5.618263473053892, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1644, "step": 3753 }, { "epoch": 5.619760479041916, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1664, "step": 3754 }, { "epoch": 5.62125748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1665, "step": 3755 }, { "epoch": 5.6227544910179645, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1632, "step": 3756 }, { "epoch": 5.624251497005988, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1623, "step": 3757 }, { "epoch": 5.625748502994012, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1548, "step": 3758 }, { "epoch": 5.6272455089820355, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1658, "step": 3759 }, { "epoch": 5.62874251497006, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1544, "step": 3760 }, { "epoch": 5.630239520958084, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1625, "step": 3761 }, { "epoch": 5.631736526946108, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1593, "step": 3762 }, { "epoch": 5.633233532934132, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.159, "step": 3763 }, { "epoch": 5.634730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1634, "step": 3764 }, { "epoch": 5.63622754491018, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1559, "step": 3765 }, { "epoch": 5.637724550898204, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1667, "step": 3766 }, { "epoch": 5.639221556886228, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1633, "step": 3767 }, { "epoch": 5.640718562874252, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1593, "step": 3768 }, { "epoch": 5.642215568862276, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1652, "step": 3769 }, { "epoch": 5.6437125748503, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1628, "step": 3770 }, { "epoch": 5.645209580838324, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1647, "step": 3771 }, { "epoch": 5.6467065868263475, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1598, "step": 3772 }, { "epoch": 5.6482035928143715, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1643, "step": 3773 }, { "epoch": 5.649700598802395, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1568, "step": 3774 }, { "epoch": 5.651197604790419, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1623, "step": 3775 }, { "epoch": 5.652694610778443, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1624, "step": 3776 }, { "epoch": 5.654191616766467, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1646, "step": 3777 }, { "epoch": 5.655688622754491, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1565, "step": 3778 }, { "epoch": 5.657185628742515, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1603, "step": 3779 }, { "epoch": 5.658682634730539, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1607, "step": 3780 }, { "epoch": 5.660179640718563, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.159, "step": 3781 }, { "epoch": 5.661676646706587, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1696, "step": 3782 }, { "epoch": 5.663173652694611, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1634, "step": 3783 }, { "epoch": 5.664670658682635, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1617, "step": 3784 }, { "epoch": 5.666167664670659, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1559, "step": 3785 }, { "epoch": 5.667664670658683, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1617, "step": 3786 }, { "epoch": 5.669161676646707, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.156, "step": 3787 }, { "epoch": 5.6706586826347305, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1612, "step": 3788 }, { "epoch": 5.672155688622754, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1607, "step": 3789 }, { "epoch": 5.673652694610778, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1665, "step": 3790 }, { "epoch": 5.675149700598802, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1597, "step": 3791 }, { "epoch": 5.676646706586826, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1583, "step": 3792 }, { "epoch": 5.67814371257485, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1632, "step": 3793 }, { "epoch": 5.679640718562874, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1554, "step": 3794 }, { "epoch": 5.681137724550898, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1606, "step": 3795 }, { "epoch": 5.682634730538922, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1648, "step": 3796 }, { "epoch": 5.684131736526946, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1644, "step": 3797 }, { "epoch": 5.68562874251497, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1587, "step": 3798 }, { "epoch": 5.687125748502994, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1617, "step": 3799 }, { "epoch": 5.688622754491018, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.159, "step": 3800 }, { "epoch": 5.690119760479042, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1505, "step": 3801 }, { "epoch": 5.691616766467066, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1673, "step": 3802 }, { "epoch": 5.6931137724550895, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1621, "step": 3803 }, { "epoch": 5.6946107784431135, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.162, "step": 3804 }, { "epoch": 5.696107784431137, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1663, "step": 3805 }, { "epoch": 5.697604790419161, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1538, "step": 3806 }, { "epoch": 5.699101796407185, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1562, "step": 3807 }, { "epoch": 5.700598802395209, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1507, "step": 3808 }, { "epoch": 5.702095808383233, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1628, "step": 3809 }, { "epoch": 5.703592814371257, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1615, "step": 3810 }, { "epoch": 5.705089820359282, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1652, "step": 3811 }, { "epoch": 5.706586826347305, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1568, "step": 3812 }, { "epoch": 5.70808383233533, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1552, "step": 3813 }, { "epoch": 5.709580838323353, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1607, "step": 3814 }, { "epoch": 5.711077844311378, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.161, "step": 3815 }, { "epoch": 5.712574850299401, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.159, "step": 3816 }, { "epoch": 5.7140718562874255, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1604, "step": 3817 }, { "epoch": 5.7155688622754495, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1636, "step": 3818 }, { "epoch": 5.717065868263473, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1586, "step": 3819 }, { "epoch": 5.718562874251497, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1592, "step": 3820 }, { "epoch": 5.720059880239521, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1522, "step": 3821 }, { "epoch": 5.721556886227545, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1616, "step": 3822 }, { "epoch": 5.723053892215569, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1573, "step": 3823 }, { "epoch": 5.724550898203593, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1559, "step": 3824 }, { "epoch": 5.726047904191617, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1603, "step": 3825 }, { "epoch": 5.727544910179641, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1643, "step": 3826 }, { "epoch": 5.729041916167665, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1673, "step": 3827 }, { "epoch": 5.730538922155689, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1622, "step": 3828 }, { "epoch": 5.732035928143713, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1656, "step": 3829 }, { "epoch": 5.733532934131737, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1578, "step": 3830 }, { "epoch": 5.735029940119761, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1616, "step": 3831 }, { "epoch": 5.736526946107785, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1616, "step": 3832 }, { "epoch": 5.7380239520958085, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1657, "step": 3833 }, { "epoch": 5.7395209580838324, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1581, "step": 3834 }, { "epoch": 5.741017964071856, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1697, "step": 3835 }, { "epoch": 5.74251497005988, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1644, "step": 3836 }, { "epoch": 5.744011976047904, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.156, "step": 3837 }, { "epoch": 5.745508982035928, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1641, "step": 3838 }, { "epoch": 5.747005988023952, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1617, "step": 3839 }, { "epoch": 5.748502994011976, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.158, "step": 3840 }, { "epoch": 5.75, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1574, "step": 3841 }, { "epoch": 5.751497005988024, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1579, "step": 3842 }, { "epoch": 5.752994011976048, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3843 }, { "epoch": 5.754491017964072, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1687, "step": 3844 }, { "epoch": 5.755988023952096, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1596, "step": 3845 }, { "epoch": 5.75748502994012, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1679, "step": 3846 }, { "epoch": 5.758982035928144, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1653, "step": 3847 }, { "epoch": 5.7604790419161676, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1633, "step": 3848 }, { "epoch": 5.7619760479041915, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1629, "step": 3849 }, { "epoch": 5.763473053892215, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1679, "step": 3850 }, { "epoch": 5.764970059880239, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1606, "step": 3851 }, { "epoch": 5.766467065868263, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.164, "step": 3852 }, { "epoch": 5.767964071856287, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1652, "step": 3853 }, { "epoch": 5.769461077844311, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1572, "step": 3854 }, { "epoch": 5.770958083832335, "grad_norm": 0.1953125, "learning_rate": 0.0008, "loss": 1.1592, "step": 3855 }, { "epoch": 5.772455089820359, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1532, "step": 3856 }, { "epoch": 5.773952095808383, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1672, "step": 3857 }, { "epoch": 5.775449101796407, "grad_norm": 0.1787109375, "learning_rate": 0.0008, "loss": 1.1631, "step": 3858 }, { "epoch": 5.776946107784431, "grad_norm": 0.201171875, "learning_rate": 0.0008, "loss": 1.158, "step": 3859 }, { "epoch": 5.778443113772455, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.1605, "step": 3860 }, { "epoch": 5.779940119760479, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1592, "step": 3861 }, { "epoch": 5.781437125748503, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1546, "step": 3862 }, { "epoch": 5.782934131736527, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1582, "step": 3863 }, { "epoch": 5.7844311377245505, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1626, "step": 3864 }, { "epoch": 5.7859281437125745, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1553, "step": 3865 }, { "epoch": 5.787425149700599, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1667, "step": 3866 }, { "epoch": 5.788922155688622, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1556, "step": 3867 }, { "epoch": 5.790419161676647, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1645, "step": 3868 }, { "epoch": 5.79191616766467, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1613, "step": 3869 }, { "epoch": 5.793413173652695, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1553, "step": 3870 }, { "epoch": 5.794910179640718, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1563, "step": 3871 }, { "epoch": 5.796407185628743, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1603, "step": 3872 }, { "epoch": 5.797904191616767, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1715, "step": 3873 }, { "epoch": 5.799401197604791, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1677, "step": 3874 }, { "epoch": 5.800898203592815, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1706, "step": 3875 }, { "epoch": 5.802395209580839, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1504, "step": 3876 }, { "epoch": 5.803892215568863, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1596, "step": 3877 }, { "epoch": 5.8053892215568865, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1551, "step": 3878 }, { "epoch": 5.8068862275449105, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1641, "step": 3879 }, { "epoch": 5.808383233532934, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1584, "step": 3880 }, { "epoch": 5.809880239520958, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1567, "step": 3881 }, { "epoch": 5.811377245508982, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1616, "step": 3882 }, { "epoch": 5.812874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1648, "step": 3883 }, { "epoch": 5.81437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1638, "step": 3884 }, { "epoch": 5.815868263473054, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1607, "step": 3885 }, { "epoch": 5.817365269461078, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1576, "step": 3886 }, { "epoch": 5.818862275449102, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1538, "step": 3887 }, { "epoch": 5.820359281437126, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1573, "step": 3888 }, { "epoch": 5.82185628742515, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1583, "step": 3889 }, { "epoch": 5.823353293413174, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1534, "step": 3890 }, { "epoch": 5.824850299401198, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1625, "step": 3891 }, { "epoch": 5.826347305389222, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1598, "step": 3892 }, { "epoch": 5.827844311377246, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1574, "step": 3893 }, { "epoch": 5.8293413173652695, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1588, "step": 3894 }, { "epoch": 5.830838323353293, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1598, "step": 3895 }, { "epoch": 5.832335329341317, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1593, "step": 3896 }, { "epoch": 5.833832335329341, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1567, "step": 3897 }, { "epoch": 5.835329341317365, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1662, "step": 3898 }, { "epoch": 5.836826347305389, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1574, "step": 3899 }, { "epoch": 5.838323353293413, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3900 }, { "epoch": 5.839820359281437, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1617, "step": 3901 }, { "epoch": 5.841317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1585, "step": 3902 }, { "epoch": 5.842814371257485, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1551, "step": 3903 }, { "epoch": 5.844311377245509, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1671, "step": 3904 }, { "epoch": 5.845808383233533, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1593, "step": 3905 }, { "epoch": 5.847305389221557, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.155, "step": 3906 }, { "epoch": 5.848802395209581, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1632, "step": 3907 }, { "epoch": 5.850299401197605, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1621, "step": 3908 }, { "epoch": 5.8517964071856285, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1519, "step": 3909 }, { "epoch": 5.8532934131736525, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1551, "step": 3910 }, { "epoch": 5.854790419161676, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.162, "step": 3911 }, { "epoch": 5.8562874251497, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1597, "step": 3912 }, { "epoch": 5.857784431137724, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1573, "step": 3913 }, { "epoch": 5.859281437125748, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1592, "step": 3914 }, { "epoch": 5.860778443113772, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1652, "step": 3915 }, { "epoch": 5.862275449101796, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.152, "step": 3916 }, { "epoch": 5.86377245508982, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1704, "step": 3917 }, { "epoch": 5.865269461077844, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1555, "step": 3918 }, { "epoch": 5.866766467065868, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.16, "step": 3919 }, { "epoch": 5.868263473053892, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1634, "step": 3920 }, { "epoch": 5.869760479041916, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1541, "step": 3921 }, { "epoch": 5.87125748502994, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1633, "step": 3922 }, { "epoch": 5.8727544910179645, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1583, "step": 3923 }, { "epoch": 5.874251497005988, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1674, "step": 3924 }, { "epoch": 5.875748502994012, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1681, "step": 3925 }, { "epoch": 5.8772455089820355, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1585, "step": 3926 }, { "epoch": 5.87874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1612, "step": 3927 }, { "epoch": 5.880239520958084, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3928 }, { "epoch": 5.881736526946108, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.158, "step": 3929 }, { "epoch": 5.883233532934132, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1597, "step": 3930 }, { "epoch": 5.884730538922156, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1675, "step": 3931 }, { "epoch": 5.88622754491018, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1615, "step": 3932 }, { "epoch": 5.887724550898204, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1586, "step": 3933 }, { "epoch": 5.889221556886228, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1611, "step": 3934 }, { "epoch": 5.890718562874252, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1604, "step": 3935 }, { "epoch": 5.892215568862276, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1597, "step": 3936 }, { "epoch": 5.8937125748503, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1578, "step": 3937 }, { "epoch": 5.895209580838324, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1637, "step": 3938 }, { "epoch": 5.8967065868263475, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1591, "step": 3939 }, { "epoch": 5.8982035928143715, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1578, "step": 3940 }, { "epoch": 5.899700598802395, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1556, "step": 3941 }, { "epoch": 5.901197604790419, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1552, "step": 3942 }, { "epoch": 5.902694610778443, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1572, "step": 3943 }, { "epoch": 5.904191616766467, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1694, "step": 3944 }, { "epoch": 5.905688622754491, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1636, "step": 3945 }, { "epoch": 5.907185628742515, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1643, "step": 3946 }, { "epoch": 5.908682634730539, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1578, "step": 3947 }, { "epoch": 5.910179640718563, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.168, "step": 3948 }, { "epoch": 5.911676646706587, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1656, "step": 3949 }, { "epoch": 5.913173652694611, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1555, "step": 3950 }, { "epoch": 5.914670658682635, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1697, "step": 3951 }, { "epoch": 5.916167664670659, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1532, "step": 3952 }, { "epoch": 5.917664670658683, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1595, "step": 3953 }, { "epoch": 5.919161676646707, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1596, "step": 3954 }, { "epoch": 5.9206586826347305, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1592, "step": 3955 }, { "epoch": 5.922155688622754, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1684, "step": 3956 }, { "epoch": 5.923652694610778, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.163, "step": 3957 }, { "epoch": 5.925149700598802, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1562, "step": 3958 }, { "epoch": 5.926646706586826, "grad_norm": 0.1630859375, "learning_rate": 0.0008, "loss": 1.1629, "step": 3959 }, { "epoch": 5.92814371257485, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1519, "step": 3960 }, { "epoch": 5.929640718562874, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1534, "step": 3961 }, { "epoch": 5.931137724550898, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1634, "step": 3962 }, { "epoch": 5.932634730538922, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1582, "step": 3963 }, { "epoch": 5.934131736526946, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1585, "step": 3964 }, { "epoch": 5.93562874251497, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1548, "step": 3965 }, { "epoch": 5.937125748502994, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1657, "step": 3966 }, { "epoch": 5.938622754491018, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1551, "step": 3967 }, { "epoch": 5.940119760479042, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3968 }, { "epoch": 5.941616766467066, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1614, "step": 3969 }, { "epoch": 5.9431137724550895, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.156, "step": 3970 }, { "epoch": 5.9446107784431135, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1635, "step": 3971 }, { "epoch": 5.946107784431137, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1616, "step": 3972 }, { "epoch": 5.947604790419161, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1511, "step": 3973 }, { "epoch": 5.949101796407185, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1573, "step": 3974 }, { "epoch": 5.950598802395209, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1654, "step": 3975 }, { "epoch": 5.952095808383233, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1559, "step": 3976 }, { "epoch": 5.953592814371257, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1572, "step": 3977 }, { "epoch": 5.955089820359282, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.169, "step": 3978 }, { "epoch": 5.956586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1567, "step": 3979 }, { "epoch": 5.95808383233533, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1571, "step": 3980 }, { "epoch": 5.959580838323353, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1643, "step": 3981 }, { "epoch": 5.961077844311378, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1559, "step": 3982 }, { "epoch": 5.962574850299401, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1634, "step": 3983 }, { "epoch": 5.9640718562874255, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1617, "step": 3984 }, { "epoch": 5.9655688622754495, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.152, "step": 3985 }, { "epoch": 5.967065868263473, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1487, "step": 3986 }, { "epoch": 5.968562874251497, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1566, "step": 3987 }, { "epoch": 5.970059880239521, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1679, "step": 3988 }, { "epoch": 5.971556886227545, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1588, "step": 3989 }, { "epoch": 5.973053892215569, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1609, "step": 3990 }, { "epoch": 5.974550898203593, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1604, "step": 3991 }, { "epoch": 5.976047904191617, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1558, "step": 3992 }, { "epoch": 5.977544910179641, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1585, "step": 3993 }, { "epoch": 5.979041916167665, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3994 }, { "epoch": 5.980538922155689, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1654, "step": 3995 }, { "epoch": 5.982035928143713, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1579, "step": 3996 }, { "epoch": 5.983532934131737, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1494, "step": 3997 }, { "epoch": 5.985029940119761, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3998 }, { "epoch": 5.986526946107785, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1605, "step": 3999 }, { "epoch": 5.9880239520958085, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1595, "step": 4000 }, { "epoch": 5.9895209580838324, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1553, "step": 4001 }, { "epoch": 5.991017964071856, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1653, "step": 4002 }, { "epoch": 5.99251497005988, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1578, "step": 4003 }, { "epoch": 5.994011976047904, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1634, "step": 4004 }, { "epoch": 5.995508982035928, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1597, "step": 4005 }, { "epoch": 5.997005988023952, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1563, "step": 4006 }, { "epoch": 5.998502994011976, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1578, "step": 4007 }, { "epoch": 6.0, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1599, "step": 4008 }, { "epoch": 6.001497005988024, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1583, "step": 4009 }, { "epoch": 6.002994011976048, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1622, "step": 4010 }, { "epoch": 6.004491017964072, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1628, "step": 4011 }, { "epoch": 6.005988023952096, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4012 }, { "epoch": 6.00748502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1613, "step": 4013 }, { "epoch": 6.008982035928144, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.157, "step": 4014 }, { "epoch": 6.0104790419161676, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1613, "step": 4015 }, { "epoch": 6.0119760479041915, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1556, "step": 4016 }, { "epoch": 6.013473053892215, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1612, "step": 4017 }, { "epoch": 6.014970059880239, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1658, "step": 4018 }, { "epoch": 6.016467065868263, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1598, "step": 4019 }, { "epoch": 6.017964071856287, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1642, "step": 4020 }, { "epoch": 6.019461077844311, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1587, "step": 4021 }, { "epoch": 6.020958083832335, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1517, "step": 4022 }, { "epoch": 6.022455089820359, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1482, "step": 4023 }, { "epoch": 6.023952095808383, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1591, "step": 4024 }, { "epoch": 6.025449101796407, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1593, "step": 4025 }, { "epoch": 6.026946107784431, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1562, "step": 4026 }, { "epoch": 6.028443113772455, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1614, "step": 4027 }, { "epoch": 6.029940119760479, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1642, "step": 4028 }, { "epoch": 6.031437125748503, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1615, "step": 4029 }, { "epoch": 6.032934131736527, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1633, "step": 4030 }, { "epoch": 6.0344311377245505, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1578, "step": 4031 }, { "epoch": 6.0359281437125745, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.163, "step": 4032 }, { "epoch": 6.037425149700598, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1579, "step": 4033 }, { "epoch": 6.038922155688622, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1572, "step": 4034 }, { "epoch": 6.040419161676646, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1578, "step": 4035 }, { "epoch": 6.04191616766467, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1609, "step": 4036 }, { "epoch": 6.043413173652695, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1549, "step": 4037 }, { "epoch": 6.044910179640719, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1605, "step": 4038 }, { "epoch": 6.046407185628743, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1623, "step": 4039 }, { "epoch": 6.047904191616767, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1644, "step": 4040 }, { "epoch": 6.049401197604791, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1594, "step": 4041 }, { "epoch": 6.050898203592815, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1622, "step": 4042 }, { "epoch": 6.052395209580839, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1634, "step": 4043 }, { "epoch": 6.053892215568863, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1616, "step": 4044 }, { "epoch": 6.0553892215568865, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1536, "step": 4045 }, { "epoch": 6.0568862275449105, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1619, "step": 4046 }, { "epoch": 6.058383233532934, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1552, "step": 4047 }, { "epoch": 6.059880239520958, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1615, "step": 4048 }, { "epoch": 6.061377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.156, "step": 4049 }, { "epoch": 6.062874251497006, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1609, "step": 4050 }, { "epoch": 6.06437125748503, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1583, "step": 4051 }, { "epoch": 6.065868263473054, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1619, "step": 4052 }, { "epoch": 6.067365269461078, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1634, "step": 4053 }, { "epoch": 6.068862275449102, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.155, "step": 4054 }, { "epoch": 6.070359281437126, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1533, "step": 4055 }, { "epoch": 6.07185628742515, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1556, "step": 4056 }, { "epoch": 6.073353293413174, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4057 }, { "epoch": 6.074850299401198, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1549, "step": 4058 }, { "epoch": 6.076347305389222, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1566, "step": 4059 }, { "epoch": 6.077844311377246, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.153, "step": 4060 }, { "epoch": 6.0793413173652695, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4061 }, { "epoch": 6.080838323353293, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1551, "step": 4062 }, { "epoch": 6.082335329341317, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1638, "step": 4063 }, { "epoch": 6.083832335329341, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1595, "step": 4064 }, { "epoch": 6.085329341317365, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1586, "step": 4065 }, { "epoch": 6.086826347305389, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1531, "step": 4066 }, { "epoch": 6.088323353293413, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1613, "step": 4067 }, { "epoch": 6.089820359281437, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1633, "step": 4068 }, { "epoch": 6.091317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1643, "step": 4069 }, { "epoch": 6.092814371257485, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1601, "step": 4070 }, { "epoch": 6.094311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1603, "step": 4071 }, { "epoch": 6.095808383233533, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1591, "step": 4072 }, { "epoch": 6.097305389221557, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1577, "step": 4073 }, { "epoch": 6.098802395209581, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1616, "step": 4074 }, { "epoch": 6.100299401197605, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1556, "step": 4075 }, { "epoch": 6.1017964071856285, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1556, "step": 4076 }, { "epoch": 6.1032934131736525, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1613, "step": 4077 }, { "epoch": 6.104790419161676, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.162, "step": 4078 }, { "epoch": 6.1062874251497, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1545, "step": 4079 }, { "epoch": 6.107784431137724, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1586, "step": 4080 }, { "epoch": 6.109281437125748, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1573, "step": 4081 }, { "epoch": 6.110778443113772, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1625, "step": 4082 }, { "epoch": 6.112275449101796, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.164, "step": 4083 }, { "epoch": 6.11377245508982, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1633, "step": 4084 }, { "epoch": 6.115269461077844, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.158, "step": 4085 }, { "epoch": 6.116766467065868, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1571, "step": 4086 }, { "epoch": 6.118263473053892, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1652, "step": 4087 }, { "epoch": 6.119760479041916, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1579, "step": 4088 }, { "epoch": 6.12125748502994, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1617, "step": 4089 }, { "epoch": 6.122754491017964, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1536, "step": 4090 }, { "epoch": 6.124251497005988, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1629, "step": 4091 }, { "epoch": 6.125748502994012, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1603, "step": 4092 }, { "epoch": 6.127245508982036, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.157, "step": 4093 }, { "epoch": 6.12874251497006, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1591, "step": 4094 }, { "epoch": 6.130239520958084, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1612, "step": 4095 }, { "epoch": 6.131736526946108, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1609, "step": 4096 }, { "epoch": 6.133233532934132, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1586, "step": 4097 }, { "epoch": 6.134730538922156, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1631, "step": 4098 }, { "epoch": 6.13622754491018, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1514, "step": 4099 }, { "epoch": 6.137724550898204, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1637, "step": 4100 }, { "epoch": 6.139221556886228, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1605, "step": 4101 }, { "epoch": 6.140718562874252, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1572, "step": 4102 }, { "epoch": 6.142215568862276, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.156, "step": 4103 }, { "epoch": 6.1437125748503, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1548, "step": 4104 }, { "epoch": 6.145209580838324, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1609, "step": 4105 }, { "epoch": 6.1467065868263475, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4106 }, { "epoch": 6.1482035928143715, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1641, "step": 4107 }, { "epoch": 6.149700598802395, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1629, "step": 4108 }, { "epoch": 6.151197604790419, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1637, "step": 4109 }, { "epoch": 6.152694610778443, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1605, "step": 4110 }, { "epoch": 6.154191616766467, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1499, "step": 4111 }, { "epoch": 6.155688622754491, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1546, "step": 4112 }, { "epoch": 6.157185628742515, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1532, "step": 4113 }, { "epoch": 6.158682634730539, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.159, "step": 4114 }, { "epoch": 6.160179640718563, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1627, "step": 4115 }, { "epoch": 6.161676646706587, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1577, "step": 4116 }, { "epoch": 6.163173652694611, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1562, "step": 4117 }, { "epoch": 6.164670658682635, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1543, "step": 4118 }, { "epoch": 6.166167664670659, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4119 }, { "epoch": 6.167664670658683, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4120 }, { "epoch": 6.169161676646707, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1574, "step": 4121 }, { "epoch": 6.1706586826347305, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1577, "step": 4122 }, { "epoch": 6.172155688622754, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1594, "step": 4123 }, { "epoch": 6.173652694610778, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1538, "step": 4124 }, { "epoch": 6.175149700598802, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1553, "step": 4125 }, { "epoch": 6.176646706586826, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1514, "step": 4126 }, { "epoch": 6.17814371257485, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1688, "step": 4127 }, { "epoch": 6.179640718562874, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1612, "step": 4128 }, { "epoch": 6.181137724550898, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1569, "step": 4129 }, { "epoch": 6.182634730538922, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1568, "step": 4130 }, { "epoch": 6.184131736526946, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1585, "step": 4131 }, { "epoch": 6.18562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1519, "step": 4132 }, { "epoch": 6.187125748502994, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1607, "step": 4133 }, { "epoch": 6.188622754491018, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.163, "step": 4134 }, { "epoch": 6.190119760479042, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1564, "step": 4135 }, { "epoch": 6.191616766467066, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1604, "step": 4136 }, { "epoch": 6.1931137724550895, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1622, "step": 4137 }, { "epoch": 6.1946107784431135, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1555, "step": 4138 }, { "epoch": 6.196107784431137, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.16, "step": 4139 }, { "epoch": 6.197604790419161, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1532, "step": 4140 }, { "epoch": 6.199101796407185, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1596, "step": 4141 }, { "epoch": 6.200598802395209, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1615, "step": 4142 }, { "epoch": 6.202095808383233, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1552, "step": 4143 }, { "epoch": 6.203592814371257, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.158, "step": 4144 }, { "epoch": 6.205089820359281, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1635, "step": 4145 }, { "epoch": 6.206586826347305, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1614, "step": 4146 }, { "epoch": 6.20808383233533, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1643, "step": 4147 }, { "epoch": 6.209580838323353, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.154, "step": 4148 }, { "epoch": 6.211077844311378, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1586, "step": 4149 }, { "epoch": 6.212574850299402, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1617, "step": 4150 }, { "epoch": 6.2140718562874255, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.168, "step": 4151 }, { "epoch": 6.2155688622754495, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1502, "step": 4152 }, { "epoch": 6.217065868263473, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1571, "step": 4153 }, { "epoch": 6.218562874251497, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1568, "step": 4154 }, { "epoch": 6.220059880239521, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1536, "step": 4155 }, { "epoch": 6.221556886227545, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1604, "step": 4156 }, { "epoch": 6.223053892215569, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1538, "step": 4157 }, { "epoch": 6.224550898203593, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1544, "step": 4158 }, { "epoch": 6.226047904191617, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.16, "step": 4159 }, { "epoch": 6.227544910179641, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1569, "step": 4160 }, { "epoch": 6.229041916167665, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1669, "step": 4161 }, { "epoch": 6.230538922155689, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4162 }, { "epoch": 6.232035928143713, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1568, "step": 4163 }, { "epoch": 6.233532934131737, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1545, "step": 4164 }, { "epoch": 6.235029940119761, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1507, "step": 4165 }, { "epoch": 6.236526946107785, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1606, "step": 4166 }, { "epoch": 6.2380239520958085, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.152, "step": 4167 }, { "epoch": 6.2395209580838324, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4168 }, { "epoch": 6.241017964071856, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1648, "step": 4169 }, { "epoch": 6.24251497005988, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1556, "step": 4170 }, { "epoch": 6.244011976047904, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.157, "step": 4171 }, { "epoch": 6.245508982035928, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1563, "step": 4172 }, { "epoch": 6.247005988023952, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1592, "step": 4173 }, { "epoch": 6.248502994011976, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1605, "step": 4174 }, { "epoch": 6.25, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1616, "step": 4175 }, { "epoch": 6.251497005988024, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1542, "step": 4176 }, { "epoch": 6.252994011976048, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.163, "step": 4177 }, { "epoch": 6.254491017964072, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1551, "step": 4178 }, { "epoch": 6.255988023952096, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1624, "step": 4179 }, { "epoch": 6.25748502994012, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1611, "step": 4180 }, { "epoch": 6.258982035928144, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1646, "step": 4181 }, { "epoch": 6.2604790419161676, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1584, "step": 4182 }, { "epoch": 6.2619760479041915, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1599, "step": 4183 }, { "epoch": 6.263473053892215, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1591, "step": 4184 }, { "epoch": 6.264970059880239, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1612, "step": 4185 }, { "epoch": 6.266467065868263, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1562, "step": 4186 }, { "epoch": 6.267964071856287, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1572, "step": 4187 }, { "epoch": 6.269461077844311, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1636, "step": 4188 }, { "epoch": 6.270958083832335, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1574, "step": 4189 }, { "epoch": 6.272455089820359, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1566, "step": 4190 }, { "epoch": 6.273952095808383, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1634, "step": 4191 }, { "epoch": 6.275449101796407, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1664, "step": 4192 }, { "epoch": 6.276946107784431, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1589, "step": 4193 }, { "epoch": 6.278443113772455, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1503, "step": 4194 }, { "epoch": 6.279940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1613, "step": 4195 }, { "epoch": 6.281437125748503, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1525, "step": 4196 }, { "epoch": 6.282934131736527, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1668, "step": 4197 }, { "epoch": 6.2844311377245505, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.166, "step": 4198 }, { "epoch": 6.2859281437125745, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1543, "step": 4199 }, { "epoch": 6.287425149700598, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1628, "step": 4200 }, { "epoch": 6.288922155688622, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1582, "step": 4201 }, { "epoch": 6.290419161676647, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.163, "step": 4202 }, { "epoch": 6.29191616766467, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1582, "step": 4203 }, { "epoch": 6.293413173652695, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.1568, "step": 4204 }, { "epoch": 6.294910179640719, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1603, "step": 4205 }, { "epoch": 6.296407185628743, "grad_norm": 0.1650390625, "learning_rate": 0.0008, "loss": 1.1553, "step": 4206 }, { "epoch": 6.297904191616767, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1602, "step": 4207 }, { "epoch": 6.299401197604791, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1601, "step": 4208 }, { "epoch": 6.300898203592815, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1569, "step": 4209 }, { "epoch": 6.302395209580839, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1573, "step": 4210 }, { "epoch": 6.303892215568863, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1609, "step": 4211 }, { "epoch": 6.3053892215568865, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1631, "step": 4212 }, { "epoch": 6.3068862275449105, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1578, "step": 4213 }, { "epoch": 6.308383233532934, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1635, "step": 4214 }, { "epoch": 6.309880239520958, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1605, "step": 4215 }, { "epoch": 6.311377245508982, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1609, "step": 4216 }, { "epoch": 6.312874251497006, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1568, "step": 4217 }, { "epoch": 6.31437125748503, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1547, "step": 4218 }, { "epoch": 6.315868263473054, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1561, "step": 4219 }, { "epoch": 6.317365269461078, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1547, "step": 4220 }, { "epoch": 6.318862275449102, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1526, "step": 4221 }, { "epoch": 6.320359281437126, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1536, "step": 4222 }, { "epoch": 6.32185628742515, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1523, "step": 4223 }, { "epoch": 6.323353293413174, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1604, "step": 4224 }, { "epoch": 6.324850299401198, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4225 }, { "epoch": 6.326347305389222, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1569, "step": 4226 }, { "epoch": 6.327844311377246, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1596, "step": 4227 }, { "epoch": 6.3293413173652695, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.155, "step": 4228 }, { "epoch": 6.330838323353293, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1531, "step": 4229 }, { "epoch": 6.332335329341317, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1573, "step": 4230 }, { "epoch": 6.333832335329341, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1622, "step": 4231 }, { "epoch": 6.335329341317365, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1609, "step": 4232 }, { "epoch": 6.336826347305389, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1575, "step": 4233 }, { "epoch": 6.338323353293413, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1594, "step": 4234 }, { "epoch": 6.339820359281437, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1535, "step": 4235 }, { "epoch": 6.341317365269461, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1585, "step": 4236 }, { "epoch": 6.342814371257485, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1631, "step": 4237 }, { "epoch": 6.344311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1523, "step": 4238 }, { "epoch": 6.345808383233533, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1514, "step": 4239 }, { "epoch": 6.347305389221557, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1608, "step": 4240 }, { "epoch": 6.348802395209581, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1496, "step": 4241 }, { "epoch": 6.350299401197605, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1597, "step": 4242 }, { "epoch": 6.3517964071856285, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1648, "step": 4243 }, { "epoch": 6.3532934131736525, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1621, "step": 4244 }, { "epoch": 6.354790419161676, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.16, "step": 4245 }, { "epoch": 6.3562874251497, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1579, "step": 4246 }, { "epoch": 6.357784431137724, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1555, "step": 4247 }, { "epoch": 6.359281437125748, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1642, "step": 4248 }, { "epoch": 6.360778443113772, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1527, "step": 4249 }, { "epoch": 6.362275449101796, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1577, "step": 4250 }, { "epoch": 6.36377245508982, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.164, "step": 4251 }, { "epoch": 6.365269461077844, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1638, "step": 4252 }, { "epoch": 6.366766467065868, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1522, "step": 4253 }, { "epoch": 6.368263473053892, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4254 }, { "epoch": 6.369760479041916, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1551, "step": 4255 }, { "epoch": 6.37125748502994, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1614, "step": 4256 }, { "epoch": 6.3727544910179645, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1607, "step": 4257 }, { "epoch": 6.374251497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1656, "step": 4258 }, { "epoch": 6.375748502994012, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1595, "step": 4259 }, { "epoch": 6.3772455089820355, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1634, "step": 4260 }, { "epoch": 6.37874251497006, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1592, "step": 4261 }, { "epoch": 6.380239520958084, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1552, "step": 4262 }, { "epoch": 6.381736526946108, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4263 }, { "epoch": 6.383233532934132, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1537, "step": 4264 }, { "epoch": 6.384730538922156, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1587, "step": 4265 }, { "epoch": 6.38622754491018, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1563, "step": 4266 }, { "epoch": 6.387724550898204, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1584, "step": 4267 }, { "epoch": 6.389221556886228, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1535, "step": 4268 }, { "epoch": 6.390718562874252, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1542, "step": 4269 }, { "epoch": 6.392215568862276, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1551, "step": 4270 }, { "epoch": 6.3937125748503, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1549, "step": 4271 }, { "epoch": 6.395209580838324, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1618, "step": 4272 }, { "epoch": 6.3967065868263475, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1636, "step": 4273 }, { "epoch": 6.3982035928143715, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1565, "step": 4274 }, { "epoch": 6.399700598802395, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1576, "step": 4275 }, { "epoch": 6.401197604790419, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1573, "step": 4276 }, { "epoch": 6.402694610778443, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1584, "step": 4277 }, { "epoch": 6.404191616766467, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1567, "step": 4278 }, { "epoch": 6.405688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1526, "step": 4279 }, { "epoch": 6.407185628742515, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1588, "step": 4280 }, { "epoch": 6.408682634730539, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1481, "step": 4281 }, { "epoch": 6.410179640718563, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.158, "step": 4282 }, { "epoch": 6.411676646706587, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1634, "step": 4283 }, { "epoch": 6.413173652694611, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1627, "step": 4284 }, { "epoch": 6.414670658682635, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4285 }, { "epoch": 6.416167664670659, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1544, "step": 4286 }, { "epoch": 6.417664670658683, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1604, "step": 4287 }, { "epoch": 6.419161676646707, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1597, "step": 4288 }, { "epoch": 6.4206586826347305, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1538, "step": 4289 }, { "epoch": 6.422155688622754, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1552, "step": 4290 }, { "epoch": 6.423652694610778, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1526, "step": 4291 }, { "epoch": 6.425149700598802, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1494, "step": 4292 }, { "epoch": 6.426646706586826, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1622, "step": 4293 }, { "epoch": 6.42814371257485, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1595, "step": 4294 }, { "epoch": 6.429640718562874, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.16, "step": 4295 }, { "epoch": 6.431137724550898, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1626, "step": 4296 }, { "epoch": 6.432634730538922, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1581, "step": 4297 }, { "epoch": 6.434131736526946, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1586, "step": 4298 }, { "epoch": 6.43562874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1548, "step": 4299 }, { "epoch": 6.437125748502994, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1582, "step": 4300 }, { "epoch": 6.438622754491018, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1614, "step": 4301 }, { "epoch": 6.440119760479042, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1546, "step": 4302 }, { "epoch": 6.441616766467066, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1541, "step": 4303 }, { "epoch": 6.4431137724550895, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1555, "step": 4304 }, { "epoch": 6.4446107784431135, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1659, "step": 4305 }, { "epoch": 6.446107784431137, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1582, "step": 4306 }, { "epoch": 6.447604790419161, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1545, "step": 4307 }, { "epoch": 6.449101796407185, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1563, "step": 4308 }, { "epoch": 6.450598802395209, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.154, "step": 4309 }, { "epoch": 6.452095808383233, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1584, "step": 4310 }, { "epoch": 6.453592814371257, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1546, "step": 4311 }, { "epoch": 6.455089820359281, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1559, "step": 4312 }, { "epoch": 6.456586826347305, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1598, "step": 4313 }, { "epoch": 6.45808383233533, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.155, "step": 4314 }, { "epoch": 6.459580838323353, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1643, "step": 4315 }, { "epoch": 6.461077844311378, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1656, "step": 4316 }, { "epoch": 6.462574850299402, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.154, "step": 4317 }, { "epoch": 6.4640718562874255, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1538, "step": 4318 }, { "epoch": 6.4655688622754495, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1549, "step": 4319 }, { "epoch": 6.467065868263473, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1515, "step": 4320 }, { "epoch": 6.468562874251497, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1612, "step": 4321 }, { "epoch": 6.470059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1575, "step": 4322 }, { "epoch": 6.471556886227545, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1539, "step": 4323 }, { "epoch": 6.473053892215569, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1593, "step": 4324 }, { "epoch": 6.474550898203593, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4325 }, { "epoch": 6.476047904191617, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1544, "step": 4326 }, { "epoch": 6.477544910179641, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4327 }, { "epoch": 6.479041916167665, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.152, "step": 4328 }, { "epoch": 6.480538922155689, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1593, "step": 4329 }, { "epoch": 6.482035928143713, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1549, "step": 4330 }, { "epoch": 6.483532934131737, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1569, "step": 4331 }, { "epoch": 6.485029940119761, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1608, "step": 4332 }, { "epoch": 6.486526946107785, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1566, "step": 4333 }, { "epoch": 6.4880239520958085, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1527, "step": 4334 }, { "epoch": 6.4895209580838324, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.155, "step": 4335 }, { "epoch": 6.491017964071856, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4336 }, { "epoch": 6.49251497005988, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.164, "step": 4337 }, { "epoch": 6.494011976047904, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1586, "step": 4338 }, { "epoch": 6.495508982035928, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4339 }, { "epoch": 6.497005988023952, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1585, "step": 4340 }, { "epoch": 6.498502994011976, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1538, "step": 4341 }, { "epoch": 6.5, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1678, "step": 4342 }, { "epoch": 6.501497005988024, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1581, "step": 4343 }, { "epoch": 6.502994011976048, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1525, "step": 4344 }, { "epoch": 6.504491017964072, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1574, "step": 4345 }, { "epoch": 6.505988023952096, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1606, "step": 4346 }, { "epoch": 6.50748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4347 }, { "epoch": 6.508982035928144, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1531, "step": 4348 }, { "epoch": 6.5104790419161676, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1566, "step": 4349 }, { "epoch": 6.5119760479041915, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.167, "step": 4350 }, { "epoch": 6.513473053892215, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1611, "step": 4351 }, { "epoch": 6.514970059880239, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1631, "step": 4352 }, { "epoch": 6.516467065868263, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1601, "step": 4353 }, { "epoch": 6.517964071856287, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1552, "step": 4354 }, { "epoch": 6.519461077844311, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1639, "step": 4355 }, { "epoch": 6.520958083832335, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1594, "step": 4356 }, { "epoch": 6.522455089820359, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1558, "step": 4357 }, { "epoch": 6.523952095808383, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1518, "step": 4358 }, { "epoch": 6.525449101796407, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.161, "step": 4359 }, { "epoch": 6.526946107784431, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1512, "step": 4360 }, { "epoch": 6.528443113772455, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1539, "step": 4361 }, { "epoch": 6.529940119760479, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4362 }, { "epoch": 6.531437125748503, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1534, "step": 4363 }, { "epoch": 6.532934131736527, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4364 }, { "epoch": 6.5344311377245505, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1609, "step": 4365 }, { "epoch": 6.5359281437125745, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1527, "step": 4366 }, { "epoch": 6.537425149700599, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1621, "step": 4367 }, { "epoch": 6.538922155688622, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1558, "step": 4368 }, { "epoch": 6.540419161676647, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1523, "step": 4369 }, { "epoch": 6.54191616766467, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.16, "step": 4370 }, { "epoch": 6.543413173652695, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1581, "step": 4371 }, { "epoch": 6.544910179640718, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1609, "step": 4372 }, { "epoch": 6.546407185628743, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1549, "step": 4373 }, { "epoch": 6.547904191616767, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1553, "step": 4374 }, { "epoch": 6.549401197604791, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1587, "step": 4375 }, { "epoch": 6.550898203592815, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1594, "step": 4376 }, { "epoch": 6.552395209580839, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1549, "step": 4377 }, { "epoch": 6.553892215568863, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1624, "step": 4378 }, { "epoch": 6.5553892215568865, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1482, "step": 4379 }, { "epoch": 6.5568862275449105, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1549, "step": 4380 }, { "epoch": 6.558383233532934, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1581, "step": 4381 }, { "epoch": 6.559880239520958, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4382 }, { "epoch": 6.561377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1543, "step": 4383 }, { "epoch": 6.562874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1551, "step": 4384 }, { "epoch": 6.56437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1602, "step": 4385 }, { "epoch": 6.565868263473054, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1543, "step": 4386 }, { "epoch": 6.567365269461078, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1471, "step": 4387 }, { "epoch": 6.568862275449102, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1591, "step": 4388 }, { "epoch": 6.570359281437126, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1577, "step": 4389 }, { "epoch": 6.57185628742515, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1595, "step": 4390 }, { "epoch": 6.573353293413174, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1568, "step": 4391 }, { "epoch": 6.574850299401198, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1532, "step": 4392 }, { "epoch": 6.576347305389222, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1553, "step": 4393 }, { "epoch": 6.577844311377246, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1607, "step": 4394 }, { "epoch": 6.5793413173652695, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1612, "step": 4395 }, { "epoch": 6.580838323353293, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1565, "step": 4396 }, { "epoch": 6.582335329341317, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1573, "step": 4397 }, { "epoch": 6.583832335329341, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1559, "step": 4398 }, { "epoch": 6.585329341317365, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1592, "step": 4399 }, { "epoch": 6.586826347305389, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1546, "step": 4400 }, { "epoch": 6.588323353293413, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1595, "step": 4401 }, { "epoch": 6.589820359281437, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1601, "step": 4402 }, { "epoch": 6.591317365269461, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1617, "step": 4403 }, { "epoch": 6.592814371257485, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1613, "step": 4404 }, { "epoch": 6.594311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1602, "step": 4405 }, { "epoch": 6.595808383233533, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1538, "step": 4406 }, { "epoch": 6.597305389221557, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1544, "step": 4407 }, { "epoch": 6.598802395209581, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.168, "step": 4408 }, { "epoch": 6.600299401197605, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1551, "step": 4409 }, { "epoch": 6.6017964071856285, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1592, "step": 4410 }, { "epoch": 6.6032934131736525, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1539, "step": 4411 }, { "epoch": 6.604790419161676, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1595, "step": 4412 }, { "epoch": 6.6062874251497, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1621, "step": 4413 }, { "epoch": 6.607784431137724, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1509, "step": 4414 }, { "epoch": 6.609281437125748, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1551, "step": 4415 }, { "epoch": 6.610778443113772, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1618, "step": 4416 }, { "epoch": 6.612275449101796, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1591, "step": 4417 }, { "epoch": 6.61377245508982, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1645, "step": 4418 }, { "epoch": 6.615269461077844, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1576, "step": 4419 }, { "epoch": 6.616766467065868, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1515, "step": 4420 }, { "epoch": 6.618263473053892, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1526, "step": 4421 }, { "epoch": 6.619760479041916, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4422 }, { "epoch": 6.62125748502994, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4423 }, { "epoch": 6.6227544910179645, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1573, "step": 4424 }, { "epoch": 6.624251497005988, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.159, "step": 4425 }, { "epoch": 6.625748502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1497, "step": 4426 }, { "epoch": 6.6272455089820355, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1553, "step": 4427 }, { "epoch": 6.62874251497006, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1602, "step": 4428 }, { "epoch": 6.630239520958084, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1577, "step": 4429 }, { "epoch": 6.631736526946108, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1624, "step": 4430 }, { "epoch": 6.633233532934132, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1586, "step": 4431 }, { "epoch": 6.634730538922156, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1607, "step": 4432 }, { "epoch": 6.63622754491018, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1593, "step": 4433 }, { "epoch": 6.637724550898204, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.159, "step": 4434 }, { "epoch": 6.639221556886228, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.161, "step": 4435 }, { "epoch": 6.640718562874252, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1548, "step": 4436 }, { "epoch": 6.642215568862276, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.16, "step": 4437 }, { "epoch": 6.6437125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1565, "step": 4438 }, { "epoch": 6.645209580838324, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1585, "step": 4439 }, { "epoch": 6.6467065868263475, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1621, "step": 4440 }, { "epoch": 6.6482035928143715, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1557, "step": 4441 }, { "epoch": 6.649700598802395, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1554, "step": 4442 }, { "epoch": 6.651197604790419, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4443 }, { "epoch": 6.652694610778443, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1563, "step": 4444 }, { "epoch": 6.654191616766467, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1597, "step": 4445 }, { "epoch": 6.655688622754491, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1533, "step": 4446 }, { "epoch": 6.657185628742515, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1591, "step": 4447 }, { "epoch": 6.658682634730539, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1534, "step": 4448 }, { "epoch": 6.660179640718563, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1552, "step": 4449 }, { "epoch": 6.661676646706587, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1556, "step": 4450 }, { "epoch": 6.663173652694611, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.156, "step": 4451 }, { "epoch": 6.664670658682635, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1491, "step": 4452 }, { "epoch": 6.666167664670659, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1492, "step": 4453 }, { "epoch": 6.667664670658683, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1548, "step": 4454 }, { "epoch": 6.669161676646707, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1529, "step": 4455 }, { "epoch": 6.6706586826347305, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4456 }, { "epoch": 6.672155688622754, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1643, "step": 4457 }, { "epoch": 6.673652694610778, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1506, "step": 4458 }, { "epoch": 6.675149700598802, "grad_norm": 0.171875, "learning_rate": 0.0008, "loss": 1.1503, "step": 4459 }, { "epoch": 6.676646706586826, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1637, "step": 4460 }, { "epoch": 6.67814371257485, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1612, "step": 4461 }, { "epoch": 6.679640718562874, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1584, "step": 4462 }, { "epoch": 6.681137724550898, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.154, "step": 4463 }, { "epoch": 6.682634730538922, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1519, "step": 4464 }, { "epoch": 6.684131736526946, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.161, "step": 4465 }, { "epoch": 6.68562874251497, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4466 }, { "epoch": 6.687125748502994, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1658, "step": 4467 }, { "epoch": 6.688622754491018, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1567, "step": 4468 }, { "epoch": 6.690119760479042, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1535, "step": 4469 }, { "epoch": 6.691616766467066, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1498, "step": 4470 }, { "epoch": 6.6931137724550895, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.153, "step": 4471 }, { "epoch": 6.6946107784431135, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.159, "step": 4472 }, { "epoch": 6.696107784431137, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1575, "step": 4473 }, { "epoch": 6.697604790419161, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1593, "step": 4474 }, { "epoch": 6.699101796407185, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1571, "step": 4475 }, { "epoch": 6.700598802395209, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4476 }, { "epoch": 6.702095808383233, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1659, "step": 4477 }, { "epoch": 6.703592814371257, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.155, "step": 4478 }, { "epoch": 6.705089820359282, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1515, "step": 4479 }, { "epoch": 6.706586826347305, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1579, "step": 4480 }, { "epoch": 6.70808383233533, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4481 }, { "epoch": 6.709580838323353, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1598, "step": 4482 }, { "epoch": 6.711077844311378, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1664, "step": 4483 }, { "epoch": 6.712574850299401, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1508, "step": 4484 }, { "epoch": 6.7140718562874255, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4485 }, { "epoch": 6.7155688622754495, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1566, "step": 4486 }, { "epoch": 6.717065868263473, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1643, "step": 4487 }, { "epoch": 6.718562874251497, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1602, "step": 4488 }, { "epoch": 6.720059880239521, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1543, "step": 4489 }, { "epoch": 6.721556886227545, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1533, "step": 4490 }, { "epoch": 6.723053892215569, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1545, "step": 4491 }, { "epoch": 6.724550898203593, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1579, "step": 4492 }, { "epoch": 6.726047904191617, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1598, "step": 4493 }, { "epoch": 6.727544910179641, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1546, "step": 4494 }, { "epoch": 6.729041916167665, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1438, "step": 4495 }, { "epoch": 6.730538922155689, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1524, "step": 4496 }, { "epoch": 6.732035928143713, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1575, "step": 4497 }, { "epoch": 6.733532934131737, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1545, "step": 4498 }, { "epoch": 6.735029940119761, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1511, "step": 4499 }, { "epoch": 6.736526946107785, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1617, "step": 4500 }, { "epoch": 6.7380239520958085, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.16, "step": 4501 }, { "epoch": 6.7395209580838324, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1516, "step": 4502 }, { "epoch": 6.741017964071856, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1585, "step": 4503 }, { "epoch": 6.74251497005988, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1543, "step": 4504 }, { "epoch": 6.744011976047904, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.162, "step": 4505 }, { "epoch": 6.745508982035928, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1656, "step": 4506 }, { "epoch": 6.747005988023952, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1579, "step": 4507 }, { "epoch": 6.748502994011976, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4508 }, { "epoch": 6.75, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1523, "step": 4509 }, { "epoch": 6.751497005988024, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1478, "step": 4510 }, { "epoch": 6.752994011976048, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1555, "step": 4511 }, { "epoch": 6.754491017964072, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.158, "step": 4512 }, { "epoch": 6.755988023952096, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1506, "step": 4513 }, { "epoch": 6.75748502994012, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1531, "step": 4514 }, { "epoch": 6.758982035928144, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1534, "step": 4515 }, { "epoch": 6.7604790419161676, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1555, "step": 4516 }, { "epoch": 6.7619760479041915, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1594, "step": 4517 }, { "epoch": 6.763473053892215, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1586, "step": 4518 }, { "epoch": 6.764970059880239, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1582, "step": 4519 }, { "epoch": 6.766467065868263, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.151, "step": 4520 }, { "epoch": 6.767964071856287, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1528, "step": 4521 }, { "epoch": 6.769461077844311, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1633, "step": 4522 }, { "epoch": 6.770958083832335, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1611, "step": 4523 }, { "epoch": 6.772455089820359, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1554, "step": 4524 }, { "epoch": 6.773952095808383, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1598, "step": 4525 }, { "epoch": 6.775449101796407, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1608, "step": 4526 }, { "epoch": 6.776946107784431, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.161, "step": 4527 }, { "epoch": 6.778443113772455, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1546, "step": 4528 }, { "epoch": 6.779940119760479, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4529 }, { "epoch": 6.781437125748503, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1596, "step": 4530 }, { "epoch": 6.782934131736527, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4531 }, { "epoch": 6.7844311377245505, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1494, "step": 4532 }, { "epoch": 6.7859281437125745, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1545, "step": 4533 }, { "epoch": 6.787425149700599, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1542, "step": 4534 }, { "epoch": 6.788922155688622, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4535 }, { "epoch": 6.790419161676647, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1541, "step": 4536 }, { "epoch": 6.79191616766467, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1551, "step": 4537 }, { "epoch": 6.793413173652695, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4538 }, { "epoch": 6.794910179640718, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1561, "step": 4539 }, { "epoch": 6.796407185628743, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.154, "step": 4540 }, { "epoch": 6.797904191616767, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1503, "step": 4541 }, { "epoch": 6.799401197604791, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1623, "step": 4542 }, { "epoch": 6.800898203592815, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1581, "step": 4543 }, { "epoch": 6.802395209580839, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1547, "step": 4544 }, { "epoch": 6.803892215568863, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1584, "step": 4545 }, { "epoch": 6.8053892215568865, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1583, "step": 4546 }, { "epoch": 6.8068862275449105, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1564, "step": 4547 }, { "epoch": 6.808383233532934, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1581, "step": 4548 }, { "epoch": 6.809880239520958, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1533, "step": 4549 }, { "epoch": 6.811377245508982, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1602, "step": 4550 }, { "epoch": 6.812874251497006, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1553, "step": 4551 }, { "epoch": 6.81437125748503, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1543, "step": 4552 }, { "epoch": 6.815868263473054, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1487, "step": 4553 }, { "epoch": 6.817365269461078, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1563, "step": 4554 }, { "epoch": 6.818862275449102, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1582, "step": 4555 }, { "epoch": 6.820359281437126, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1554, "step": 4556 }, { "epoch": 6.82185628742515, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1592, "step": 4557 }, { "epoch": 6.823353293413174, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1589, "step": 4558 }, { "epoch": 6.824850299401198, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1565, "step": 4559 }, { "epoch": 6.826347305389222, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1527, "step": 4560 }, { "epoch": 6.827844311377246, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1562, "step": 4561 }, { "epoch": 6.8293413173652695, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.158, "step": 4562 }, { "epoch": 6.830838323353293, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1606, "step": 4563 }, { "epoch": 6.832335329341317, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1528, "step": 4564 }, { "epoch": 6.833832335329341, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.156, "step": 4565 }, { "epoch": 6.835329341317365, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1537, "step": 4566 }, { "epoch": 6.836826347305389, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1559, "step": 4567 }, { "epoch": 6.838323353293413, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1572, "step": 4568 }, { "epoch": 6.839820359281437, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1549, "step": 4569 }, { "epoch": 6.841317365269461, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1569, "step": 4570 }, { "epoch": 6.842814371257485, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1552, "step": 4571 }, { "epoch": 6.844311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1471, "step": 4572 }, { "epoch": 6.845808383233533, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1608, "step": 4573 }, { "epoch": 6.847305389221557, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1536, "step": 4574 }, { "epoch": 6.848802395209581, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1531, "step": 4575 }, { "epoch": 6.850299401197605, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1484, "step": 4576 }, { "epoch": 6.8517964071856285, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1551, "step": 4577 }, { "epoch": 6.8532934131736525, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.153, "step": 4578 }, { "epoch": 6.854790419161676, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.156, "step": 4579 }, { "epoch": 6.8562874251497, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1617, "step": 4580 }, { "epoch": 6.857784431137724, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4581 }, { "epoch": 6.859281437125748, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1596, "step": 4582 }, { "epoch": 6.860778443113772, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.16, "step": 4583 }, { "epoch": 6.862275449101796, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1532, "step": 4584 }, { "epoch": 6.86377245508982, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1509, "step": 4585 }, { "epoch": 6.865269461077844, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1598, "step": 4586 }, { "epoch": 6.866766467065868, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1492, "step": 4587 }, { "epoch": 6.868263473053892, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1619, "step": 4588 }, { "epoch": 6.869760479041916, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1594, "step": 4589 }, { "epoch": 6.87125748502994, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.1551, "step": 4590 }, { "epoch": 6.8727544910179645, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.161, "step": 4591 }, { "epoch": 6.874251497005988, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.153, "step": 4592 }, { "epoch": 6.875748502994012, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1583, "step": 4593 }, { "epoch": 6.8772455089820355, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1583, "step": 4594 }, { "epoch": 6.87874251497006, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4595 }, { "epoch": 6.880239520958084, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1581, "step": 4596 }, { "epoch": 6.881736526946108, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1548, "step": 4597 }, { "epoch": 6.883233532934132, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1519, "step": 4598 }, { "epoch": 6.884730538922156, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1505, "step": 4599 }, { "epoch": 6.88622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1521, "step": 4600 }, { "epoch": 6.887724550898204, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1516, "step": 4601 }, { "epoch": 6.889221556886228, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1588, "step": 4602 }, { "epoch": 6.890718562874252, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4603 }, { "epoch": 6.892215568862276, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1571, "step": 4604 }, { "epoch": 6.8937125748503, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1528, "step": 4605 }, { "epoch": 6.895209580838324, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1529, "step": 4606 }, { "epoch": 6.8967065868263475, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1538, "step": 4607 }, { "epoch": 6.8982035928143715, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1546, "step": 4608 }, { "epoch": 6.899700598802395, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1516, "step": 4609 }, { "epoch": 6.901197604790419, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.15, "step": 4610 }, { "epoch": 6.902694610778443, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1537, "step": 4611 }, { "epoch": 6.904191616766467, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4612 }, { "epoch": 6.905688622754491, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.158, "step": 4613 }, { "epoch": 6.907185628742515, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4614 }, { "epoch": 6.908682634730539, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1538, "step": 4615 }, { "epoch": 6.910179640718563, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1578, "step": 4616 }, { "epoch": 6.911676646706587, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1522, "step": 4617 }, { "epoch": 6.913173652694611, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1618, "step": 4618 }, { "epoch": 6.914670658682635, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1523, "step": 4619 }, { "epoch": 6.916167664670659, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.157, "step": 4620 }, { "epoch": 6.917664670658683, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1597, "step": 4621 }, { "epoch": 6.919161676646707, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1594, "step": 4622 }, { "epoch": 6.9206586826347305, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.161, "step": 4623 }, { "epoch": 6.922155688622754, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1541, "step": 4624 }, { "epoch": 6.923652694610778, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4625 }, { "epoch": 6.925149700598802, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1644, "step": 4626 }, { "epoch": 6.926646706586826, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1543, "step": 4627 }, { "epoch": 6.92814371257485, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1557, "step": 4628 }, { "epoch": 6.929640718562874, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1597, "step": 4629 }, { "epoch": 6.931137724550898, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.151, "step": 4630 }, { "epoch": 6.932634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1566, "step": 4631 }, { "epoch": 6.934131736526946, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1605, "step": 4632 }, { "epoch": 6.93562874251497, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.159, "step": 4633 }, { "epoch": 6.937125748502994, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.153, "step": 4634 }, { "epoch": 6.938622754491018, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1532, "step": 4635 }, { "epoch": 6.940119760479042, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1555, "step": 4636 }, { "epoch": 6.941616766467066, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1646, "step": 4637 }, { "epoch": 6.9431137724550895, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1598, "step": 4638 }, { "epoch": 6.9446107784431135, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1531, "step": 4639 }, { "epoch": 6.946107784431137, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1601, "step": 4640 }, { "epoch": 6.947604790419161, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1565, "step": 4641 }, { "epoch": 6.949101796407185, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.153, "step": 4642 }, { "epoch": 6.950598802395209, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1601, "step": 4643 }, { "epoch": 6.952095808383233, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1555, "step": 4644 }, { "epoch": 6.953592814371257, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.16, "step": 4645 }, { "epoch": 6.955089820359282, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1599, "step": 4646 }, { "epoch": 6.956586826347305, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1533, "step": 4647 }, { "epoch": 6.95808383233533, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1597, "step": 4648 }, { "epoch": 6.959580838323353, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1539, "step": 4649 }, { "epoch": 6.961077844311378, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1583, "step": 4650 }, { "epoch": 6.962574850299401, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1547, "step": 4651 }, { "epoch": 6.9640718562874255, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.16, "step": 4652 }, { "epoch": 6.9655688622754495, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1578, "step": 4653 }, { "epoch": 6.967065868263473, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1522, "step": 4654 }, { "epoch": 6.968562874251497, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1494, "step": 4655 }, { "epoch": 6.970059880239521, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1634, "step": 4656 }, { "epoch": 6.971556886227545, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1506, "step": 4657 }, { "epoch": 6.973053892215569, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1566, "step": 4658 }, { "epoch": 6.974550898203593, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1505, "step": 4659 }, { "epoch": 6.976047904191617, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1625, "step": 4660 }, { "epoch": 6.977544910179641, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1603, "step": 4661 }, { "epoch": 6.979041916167665, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.157, "step": 4662 }, { "epoch": 6.980538922155689, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4663 }, { "epoch": 6.982035928143713, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1509, "step": 4664 }, { "epoch": 6.983532934131737, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1619, "step": 4665 }, { "epoch": 6.985029940119761, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1596, "step": 4666 }, { "epoch": 6.986526946107785, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1537, "step": 4667 }, { "epoch": 6.9880239520958085, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.157, "step": 4668 }, { "epoch": 6.9895209580838324, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1564, "step": 4669 }, { "epoch": 6.991017964071856, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1537, "step": 4670 }, { "epoch": 6.99251497005988, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1597, "step": 4671 }, { "epoch": 6.994011976047904, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1584, "step": 4672 }, { "epoch": 6.995508982035928, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1563, "step": 4673 }, { "epoch": 6.997005988023952, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1561, "step": 4674 }, { "epoch": 6.998502994011976, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1582, "step": 4675 }, { "epoch": 7.0, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1549, "step": 4676 }, { "epoch": 7.001497005988024, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1562, "step": 4677 }, { "epoch": 7.002994011976048, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1631, "step": 4678 }, { "epoch": 7.004491017964072, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1555, "step": 4679 }, { "epoch": 7.005988023952096, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1591, "step": 4680 }, { "epoch": 7.00748502994012, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1508, "step": 4681 }, { "epoch": 7.008982035928144, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1619, "step": 4682 }, { "epoch": 7.0104790419161676, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1511, "step": 4683 }, { "epoch": 7.0119760479041915, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4684 }, { "epoch": 7.013473053892215, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1527, "step": 4685 }, { "epoch": 7.014970059880239, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.148, "step": 4686 }, { "epoch": 7.016467065868263, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1663, "step": 4687 }, { "epoch": 7.017964071856287, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1596, "step": 4688 }, { "epoch": 7.019461077844311, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1545, "step": 4689 }, { "epoch": 7.020958083832335, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1554, "step": 4690 }, { "epoch": 7.022455089820359, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1544, "step": 4691 }, { "epoch": 7.023952095808383, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1513, "step": 4692 }, { "epoch": 7.025449101796407, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1531, "step": 4693 }, { "epoch": 7.026946107784431, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.155, "step": 4694 }, { "epoch": 7.028443113772455, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1591, "step": 4695 }, { "epoch": 7.029940119760479, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1539, "step": 4696 }, { "epoch": 7.031437125748503, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1614, "step": 4697 }, { "epoch": 7.032934131736527, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1565, "step": 4698 }, { "epoch": 7.0344311377245505, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1586, "step": 4699 }, { "epoch": 7.0359281437125745, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1545, "step": 4700 }, { "epoch": 7.037425149700598, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1591, "step": 4701 }, { "epoch": 7.038922155688622, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1565, "step": 4702 }, { "epoch": 7.040419161676646, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1458, "step": 4703 }, { "epoch": 7.04191616766467, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1548, "step": 4704 }, { "epoch": 7.043413173652695, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1565, "step": 4705 }, { "epoch": 7.044910179640719, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1562, "step": 4706 }, { "epoch": 7.046407185628743, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1436, "step": 4707 }, { "epoch": 7.047904191616767, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1601, "step": 4708 }, { "epoch": 7.049401197604791, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1557, "step": 4709 }, { "epoch": 7.050898203592815, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1563, "step": 4710 }, { "epoch": 7.052395209580839, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1543, "step": 4711 }, { "epoch": 7.053892215568863, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1478, "step": 4712 }, { "epoch": 7.0553892215568865, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1536, "step": 4713 }, { "epoch": 7.0568862275449105, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1559, "step": 4714 }, { "epoch": 7.058383233532934, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1525, "step": 4715 }, { "epoch": 7.059880239520958, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4716 }, { "epoch": 7.061377245508982, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1513, "step": 4717 }, { "epoch": 7.062874251497006, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1463, "step": 4718 }, { "epoch": 7.06437125748503, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1553, "step": 4719 }, { "epoch": 7.065868263473054, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1536, "step": 4720 }, { "epoch": 7.067365269461078, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4721 }, { "epoch": 7.068862275449102, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1471, "step": 4722 }, { "epoch": 7.070359281437126, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1593, "step": 4723 }, { "epoch": 7.07185628742515, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1575, "step": 4724 }, { "epoch": 7.073353293413174, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1511, "step": 4725 }, { "epoch": 7.074850299401198, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1533, "step": 4726 }, { "epoch": 7.076347305389222, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1526, "step": 4727 }, { "epoch": 7.077844311377246, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1583, "step": 4728 }, { "epoch": 7.0793413173652695, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1575, "step": 4729 }, { "epoch": 7.080838323353293, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1613, "step": 4730 }, { "epoch": 7.082335329341317, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4731 }, { "epoch": 7.083832335329341, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1531, "step": 4732 }, { "epoch": 7.085329341317365, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1531, "step": 4733 }, { "epoch": 7.086826347305389, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1531, "step": 4734 }, { "epoch": 7.088323353293413, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4735 }, { "epoch": 7.089820359281437, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1488, "step": 4736 }, { "epoch": 7.091317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1579, "step": 4737 }, { "epoch": 7.092814371257485, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1467, "step": 4738 }, { "epoch": 7.094311377245509, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1503, "step": 4739 }, { "epoch": 7.095808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1583, "step": 4740 }, { "epoch": 7.097305389221557, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.155, "step": 4741 }, { "epoch": 7.098802395209581, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.161, "step": 4742 }, { "epoch": 7.100299401197605, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1586, "step": 4743 }, { "epoch": 7.1017964071856285, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1537, "step": 4744 }, { "epoch": 7.1032934131736525, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1561, "step": 4745 }, { "epoch": 7.104790419161676, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1508, "step": 4746 }, { "epoch": 7.1062874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1526, "step": 4747 }, { "epoch": 7.107784431137724, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1542, "step": 4748 }, { "epoch": 7.109281437125748, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1536, "step": 4749 }, { "epoch": 7.110778443113772, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1584, "step": 4750 }, { "epoch": 7.112275449101796, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1592, "step": 4751 }, { "epoch": 7.11377245508982, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.154, "step": 4752 }, { "epoch": 7.115269461077844, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1497, "step": 4753 }, { "epoch": 7.116766467065868, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1592, "step": 4754 }, { "epoch": 7.118263473053892, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1583, "step": 4755 }, { "epoch": 7.119760479041916, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1606, "step": 4756 }, { "epoch": 7.12125748502994, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1548, "step": 4757 }, { "epoch": 7.122754491017964, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.162, "step": 4758 }, { "epoch": 7.124251497005988, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1553, "step": 4759 }, { "epoch": 7.125748502994012, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1515, "step": 4760 }, { "epoch": 7.127245508982036, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1557, "step": 4761 }, { "epoch": 7.12874251497006, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1598, "step": 4762 }, { "epoch": 7.130239520958084, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1617, "step": 4763 }, { "epoch": 7.131736526946108, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1573, "step": 4764 }, { "epoch": 7.133233532934132, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1538, "step": 4765 }, { "epoch": 7.134730538922156, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1546, "step": 4766 }, { "epoch": 7.13622754491018, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1484, "step": 4767 }, { "epoch": 7.137724550898204, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1601, "step": 4768 }, { "epoch": 7.139221556886228, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1497, "step": 4769 }, { "epoch": 7.140718562874252, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1504, "step": 4770 }, { "epoch": 7.142215568862276, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4771 }, { "epoch": 7.1437125748503, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1549, "step": 4772 }, { "epoch": 7.145209580838324, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1534, "step": 4773 }, { "epoch": 7.1467065868263475, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1618, "step": 4774 }, { "epoch": 7.1482035928143715, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1539, "step": 4775 }, { "epoch": 7.149700598802395, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1516, "step": 4776 }, { "epoch": 7.151197604790419, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1566, "step": 4777 }, { "epoch": 7.152694610778443, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1668, "step": 4778 }, { "epoch": 7.154191616766467, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1537, "step": 4779 }, { "epoch": 7.155688622754491, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.157, "step": 4780 }, { "epoch": 7.157185628742515, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1592, "step": 4781 }, { "epoch": 7.158682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1589, "step": 4782 }, { "epoch": 7.160179640718563, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1576, "step": 4783 }, { "epoch": 7.161676646706587, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1563, "step": 4784 }, { "epoch": 7.163173652694611, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1587, "step": 4785 }, { "epoch": 7.164670658682635, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1552, "step": 4786 }, { "epoch": 7.166167664670659, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1552, "step": 4787 }, { "epoch": 7.167664670658683, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1511, "step": 4788 }, { "epoch": 7.169161676646707, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1534, "step": 4789 }, { "epoch": 7.1706586826347305, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1609, "step": 4790 }, { "epoch": 7.172155688622754, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1517, "step": 4791 }, { "epoch": 7.173652694610778, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1526, "step": 4792 }, { "epoch": 7.175149700598802, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1555, "step": 4793 }, { "epoch": 7.176646706586826, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1546, "step": 4794 }, { "epoch": 7.17814371257485, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1558, "step": 4795 }, { "epoch": 7.179640718562874, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1475, "step": 4796 }, { "epoch": 7.181137724550898, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1588, "step": 4797 }, { "epoch": 7.182634730538922, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1516, "step": 4798 }, { "epoch": 7.184131736526946, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1535, "step": 4799 }, { "epoch": 7.18562874251497, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1521, "step": 4800 }, { "epoch": 7.187125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1533, "step": 4801 }, { "epoch": 7.188622754491018, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.159, "step": 4802 }, { "epoch": 7.190119760479042, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1566, "step": 4803 }, { "epoch": 7.191616766467066, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1496, "step": 4804 }, { "epoch": 7.1931137724550895, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1581, "step": 4805 }, { "epoch": 7.1946107784431135, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1511, "step": 4806 }, { "epoch": 7.196107784431137, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1503, "step": 4807 }, { "epoch": 7.197604790419161, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1568, "step": 4808 }, { "epoch": 7.199101796407185, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1549, "step": 4809 }, { "epoch": 7.200598802395209, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1572, "step": 4810 }, { "epoch": 7.202095808383233, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1534, "step": 4811 }, { "epoch": 7.203592814371257, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1575, "step": 4812 }, { "epoch": 7.205089820359281, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1495, "step": 4813 }, { "epoch": 7.206586826347305, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1621, "step": 4814 }, { "epoch": 7.20808383233533, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1536, "step": 4815 }, { "epoch": 7.209580838323353, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.157, "step": 4816 }, { "epoch": 7.211077844311378, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1567, "step": 4817 }, { "epoch": 7.212574850299402, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1612, "step": 4818 }, { "epoch": 7.2140718562874255, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1609, "step": 4819 }, { "epoch": 7.2155688622754495, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1537, "step": 4820 }, { "epoch": 7.217065868263473, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1504, "step": 4821 }, { "epoch": 7.218562874251497, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1577, "step": 4822 }, { "epoch": 7.220059880239521, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1554, "step": 4823 }, { "epoch": 7.221556886227545, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1477, "step": 4824 }, { "epoch": 7.223053892215569, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1529, "step": 4825 }, { "epoch": 7.224550898203593, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1577, "step": 4826 }, { "epoch": 7.226047904191617, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1526, "step": 4827 }, { "epoch": 7.227544910179641, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1559, "step": 4828 }, { "epoch": 7.229041916167665, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1482, "step": 4829 }, { "epoch": 7.230538922155689, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1467, "step": 4830 }, { "epoch": 7.232035928143713, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1556, "step": 4831 }, { "epoch": 7.233532934131737, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1531, "step": 4832 }, { "epoch": 7.235029940119761, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1585, "step": 4833 }, { "epoch": 7.236526946107785, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1571, "step": 4834 }, { "epoch": 7.2380239520958085, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1588, "step": 4835 }, { "epoch": 7.2395209580838324, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1494, "step": 4836 }, { "epoch": 7.241017964071856, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1565, "step": 4837 }, { "epoch": 7.24251497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1589, "step": 4838 }, { "epoch": 7.244011976047904, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1554, "step": 4839 }, { "epoch": 7.245508982035928, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1621, "step": 4840 }, { "epoch": 7.247005988023952, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1549, "step": 4841 }, { "epoch": 7.248502994011976, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1494, "step": 4842 }, { "epoch": 7.25, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.158, "step": 4843 }, { "epoch": 7.251497005988024, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1578, "step": 4844 }, { "epoch": 7.252994011976048, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1535, "step": 4845 }, { "epoch": 7.254491017964072, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1522, "step": 4846 }, { "epoch": 7.255988023952096, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1529, "step": 4847 }, { "epoch": 7.25748502994012, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1543, "step": 4848 }, { "epoch": 7.258982035928144, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1499, "step": 4849 }, { "epoch": 7.2604790419161676, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1507, "step": 4850 }, { "epoch": 7.2619760479041915, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1522, "step": 4851 }, { "epoch": 7.263473053892215, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1579, "step": 4852 }, { "epoch": 7.264970059880239, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1562, "step": 4853 }, { "epoch": 7.266467065868263, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1587, "step": 4854 }, { "epoch": 7.267964071856287, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1567, "step": 4855 }, { "epoch": 7.269461077844311, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1528, "step": 4856 }, { "epoch": 7.270958083832335, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1474, "step": 4857 }, { "epoch": 7.272455089820359, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1577, "step": 4858 }, { "epoch": 7.273952095808383, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1507, "step": 4859 }, { "epoch": 7.275449101796407, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1529, "step": 4860 }, { "epoch": 7.276946107784431, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1519, "step": 4861 }, { "epoch": 7.278443113772455, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1512, "step": 4862 }, { "epoch": 7.279940119760479, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1605, "step": 4863 }, { "epoch": 7.281437125748503, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1574, "step": 4864 }, { "epoch": 7.282934131736527, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1557, "step": 4865 }, { "epoch": 7.2844311377245505, "grad_norm": 0.050048828125, "learning_rate": 0.0008, "loss": 1.1523, "step": 4866 }, { "epoch": 7.2859281437125745, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1522, "step": 4867 }, { "epoch": 7.287425149700598, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1508, "step": 4868 }, { "epoch": 7.288922155688622, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1553, "step": 4869 }, { "epoch": 7.290419161676647, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1511, "step": 4870 }, { "epoch": 7.29191616766467, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1581, "step": 4871 }, { "epoch": 7.293413173652695, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1597, "step": 4872 }, { "epoch": 7.294910179640719, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1465, "step": 4873 }, { "epoch": 7.296407185628743, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1543, "step": 4874 }, { "epoch": 7.297904191616767, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1465, "step": 4875 }, { "epoch": 7.299401197604791, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.152, "step": 4876 }, { "epoch": 7.300898203592815, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1587, "step": 4877 }, { "epoch": 7.302395209580839, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1497, "step": 4878 }, { "epoch": 7.303892215568863, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1579, "step": 4879 }, { "epoch": 7.3053892215568865, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1602, "step": 4880 }, { "epoch": 7.3068862275449105, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4881 }, { "epoch": 7.308383233532934, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1496, "step": 4882 }, { "epoch": 7.309880239520958, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.155, "step": 4883 }, { "epoch": 7.311377245508982, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1537, "step": 4884 }, { "epoch": 7.312874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1519, "step": 4885 }, { "epoch": 7.31437125748503, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1578, "step": 4886 }, { "epoch": 7.315868263473054, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1556, "step": 4887 }, { "epoch": 7.317365269461078, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1555, "step": 4888 }, { "epoch": 7.318862275449102, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1554, "step": 4889 }, { "epoch": 7.320359281437126, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1546, "step": 4890 }, { "epoch": 7.32185628742515, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1576, "step": 4891 }, { "epoch": 7.323353293413174, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1531, "step": 4892 }, { "epoch": 7.324850299401198, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1521, "step": 4893 }, { "epoch": 7.326347305389222, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1536, "step": 4894 }, { "epoch": 7.327844311377246, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.15, "step": 4895 }, { "epoch": 7.3293413173652695, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1538, "step": 4896 }, { "epoch": 7.330838323353293, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1549, "step": 4897 }, { "epoch": 7.332335329341317, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.159, "step": 4898 }, { "epoch": 7.333832335329341, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1529, "step": 4899 }, { "epoch": 7.335329341317365, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1522, "step": 4900 }, { "epoch": 7.336826347305389, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1528, "step": 4901 }, { "epoch": 7.338323353293413, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.158, "step": 4902 }, { "epoch": 7.339820359281437, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1536, "step": 4903 }, { "epoch": 7.341317365269461, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1512, "step": 4904 }, { "epoch": 7.342814371257485, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1505, "step": 4905 }, { "epoch": 7.344311377245509, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1591, "step": 4906 }, { "epoch": 7.345808383233533, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1577, "step": 4907 }, { "epoch": 7.347305389221557, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1586, "step": 4908 }, { "epoch": 7.348802395209581, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1479, "step": 4909 }, { "epoch": 7.350299401197605, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1548, "step": 4910 }, { "epoch": 7.3517964071856285, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1536, "step": 4911 }, { "epoch": 7.3532934131736525, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4912 }, { "epoch": 7.354790419161676, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1525, "step": 4913 }, { "epoch": 7.3562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1569, "step": 4914 }, { "epoch": 7.357784431137724, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1532, "step": 4915 }, { "epoch": 7.359281437125748, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1605, "step": 4916 }, { "epoch": 7.360778443113772, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.155, "step": 4917 }, { "epoch": 7.362275449101796, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1516, "step": 4918 }, { "epoch": 7.36377245508982, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1624, "step": 4919 }, { "epoch": 7.365269461077844, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1643, "step": 4920 }, { "epoch": 7.366766467065868, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.154, "step": 4921 }, { "epoch": 7.368263473053892, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1529, "step": 4922 }, { "epoch": 7.369760479041916, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1547, "step": 4923 }, { "epoch": 7.37125748502994, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1528, "step": 4924 }, { "epoch": 7.3727544910179645, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1562, "step": 4925 }, { "epoch": 7.374251497005988, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1527, "step": 4926 }, { "epoch": 7.375748502994012, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1542, "step": 4927 }, { "epoch": 7.3772455089820355, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1605, "step": 4928 }, { "epoch": 7.37874251497006, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.157, "step": 4929 }, { "epoch": 7.380239520958084, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1589, "step": 4930 }, { "epoch": 7.381736526946108, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4931 }, { "epoch": 7.383233532934132, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1519, "step": 4932 }, { "epoch": 7.384730538922156, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1595, "step": 4933 }, { "epoch": 7.38622754491018, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1609, "step": 4934 }, { "epoch": 7.387724550898204, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1568, "step": 4935 }, { "epoch": 7.389221556886228, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1547, "step": 4936 }, { "epoch": 7.390718562874252, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4937 }, { "epoch": 7.392215568862276, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1567, "step": 4938 }, { "epoch": 7.3937125748503, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1606, "step": 4939 }, { "epoch": 7.395209580838324, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1541, "step": 4940 }, { "epoch": 7.3967065868263475, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1536, "step": 4941 }, { "epoch": 7.3982035928143715, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.156, "step": 4942 }, { "epoch": 7.399700598802395, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1561, "step": 4943 }, { "epoch": 7.401197604790419, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1527, "step": 4944 }, { "epoch": 7.402694610778443, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1549, "step": 4945 }, { "epoch": 7.404191616766467, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1525, "step": 4946 }, { "epoch": 7.405688622754491, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1526, "step": 4947 }, { "epoch": 7.407185628742515, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1601, "step": 4948 }, { "epoch": 7.408682634730539, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1565, "step": 4949 }, { "epoch": 7.410179640718563, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1547, "step": 4950 }, { "epoch": 7.411676646706587, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1625, "step": 4951 }, { "epoch": 7.413173652694611, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1565, "step": 4952 }, { "epoch": 7.414670658682635, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1525, "step": 4953 }, { "epoch": 7.416167664670659, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1553, "step": 4954 }, { "epoch": 7.417664670658683, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1555, "step": 4955 }, { "epoch": 7.419161676646707, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1535, "step": 4956 }, { "epoch": 7.4206586826347305, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1544, "step": 4957 }, { "epoch": 7.422155688622754, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1524, "step": 4958 }, { "epoch": 7.423652694610778, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1517, "step": 4959 }, { "epoch": 7.425149700598802, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1557, "step": 4960 }, { "epoch": 7.426646706586826, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1564, "step": 4961 }, { "epoch": 7.42814371257485, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1551, "step": 4962 }, { "epoch": 7.429640718562874, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1564, "step": 4963 }, { "epoch": 7.431137724550898, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1555, "step": 4964 }, { "epoch": 7.432634730538922, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1546, "step": 4965 }, { "epoch": 7.434131736526946, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1525, "step": 4966 }, { "epoch": 7.43562874251497, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1527, "step": 4967 }, { "epoch": 7.437125748502994, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1528, "step": 4968 }, { "epoch": 7.438622754491018, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1581, "step": 4969 }, { "epoch": 7.440119760479042, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1575, "step": 4970 }, { "epoch": 7.441616766467066, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1509, "step": 4971 }, { "epoch": 7.4431137724550895, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1593, "step": 4972 }, { "epoch": 7.4446107784431135, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1572, "step": 4973 }, { "epoch": 7.446107784431137, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1537, "step": 4974 }, { "epoch": 7.447604790419161, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1556, "step": 4975 }, { "epoch": 7.449101796407185, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1523, "step": 4976 }, { "epoch": 7.450598802395209, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1577, "step": 4977 }, { "epoch": 7.452095808383233, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1539, "step": 4978 }, { "epoch": 7.453592814371257, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4979 }, { "epoch": 7.455089820359281, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1529, "step": 4980 }, { "epoch": 7.456586826347305, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.154, "step": 4981 }, { "epoch": 7.45808383233533, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1578, "step": 4982 }, { "epoch": 7.459580838323353, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1541, "step": 4983 }, { "epoch": 7.461077844311378, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1547, "step": 4984 }, { "epoch": 7.462574850299402, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1493, "step": 4985 }, { "epoch": 7.4640718562874255, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1526, "step": 4986 }, { "epoch": 7.4655688622754495, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1585, "step": 4987 }, { "epoch": 7.467065868263473, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.151, "step": 4988 }, { "epoch": 7.468562874251497, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.148, "step": 4989 }, { "epoch": 7.470059880239521, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1526, "step": 4990 }, { "epoch": 7.471556886227545, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1556, "step": 4991 }, { "epoch": 7.473053892215569, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1493, "step": 4992 }, { "epoch": 7.474550898203593, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1503, "step": 4993 }, { "epoch": 7.476047904191617, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1577, "step": 4994 }, { "epoch": 7.477544910179641, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1559, "step": 4995 }, { "epoch": 7.479041916167665, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.161, "step": 4996 }, { "epoch": 7.480538922155689, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1527, "step": 4997 }, { "epoch": 7.482035928143713, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1498, "step": 4998 }, { "epoch": 7.483532934131737, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1608, "step": 4999 }, { "epoch": 7.485029940119761, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1498, "step": 5000 }, { "epoch": 7.486526946107785, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1603, "step": 5001 }, { "epoch": 7.4880239520958085, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1504, "step": 5002 }, { "epoch": 7.4895209580838324, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1586, "step": 5003 }, { "epoch": 7.491017964071856, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1525, "step": 5004 }, { "epoch": 7.49251497005988, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1533, "step": 5005 }, { "epoch": 7.494011976047904, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1571, "step": 5006 }, { "epoch": 7.495508982035928, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1496, "step": 5007 }, { "epoch": 7.497005988023952, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1509, "step": 5008 }, { "epoch": 7.498502994011976, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1565, "step": 5009 }, { "epoch": 7.5, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1549, "step": 5010 }, { "epoch": 7.501497005988024, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1522, "step": 5011 }, { "epoch": 7.502994011976048, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1506, "step": 5012 }, { "epoch": 7.504491017964072, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1601, "step": 5013 }, { "epoch": 7.505988023952096, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1518, "step": 5014 }, { "epoch": 7.50748502994012, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1482, "step": 5015 }, { "epoch": 7.508982035928144, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1543, "step": 5016 }, { "epoch": 7.5104790419161676, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1515, "step": 5017 }, { "epoch": 7.5119760479041915, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1535, "step": 5018 }, { "epoch": 7.513473053892215, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1588, "step": 5019 }, { "epoch": 7.514970059880239, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1551, "step": 5020 }, { "epoch": 7.516467065868263, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1561, "step": 5021 }, { "epoch": 7.517964071856287, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1597, "step": 5022 }, { "epoch": 7.519461077844311, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1508, "step": 5023 }, { "epoch": 7.520958083832335, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1607, "step": 5024 }, { "epoch": 7.522455089820359, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1558, "step": 5025 }, { "epoch": 7.523952095808383, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1571, "step": 5026 }, { "epoch": 7.525449101796407, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1478, "step": 5027 }, { "epoch": 7.526946107784431, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1558, "step": 5028 }, { "epoch": 7.528443113772455, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1474, "step": 5029 }, { "epoch": 7.529940119760479, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1594, "step": 5030 }, { "epoch": 7.531437125748503, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1532, "step": 5031 }, { "epoch": 7.532934131736527, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1543, "step": 5032 }, { "epoch": 7.5344311377245505, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1494, "step": 5033 }, { "epoch": 7.5359281437125745, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.151, "step": 5034 }, { "epoch": 7.537425149700599, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1571, "step": 5035 }, { "epoch": 7.538922155688622, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1534, "step": 5036 }, { "epoch": 7.540419161676647, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.154, "step": 5037 }, { "epoch": 7.54191616766467, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.153, "step": 5038 }, { "epoch": 7.543413173652695, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1552, "step": 5039 }, { "epoch": 7.544910179640718, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1517, "step": 5040 }, { "epoch": 7.546407185628743, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1564, "step": 5041 }, { "epoch": 7.547904191616767, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1511, "step": 5042 }, { "epoch": 7.549401197604791, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1621, "step": 5043 }, { "epoch": 7.550898203592815, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.1553, "step": 5044 }, { "epoch": 7.552395209580839, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1554, "step": 5045 }, { "epoch": 7.553892215568863, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1514, "step": 5046 }, { "epoch": 7.5553892215568865, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1526, "step": 5047 }, { "epoch": 7.5568862275449105, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1536, "step": 5048 }, { "epoch": 7.558383233532934, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5049 }, { "epoch": 7.559880239520958, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1516, "step": 5050 }, { "epoch": 7.561377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1502, "step": 5051 }, { "epoch": 7.562874251497006, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1632, "step": 5052 }, { "epoch": 7.56437125748503, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1573, "step": 5053 }, { "epoch": 7.565868263473054, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1562, "step": 5054 }, { "epoch": 7.567365269461078, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1554, "step": 5055 }, { "epoch": 7.568862275449102, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1486, "step": 5056 }, { "epoch": 7.570359281437126, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1515, "step": 5057 }, { "epoch": 7.57185628742515, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1573, "step": 5058 }, { "epoch": 7.573353293413174, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1492, "step": 5059 }, { "epoch": 7.574850299401198, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1541, "step": 5060 }, { "epoch": 7.576347305389222, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1565, "step": 5061 }, { "epoch": 7.577844311377246, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1528, "step": 5062 }, { "epoch": 7.5793413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1573, "step": 5063 }, { "epoch": 7.580838323353293, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1485, "step": 5064 }, { "epoch": 7.582335329341317, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.158, "step": 5065 }, { "epoch": 7.583832335329341, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1546, "step": 5066 }, { "epoch": 7.585329341317365, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1594, "step": 5067 }, { "epoch": 7.586826347305389, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1464, "step": 5068 }, { "epoch": 7.588323353293413, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1609, "step": 5069 }, { "epoch": 7.589820359281437, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1507, "step": 5070 }, { "epoch": 7.591317365269461, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1529, "step": 5071 }, { "epoch": 7.592814371257485, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1569, "step": 5072 }, { "epoch": 7.594311377245509, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1539, "step": 5073 }, { "epoch": 7.595808383233533, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1567, "step": 5074 }, { "epoch": 7.597305389221557, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1541, "step": 5075 }, { "epoch": 7.598802395209581, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1552, "step": 5076 }, { "epoch": 7.600299401197605, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1561, "step": 5077 }, { "epoch": 7.6017964071856285, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1497, "step": 5078 }, { "epoch": 7.6032934131736525, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1526, "step": 5079 }, { "epoch": 7.604790419161676, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1509, "step": 5080 }, { "epoch": 7.6062874251497, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1587, "step": 5081 }, { "epoch": 7.607784431137724, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1575, "step": 5082 }, { "epoch": 7.609281437125748, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1558, "step": 5083 }, { "epoch": 7.610778443113772, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1531, "step": 5084 }, { "epoch": 7.612275449101796, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1543, "step": 5085 }, { "epoch": 7.61377245508982, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1513, "step": 5086 }, { "epoch": 7.615269461077844, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1526, "step": 5087 }, { "epoch": 7.616766467065868, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1542, "step": 5088 }, { "epoch": 7.618263473053892, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1566, "step": 5089 }, { "epoch": 7.619760479041916, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1576, "step": 5090 }, { "epoch": 7.62125748502994, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1569, "step": 5091 }, { "epoch": 7.6227544910179645, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1499, "step": 5092 }, { "epoch": 7.624251497005988, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1519, "step": 5093 }, { "epoch": 7.625748502994012, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1571, "step": 5094 }, { "epoch": 7.6272455089820355, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1593, "step": 5095 }, { "epoch": 7.62874251497006, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1534, "step": 5096 }, { "epoch": 7.630239520958084, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1558, "step": 5097 }, { "epoch": 7.631736526946108, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1581, "step": 5098 }, { "epoch": 7.633233532934132, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1498, "step": 5099 }, { "epoch": 7.634730538922156, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1544, "step": 5100 }, { "epoch": 7.63622754491018, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1543, "step": 5101 }, { "epoch": 7.637724550898204, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1543, "step": 5102 }, { "epoch": 7.639221556886228, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1584, "step": 5103 }, { "epoch": 7.640718562874252, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1593, "step": 5104 }, { "epoch": 7.642215568862276, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1502, "step": 5105 }, { "epoch": 7.6437125748503, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1489, "step": 5106 }, { "epoch": 7.645209580838324, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1513, "step": 5107 }, { "epoch": 7.6467065868263475, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1491, "step": 5108 }, { "epoch": 7.6482035928143715, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1459, "step": 5109 }, { "epoch": 7.649700598802395, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.153, "step": 5110 }, { "epoch": 7.651197604790419, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1512, "step": 5111 }, { "epoch": 7.652694610778443, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1509, "step": 5112 }, { "epoch": 7.654191616766467, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1611, "step": 5113 }, { "epoch": 7.655688622754491, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1487, "step": 5114 }, { "epoch": 7.657185628742515, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1462, "step": 5115 }, { "epoch": 7.658682634730539, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.149, "step": 5116 }, { "epoch": 7.660179640718563, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1611, "step": 5117 }, { "epoch": 7.661676646706587, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1514, "step": 5118 }, { "epoch": 7.663173652694611, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1589, "step": 5119 }, { "epoch": 7.664670658682635, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1586, "step": 5120 }, { "epoch": 7.666167664670659, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1599, "step": 5121 }, { "epoch": 7.667664670658683, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5122 }, { "epoch": 7.669161676646707, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1478, "step": 5123 }, { "epoch": 7.6706586826347305, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1525, "step": 5124 }, { "epoch": 7.672155688622754, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.141, "step": 5125 }, { "epoch": 7.673652694610778, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1471, "step": 5126 }, { "epoch": 7.675149700598802, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5127 }, { "epoch": 7.676646706586826, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1426, "step": 5128 }, { "epoch": 7.67814371257485, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1522, "step": 5129 }, { "epoch": 7.679640718562874, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1531, "step": 5130 }, { "epoch": 7.681137724550898, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1545, "step": 5131 }, { "epoch": 7.682634730538922, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.1594, "step": 5132 }, { "epoch": 7.684131736526946, "grad_norm": 0.20703125, "learning_rate": 0.0008, "loss": 1.1525, "step": 5133 }, { "epoch": 7.68562874251497, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1554, "step": 5134 }, { "epoch": 7.687125748502994, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.1445, "step": 5135 }, { "epoch": 7.688622754491018, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1553, "step": 5136 }, { "epoch": 7.690119760479042, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1538, "step": 5137 }, { "epoch": 7.691616766467066, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1503, "step": 5138 }, { "epoch": 7.6931137724550895, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1522, "step": 5139 }, { "epoch": 7.6946107784431135, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1517, "step": 5140 }, { "epoch": 7.696107784431137, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1503, "step": 5141 }, { "epoch": 7.697604790419161, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1572, "step": 5142 }, { "epoch": 7.699101796407185, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1553, "step": 5143 }, { "epoch": 7.700598802395209, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1573, "step": 5144 }, { "epoch": 7.702095808383233, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1563, "step": 5145 }, { "epoch": 7.703592814371257, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1511, "step": 5146 }, { "epoch": 7.705089820359282, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1617, "step": 5147 }, { "epoch": 7.706586826347305, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1532, "step": 5148 }, { "epoch": 7.70808383233533, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1493, "step": 5149 }, { "epoch": 7.709580838323353, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1498, "step": 5150 }, { "epoch": 7.711077844311378, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1504, "step": 5151 }, { "epoch": 7.712574850299401, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1604, "step": 5152 }, { "epoch": 7.7140718562874255, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1522, "step": 5153 }, { "epoch": 7.7155688622754495, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1548, "step": 5154 }, { "epoch": 7.717065868263473, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1512, "step": 5155 }, { "epoch": 7.718562874251497, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1542, "step": 5156 }, { "epoch": 7.720059880239521, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1612, "step": 5157 }, { "epoch": 7.721556886227545, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1566, "step": 5158 }, { "epoch": 7.723053892215569, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1565, "step": 5159 }, { "epoch": 7.724550898203593, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1512, "step": 5160 }, { "epoch": 7.726047904191617, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1535, "step": 5161 }, { "epoch": 7.727544910179641, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1553, "step": 5162 }, { "epoch": 7.729041916167665, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1561, "step": 5163 }, { "epoch": 7.730538922155689, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1491, "step": 5164 }, { "epoch": 7.732035928143713, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1479, "step": 5165 }, { "epoch": 7.733532934131737, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1564, "step": 5166 }, { "epoch": 7.735029940119761, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1539, "step": 5167 }, { "epoch": 7.736526946107785, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1571, "step": 5168 }, { "epoch": 7.7380239520958085, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1563, "step": 5169 }, { "epoch": 7.7395209580838324, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1483, "step": 5170 }, { "epoch": 7.741017964071856, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1503, "step": 5171 }, { "epoch": 7.74251497005988, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1554, "step": 5172 }, { "epoch": 7.744011976047904, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1518, "step": 5173 }, { "epoch": 7.745508982035928, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1496, "step": 5174 }, { "epoch": 7.747005988023952, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1557, "step": 5175 }, { "epoch": 7.748502994011976, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1551, "step": 5176 }, { "epoch": 7.75, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1593, "step": 5177 }, { "epoch": 7.751497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1452, "step": 5178 }, { "epoch": 7.752994011976048, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1509, "step": 5179 }, { "epoch": 7.754491017964072, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1499, "step": 5180 }, { "epoch": 7.755988023952096, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.146, "step": 5181 }, { "epoch": 7.75748502994012, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1512, "step": 5182 }, { "epoch": 7.758982035928144, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1612, "step": 5183 }, { "epoch": 7.7604790419161676, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1576, "step": 5184 }, { "epoch": 7.7619760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1499, "step": 5185 }, { "epoch": 7.763473053892215, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1542, "step": 5186 }, { "epoch": 7.764970059880239, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1563, "step": 5187 }, { "epoch": 7.766467065868263, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1486, "step": 5188 }, { "epoch": 7.767964071856287, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1566, "step": 5189 }, { "epoch": 7.769461077844311, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1623, "step": 5190 }, { "epoch": 7.770958083832335, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1573, "step": 5191 }, { "epoch": 7.772455089820359, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1501, "step": 5192 }, { "epoch": 7.773952095808383, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1585, "step": 5193 }, { "epoch": 7.775449101796407, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1511, "step": 5194 }, { "epoch": 7.776946107784431, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1552, "step": 5195 }, { "epoch": 7.778443113772455, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1528, "step": 5196 }, { "epoch": 7.779940119760479, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1542, "step": 5197 }, { "epoch": 7.781437125748503, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1513, "step": 5198 }, { "epoch": 7.782934131736527, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1529, "step": 5199 }, { "epoch": 7.7844311377245505, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1542, "step": 5200 }, { "epoch": 7.7859281437125745, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.155, "step": 5201 }, { "epoch": 7.787425149700599, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1542, "step": 5202 }, { "epoch": 7.788922155688622, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1554, "step": 5203 }, { "epoch": 7.790419161676647, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1474, "step": 5204 }, { "epoch": 7.79191616766467, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1557, "step": 5205 }, { "epoch": 7.793413173652695, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1521, "step": 5206 }, { "epoch": 7.794910179640718, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1564, "step": 5207 }, { "epoch": 7.796407185628743, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.146, "step": 5208 }, { "epoch": 7.797904191616767, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1574, "step": 5209 }, { "epoch": 7.799401197604791, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1561, "step": 5210 }, { "epoch": 7.800898203592815, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1508, "step": 5211 }, { "epoch": 7.802395209580839, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1517, "step": 5212 }, { "epoch": 7.803892215568863, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1525, "step": 5213 }, { "epoch": 7.8053892215568865, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1514, "step": 5214 }, { "epoch": 7.8068862275449105, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1517, "step": 5215 }, { "epoch": 7.808383233532934, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1433, "step": 5216 }, { "epoch": 7.809880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1537, "step": 5217 }, { "epoch": 7.811377245508982, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5218 }, { "epoch": 7.812874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1565, "step": 5219 }, { "epoch": 7.81437125748503, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1511, "step": 5220 }, { "epoch": 7.815868263473054, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1471, "step": 5221 }, { "epoch": 7.817365269461078, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1552, "step": 5222 }, { "epoch": 7.818862275449102, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1455, "step": 5223 }, { "epoch": 7.820359281437126, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1499, "step": 5224 }, { "epoch": 7.82185628742515, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1558, "step": 5225 }, { "epoch": 7.823353293413174, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1515, "step": 5226 }, { "epoch": 7.824850299401198, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1523, "step": 5227 }, { "epoch": 7.826347305389222, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1542, "step": 5228 }, { "epoch": 7.827844311377246, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1532, "step": 5229 }, { "epoch": 7.8293413173652695, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.152, "step": 5230 }, { "epoch": 7.830838323353293, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5231 }, { "epoch": 7.832335329341317, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.151, "step": 5232 }, { "epoch": 7.833832335329341, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1586, "step": 5233 }, { "epoch": 7.835329341317365, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1484, "step": 5234 }, { "epoch": 7.836826347305389, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1487, "step": 5235 }, { "epoch": 7.838323353293413, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1498, "step": 5236 }, { "epoch": 7.839820359281437, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1504, "step": 5237 }, { "epoch": 7.841317365269461, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1531, "step": 5238 }, { "epoch": 7.842814371257485, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1551, "step": 5239 }, { "epoch": 7.844311377245509, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1478, "step": 5240 }, { "epoch": 7.845808383233533, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1506, "step": 5241 }, { "epoch": 7.847305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1521, "step": 5242 }, { "epoch": 7.848802395209581, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1473, "step": 5243 }, { "epoch": 7.850299401197605, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1546, "step": 5244 }, { "epoch": 7.8517964071856285, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1568, "step": 5245 }, { "epoch": 7.8532934131736525, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1507, "step": 5246 }, { "epoch": 7.854790419161676, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1453, "step": 5247 }, { "epoch": 7.8562874251497, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1583, "step": 5248 }, { "epoch": 7.857784431137724, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1497, "step": 5249 }, { "epoch": 7.859281437125748, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1478, "step": 5250 }, { "epoch": 7.860778443113772, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1493, "step": 5251 }, { "epoch": 7.862275449101796, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1561, "step": 5252 }, { "epoch": 7.86377245508982, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1557, "step": 5253 }, { "epoch": 7.865269461077844, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1601, "step": 5254 }, { "epoch": 7.866766467065868, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1528, "step": 5255 }, { "epoch": 7.868263473053892, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1501, "step": 5256 }, { "epoch": 7.869760479041916, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1586, "step": 5257 }, { "epoch": 7.87125748502994, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1545, "step": 5258 }, { "epoch": 7.8727544910179645, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1576, "step": 5259 }, { "epoch": 7.874251497005988, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1532, "step": 5260 }, { "epoch": 7.875748502994012, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1481, "step": 5261 }, { "epoch": 7.8772455089820355, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1514, "step": 5262 }, { "epoch": 7.87874251497006, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1514, "step": 5263 }, { "epoch": 7.880239520958084, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1575, "step": 5264 }, { "epoch": 7.881736526946108, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1581, "step": 5265 }, { "epoch": 7.883233532934132, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1552, "step": 5266 }, { "epoch": 7.884730538922156, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1505, "step": 5267 }, { "epoch": 7.88622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1509, "step": 5268 }, { "epoch": 7.887724550898204, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.152, "step": 5269 }, { "epoch": 7.889221556886228, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1553, "step": 5270 }, { "epoch": 7.890718562874252, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1582, "step": 5271 }, { "epoch": 7.892215568862276, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.156, "step": 5272 }, { "epoch": 7.8937125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1514, "step": 5273 }, { "epoch": 7.895209580838324, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1585, "step": 5274 }, { "epoch": 7.8967065868263475, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1558, "step": 5275 }, { "epoch": 7.8982035928143715, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1551, "step": 5276 }, { "epoch": 7.899700598802395, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1466, "step": 5277 }, { "epoch": 7.901197604790419, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1516, "step": 5278 }, { "epoch": 7.902694610778443, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1582, "step": 5279 }, { "epoch": 7.904191616766467, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.153, "step": 5280 }, { "epoch": 7.905688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1525, "step": 5281 }, { "epoch": 7.907185628742515, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1518, "step": 5282 }, { "epoch": 7.908682634730539, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1496, "step": 5283 }, { "epoch": 7.910179640718563, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5284 }, { "epoch": 7.911676646706587, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1541, "step": 5285 }, { "epoch": 7.913173652694611, "grad_norm": 0.049072265625, "learning_rate": 0.0008, "loss": 1.1513, "step": 5286 }, { "epoch": 7.914670658682635, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1522, "step": 5287 }, { "epoch": 7.916167664670659, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1534, "step": 5288 }, { "epoch": 7.917664670658683, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1559, "step": 5289 }, { "epoch": 7.919161676646707, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1511, "step": 5290 }, { "epoch": 7.9206586826347305, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1527, "step": 5291 }, { "epoch": 7.922155688622754, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1553, "step": 5292 }, { "epoch": 7.923652694610778, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1525, "step": 5293 }, { "epoch": 7.925149700598802, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1488, "step": 5294 }, { "epoch": 7.926646706586826, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1505, "step": 5295 }, { "epoch": 7.92814371257485, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.152, "step": 5296 }, { "epoch": 7.929640718562874, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1554, "step": 5297 }, { "epoch": 7.931137724550898, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1462, "step": 5298 }, { "epoch": 7.932634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1425, "step": 5299 }, { "epoch": 7.934131736526946, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1528, "step": 5300 }, { "epoch": 7.93562874251497, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1516, "step": 5301 }, { "epoch": 7.937125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1424, "step": 5302 }, { "epoch": 7.938622754491018, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1437, "step": 5303 }, { "epoch": 7.940119760479042, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1446, "step": 5304 }, { "epoch": 7.941616766467066, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1502, "step": 5305 }, { "epoch": 7.9431137724550895, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1485, "step": 5306 }, { "epoch": 7.9446107784431135, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1468, "step": 5307 }, { "epoch": 7.946107784431137, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1496, "step": 5308 }, { "epoch": 7.947604790419161, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1501, "step": 5309 }, { "epoch": 7.949101796407185, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1516, "step": 5310 }, { "epoch": 7.950598802395209, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.152, "step": 5311 }, { "epoch": 7.952095808383233, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1606, "step": 5312 }, { "epoch": 7.953592814371257, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1541, "step": 5313 }, { "epoch": 7.955089820359282, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1521, "step": 5314 }, { "epoch": 7.956586826347305, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1478, "step": 5315 }, { "epoch": 7.95808383233533, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1562, "step": 5316 }, { "epoch": 7.959580838323353, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.154, "step": 5317 }, { "epoch": 7.961077844311378, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1518, "step": 5318 }, { "epoch": 7.962574850299401, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1548, "step": 5319 }, { "epoch": 7.9640718562874255, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1507, "step": 5320 }, { "epoch": 7.9655688622754495, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1539, "step": 5321 }, { "epoch": 7.967065868263473, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1498, "step": 5322 }, { "epoch": 7.968562874251497, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1556, "step": 5323 }, { "epoch": 7.970059880239521, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1482, "step": 5324 }, { "epoch": 7.971556886227545, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1545, "step": 5325 }, { "epoch": 7.973053892215569, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1581, "step": 5326 }, { "epoch": 7.974550898203593, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1533, "step": 5327 }, { "epoch": 7.976047904191617, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1531, "step": 5328 }, { "epoch": 7.977544910179641, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1522, "step": 5329 }, { "epoch": 7.979041916167665, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1603, "step": 5330 }, { "epoch": 7.980538922155689, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1488, "step": 5331 }, { "epoch": 7.982035928143713, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.153, "step": 5332 }, { "epoch": 7.983532934131737, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1584, "step": 5333 }, { "epoch": 7.985029940119761, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.143, "step": 5334 }, { "epoch": 7.986526946107785, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1529, "step": 5335 }, { "epoch": 7.9880239520958085, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1563, "step": 5336 }, { "epoch": 7.9895209580838324, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1513, "step": 5337 }, { "epoch": 7.991017964071856, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1549, "step": 5338 }, { "epoch": 7.99251497005988, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1519, "step": 5339 }, { "epoch": 7.994011976047904, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1513, "step": 5340 }, { "epoch": 7.995508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1549, "step": 5341 }, { "epoch": 7.997005988023952, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1427, "step": 5342 }, { "epoch": 7.998502994011976, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1478, "step": 5343 }, { "epoch": 8.0, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1479, "step": 5344 }, { "epoch": 8.001497005988025, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1474, "step": 5345 }, { "epoch": 8.002994011976048, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1516, "step": 5346 }, { "epoch": 8.004491017964073, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1493, "step": 5347 }, { "epoch": 8.005988023952096, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1444, "step": 5348 }, { "epoch": 8.00748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1553, "step": 5349 }, { "epoch": 8.008982035928144, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1521, "step": 5350 }, { "epoch": 8.010479041916168, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1549, "step": 5351 }, { "epoch": 8.011976047904191, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1572, "step": 5352 }, { "epoch": 8.013473053892216, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5353 }, { "epoch": 8.01497005988024, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1518, "step": 5354 }, { "epoch": 8.016467065868264, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1525, "step": 5355 }, { "epoch": 8.017964071856287, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1541, "step": 5356 }, { "epoch": 8.019461077844312, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1514, "step": 5357 }, { "epoch": 8.020958083832335, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1521, "step": 5358 }, { "epoch": 8.02245508982036, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1489, "step": 5359 }, { "epoch": 8.023952095808383, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1494, "step": 5360 }, { "epoch": 8.025449101796408, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1511, "step": 5361 }, { "epoch": 8.02694610778443, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1567, "step": 5362 }, { "epoch": 8.028443113772456, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1485, "step": 5363 }, { "epoch": 8.029940119760479, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1529, "step": 5364 }, { "epoch": 8.031437125748504, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1472, "step": 5365 }, { "epoch": 8.032934131736527, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.156, "step": 5366 }, { "epoch": 8.034431137724551, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1605, "step": 5367 }, { "epoch": 8.035928143712574, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1501, "step": 5368 }, { "epoch": 8.0374251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.15, "step": 5369 }, { "epoch": 8.038922155688622, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1559, "step": 5370 }, { "epoch": 8.040419161676647, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1546, "step": 5371 }, { "epoch": 8.04191616766467, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1456, "step": 5372 }, { "epoch": 8.043413173652695, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 5373 }, { "epoch": 8.044910179640718, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1536, "step": 5374 }, { "epoch": 8.046407185628743, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1474, "step": 5375 }, { "epoch": 8.047904191616766, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1511, "step": 5376 }, { "epoch": 8.04940119760479, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1578, "step": 5377 }, { "epoch": 8.050898203592814, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1604, "step": 5378 }, { "epoch": 8.052395209580839, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1497, "step": 5379 }, { "epoch": 8.053892215568862, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.144, "step": 5380 }, { "epoch": 8.055389221556887, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1554, "step": 5381 }, { "epoch": 8.05688622754491, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1545, "step": 5382 }, { "epoch": 8.058383233532934, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.158, "step": 5383 }, { "epoch": 8.059880239520957, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1507, "step": 5384 }, { "epoch": 8.061377245508982, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1529, "step": 5385 }, { "epoch": 8.062874251497005, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1544, "step": 5386 }, { "epoch": 8.06437125748503, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1453, "step": 5387 }, { "epoch": 8.065868263473053, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1517, "step": 5388 }, { "epoch": 8.067365269461078, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1453, "step": 5389 }, { "epoch": 8.068862275449101, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1427, "step": 5390 }, { "epoch": 8.070359281437126, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1447, "step": 5391 }, { "epoch": 8.071856287425149, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1503, "step": 5392 }, { "epoch": 8.073353293413174, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1553, "step": 5393 }, { "epoch": 8.074850299401197, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1554, "step": 5394 }, { "epoch": 8.076347305389222, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1499, "step": 5395 }, { "epoch": 8.077844311377245, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1489, "step": 5396 }, { "epoch": 8.07934131736527, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.149, "step": 5397 }, { "epoch": 8.080838323353293, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1492, "step": 5398 }, { "epoch": 8.082335329341317, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1553, "step": 5399 }, { "epoch": 8.08383233532934, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1495, "step": 5400 }, { "epoch": 8.085329341317365, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1517, "step": 5401 }, { "epoch": 8.08682634730539, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1506, "step": 5402 }, { "epoch": 8.088323353293413, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1427, "step": 5403 }, { "epoch": 8.089820359281438, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1527, "step": 5404 }, { "epoch": 8.091317365269461, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1444, "step": 5405 }, { "epoch": 8.092814371257486, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1481, "step": 5406 }, { "epoch": 8.094311377245509, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1546, "step": 5407 }, { "epoch": 8.095808383233534, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1541, "step": 5408 }, { "epoch": 8.097305389221557, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1474, "step": 5409 }, { "epoch": 8.098802395209582, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1569, "step": 5410 }, { "epoch": 8.100299401197605, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1499, "step": 5411 }, { "epoch": 8.10179640718563, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1511, "step": 5412 }, { "epoch": 8.103293413173652, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1558, "step": 5413 }, { "epoch": 8.104790419161677, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1492, "step": 5414 }, { "epoch": 8.1062874251497, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.1491, "step": 5415 }, { "epoch": 8.107784431137725, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1527, "step": 5416 }, { "epoch": 8.109281437125748, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1494, "step": 5417 }, { "epoch": 8.110778443113773, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1574, "step": 5418 }, { "epoch": 8.112275449101796, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1582, "step": 5419 }, { "epoch": 8.113772455089821, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.1508, "step": 5420 }, { "epoch": 8.115269461077844, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1577, "step": 5421 }, { "epoch": 8.116766467065869, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1436, "step": 5422 }, { "epoch": 8.118263473053892, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1602, "step": 5423 }, { "epoch": 8.119760479041917, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1535, "step": 5424 }, { "epoch": 8.12125748502994, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1466, "step": 5425 }, { "epoch": 8.122754491017965, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1475, "step": 5426 }, { "epoch": 8.124251497005988, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.152, "step": 5427 }, { "epoch": 8.125748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.152, "step": 5428 }, { "epoch": 8.127245508982035, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1501, "step": 5429 }, { "epoch": 8.12874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1518, "step": 5430 }, { "epoch": 8.130239520958083, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1504, "step": 5431 }, { "epoch": 8.131736526946108, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1496, "step": 5432 }, { "epoch": 8.133233532934131, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1576, "step": 5433 }, { "epoch": 8.134730538922156, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1551, "step": 5434 }, { "epoch": 8.136227544910179, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1482, "step": 5435 }, { "epoch": 8.137724550898204, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.155, "step": 5436 }, { "epoch": 8.139221556886227, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1517, "step": 5437 }, { "epoch": 8.140718562874252, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.1565, "step": 5438 }, { "epoch": 8.142215568862275, "grad_norm": 0.1943359375, "learning_rate": 0.0008, "loss": 1.1551, "step": 5439 }, { "epoch": 8.1437125748503, "grad_norm": 0.220703125, "learning_rate": 0.0008, "loss": 1.1553, "step": 5440 }, { "epoch": 8.145209580838323, "grad_norm": 0.2412109375, "learning_rate": 0.0008, "loss": 1.1542, "step": 5441 }, { "epoch": 8.146706586826348, "grad_norm": 0.25, "learning_rate": 0.0008, "loss": 1.1539, "step": 5442 }, { "epoch": 8.14820359281437, "grad_norm": 0.232421875, "learning_rate": 0.0008, "loss": 1.157, "step": 5443 }, { "epoch": 8.149700598802395, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.152, "step": 5444 }, { "epoch": 8.151197604790418, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1522, "step": 5445 }, { "epoch": 8.152694610778443, "grad_norm": 0.212890625, "learning_rate": 0.0008, "loss": 1.1497, "step": 5446 }, { "epoch": 8.154191616766466, "grad_norm": 0.28515625, "learning_rate": 0.0008, "loss": 1.158, "step": 5447 }, { "epoch": 8.155688622754491, "grad_norm": 0.408203125, "learning_rate": 0.0008, "loss": 1.1523, "step": 5448 }, { "epoch": 8.157185628742514, "grad_norm": 0.734375, "learning_rate": 0.0008, "loss": 1.1533, "step": 5449 }, { "epoch": 8.158682634730539, "grad_norm": 1.28125, "learning_rate": 0.0008, "loss": 1.1855, "step": 5450 }, { "epoch": 8.160179640718562, "grad_norm": 0.671875, "learning_rate": 0.0008, "loss": 1.1727, "step": 5451 }, { "epoch": 8.161676646706587, "grad_norm": 0.369140625, "learning_rate": 0.0008, "loss": 1.161, "step": 5452 }, { "epoch": 8.16317365269461, "grad_norm": 0.5625, "learning_rate": 0.0008, "loss": 1.1676, "step": 5453 }, { "epoch": 8.164670658682635, "grad_norm": 0.1748046875, "learning_rate": 0.0008, "loss": 1.1607, "step": 5454 }, { "epoch": 8.16616766467066, "grad_norm": 0.47265625, "learning_rate": 0.0008, "loss": 1.1718, "step": 5455 }, { "epoch": 8.167664670658683, "grad_norm": 0.376953125, "learning_rate": 0.0008, "loss": 1.1675, "step": 5456 }, { "epoch": 8.169161676646707, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.1579, "step": 5457 }, { "epoch": 8.17065868263473, "grad_norm": 0.53125, "learning_rate": 0.0008, "loss": 1.162, "step": 5458 }, { "epoch": 8.172155688622755, "grad_norm": 0.37109375, "learning_rate": 0.0008, "loss": 1.1661, "step": 5459 }, { "epoch": 8.173652694610778, "grad_norm": 0.18359375, "learning_rate": 0.0008, "loss": 1.1658, "step": 5460 }, { "epoch": 8.175149700598803, "grad_norm": 0.28125, "learning_rate": 0.0008, "loss": 1.1597, "step": 5461 }, { "epoch": 8.176646706586826, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.1678, "step": 5462 }, { "epoch": 8.178143712574851, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.1571, "step": 5463 }, { "epoch": 8.179640718562874, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.1588, "step": 5464 }, { "epoch": 8.181137724550899, "grad_norm": 0.30078125, "learning_rate": 0.0008, "loss": 1.1642, "step": 5465 }, { "epoch": 8.182634730538922, "grad_norm": 0.1748046875, "learning_rate": 0.0008, "loss": 1.1643, "step": 5466 }, { "epoch": 8.184131736526947, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1508, "step": 5467 }, { "epoch": 8.18562874251497, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1541, "step": 5468 }, { "epoch": 8.187125748502995, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1534, "step": 5469 }, { "epoch": 8.188622754491018, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1591, "step": 5470 }, { "epoch": 8.190119760479043, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1576, "step": 5471 }, { "epoch": 8.191616766467066, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.15, "step": 5472 }, { "epoch": 8.19311377245509, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1573, "step": 5473 }, { "epoch": 8.194610778443113, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1593, "step": 5474 }, { "epoch": 8.196107784431138, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1545, "step": 5475 }, { "epoch": 8.197604790419161, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1544, "step": 5476 }, { "epoch": 8.199101796407186, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1593, "step": 5477 }, { "epoch": 8.20059880239521, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1527, "step": 5478 }, { "epoch": 8.202095808383234, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1515, "step": 5479 }, { "epoch": 8.203592814371257, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1555, "step": 5480 }, { "epoch": 8.205089820359282, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1517, "step": 5481 }, { "epoch": 8.206586826347305, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1513, "step": 5482 }, { "epoch": 8.20808383233533, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1593, "step": 5483 }, { "epoch": 8.209580838323353, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1579, "step": 5484 }, { "epoch": 8.211077844311378, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1579, "step": 5485 }, { "epoch": 8.2125748502994, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1486, "step": 5486 }, { "epoch": 8.214071856287426, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1512, "step": 5487 }, { "epoch": 8.215568862275449, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1549, "step": 5488 }, { "epoch": 8.217065868263473, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1556, "step": 5489 }, { "epoch": 8.218562874251496, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1563, "step": 5490 }, { "epoch": 8.220059880239521, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1567, "step": 5491 }, { "epoch": 8.221556886227544, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1577, "step": 5492 }, { "epoch": 8.22305389221557, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1519, "step": 5493 }, { "epoch": 8.224550898203592, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1469, "step": 5494 }, { "epoch": 8.226047904191617, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1571, "step": 5495 }, { "epoch": 8.22754491017964, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1567, "step": 5496 }, { "epoch": 8.229041916167665, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1536, "step": 5497 }, { "epoch": 8.230538922155688, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1527, "step": 5498 }, { "epoch": 8.232035928143713, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1541, "step": 5499 }, { "epoch": 8.233532934131736, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1611, "step": 5500 }, { "epoch": 8.23502994011976, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1486, "step": 5501 }, { "epoch": 8.236526946107784, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1554, "step": 5502 }, { "epoch": 8.238023952095809, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1534, "step": 5503 }, { "epoch": 8.239520958083832, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1541, "step": 5504 }, { "epoch": 8.241017964071856, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1515, "step": 5505 }, { "epoch": 8.24251497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1506, "step": 5506 }, { "epoch": 8.244011976047904, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1526, "step": 5507 }, { "epoch": 8.245508982035927, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1509, "step": 5508 }, { "epoch": 8.247005988023952, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1493, "step": 5509 }, { "epoch": 8.248502994011975, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1502, "step": 5510 }, { "epoch": 8.25, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1541, "step": 5511 }, { "epoch": 8.251497005988025, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.156, "step": 5512 }, { "epoch": 8.252994011976048, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1565, "step": 5513 }, { "epoch": 8.254491017964073, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1494, "step": 5514 }, { "epoch": 8.255988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1593, "step": 5515 }, { "epoch": 8.25748502994012, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1456, "step": 5516 }, { "epoch": 8.258982035928144, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1524, "step": 5517 }, { "epoch": 8.260479041916168, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1567, "step": 5518 }, { "epoch": 8.261976047904191, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1553, "step": 5519 }, { "epoch": 8.263473053892216, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.148, "step": 5520 }, { "epoch": 8.26497005988024, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1478, "step": 5521 }, { "epoch": 8.266467065868264, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1544, "step": 5522 }, { "epoch": 8.267964071856287, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1565, "step": 5523 }, { "epoch": 8.269461077844312, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1589, "step": 5524 }, { "epoch": 8.270958083832335, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1472, "step": 5525 }, { "epoch": 8.27245508982036, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1517, "step": 5526 }, { "epoch": 8.273952095808383, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1571, "step": 5527 }, { "epoch": 8.275449101796408, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1532, "step": 5528 }, { "epoch": 8.27694610778443, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1547, "step": 5529 }, { "epoch": 8.278443113772456, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1555, "step": 5530 }, { "epoch": 8.279940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1502, "step": 5531 }, { "epoch": 8.281437125748504, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1458, "step": 5532 }, { "epoch": 8.282934131736527, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1547, "step": 5533 }, { "epoch": 8.284431137724551, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1564, "step": 5534 }, { "epoch": 8.285928143712574, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1552, "step": 5535 }, { "epoch": 8.2874251497006, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1491, "step": 5536 }, { "epoch": 8.288922155688622, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1544, "step": 5537 }, { "epoch": 8.290419161676647, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1574, "step": 5538 }, { "epoch": 8.29191616766467, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1564, "step": 5539 }, { "epoch": 8.293413173652695, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1497, "step": 5540 }, { "epoch": 8.294910179640718, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1504, "step": 5541 }, { "epoch": 8.296407185628743, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1528, "step": 5542 }, { "epoch": 8.297904191616766, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1551, "step": 5543 }, { "epoch": 8.29940119760479, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1453, "step": 5544 }, { "epoch": 8.300898203592814, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1516, "step": 5545 }, { "epoch": 8.302395209580839, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1491, "step": 5546 }, { "epoch": 8.303892215568862, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1527, "step": 5547 }, { "epoch": 8.305389221556887, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1456, "step": 5548 }, { "epoch": 8.30688622754491, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1493, "step": 5549 }, { "epoch": 8.308383233532934, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1559, "step": 5550 }, { "epoch": 8.309880239520957, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1513, "step": 5551 }, { "epoch": 8.311377245508982, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1523, "step": 5552 }, { "epoch": 8.312874251497005, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1485, "step": 5553 }, { "epoch": 8.31437125748503, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1471, "step": 5554 }, { "epoch": 8.315868263473053, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1499, "step": 5555 }, { "epoch": 8.317365269461078, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.153, "step": 5556 }, { "epoch": 8.318862275449101, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1568, "step": 5557 }, { "epoch": 8.320359281437126, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.152, "step": 5558 }, { "epoch": 8.321856287425149, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1586, "step": 5559 }, { "epoch": 8.323353293413174, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1481, "step": 5560 }, { "epoch": 8.324850299401197, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1534, "step": 5561 }, { "epoch": 8.326347305389222, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1518, "step": 5562 }, { "epoch": 8.327844311377245, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1518, "step": 5563 }, { "epoch": 8.32934131736527, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1461, "step": 5564 }, { "epoch": 8.330838323353294, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1486, "step": 5565 }, { "epoch": 8.332335329341317, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1489, "step": 5566 }, { "epoch": 8.33383233532934, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1555, "step": 5567 }, { "epoch": 8.335329341317365, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1509, "step": 5568 }, { "epoch": 8.33682634730539, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1534, "step": 5569 }, { "epoch": 8.338323353293413, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1462, "step": 5570 }, { "epoch": 8.339820359281438, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1486, "step": 5571 }, { "epoch": 8.341317365269461, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1485, "step": 5572 }, { "epoch": 8.342814371257486, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5573 }, { "epoch": 8.344311377245509, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1522, "step": 5574 }, { "epoch": 8.345808383233534, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1529, "step": 5575 }, { "epoch": 8.347305389221557, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.15, "step": 5576 }, { "epoch": 8.348802395209582, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1466, "step": 5577 }, { "epoch": 8.350299401197605, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1472, "step": 5578 }, { "epoch": 8.35179640718563, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1547, "step": 5579 }, { "epoch": 8.353293413173652, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1489, "step": 5580 }, { "epoch": 8.354790419161677, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.153, "step": 5581 }, { "epoch": 8.3562874251497, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1581, "step": 5582 }, { "epoch": 8.357784431137725, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.157, "step": 5583 }, { "epoch": 8.359281437125748, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1478, "step": 5584 }, { "epoch": 8.360778443113773, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.157, "step": 5585 }, { "epoch": 8.362275449101796, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1498, "step": 5586 }, { "epoch": 8.363772455089821, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1492, "step": 5587 }, { "epoch": 8.365269461077844, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1453, "step": 5588 }, { "epoch": 8.366766467065869, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5589 }, { "epoch": 8.368263473053892, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1554, "step": 5590 }, { "epoch": 8.369760479041917, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.154, "step": 5591 }, { "epoch": 8.37125748502994, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1544, "step": 5592 }, { "epoch": 8.372754491017965, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1461, "step": 5593 }, { "epoch": 8.374251497005988, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1433, "step": 5594 }, { "epoch": 8.375748502994012, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1476, "step": 5595 }, { "epoch": 8.377245508982035, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1573, "step": 5596 }, { "epoch": 8.37874251497006, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1473, "step": 5597 }, { "epoch": 8.380239520958083, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1523, "step": 5598 }, { "epoch": 8.381736526946108, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1473, "step": 5599 }, { "epoch": 8.383233532934131, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1486, "step": 5600 }, { "epoch": 8.384730538922156, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1549, "step": 5601 }, { "epoch": 8.386227544910179, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1504, "step": 5602 }, { "epoch": 8.387724550898204, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1488, "step": 5603 }, { "epoch": 8.389221556886227, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1531, "step": 5604 }, { "epoch": 8.390718562874252, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1478, "step": 5605 }, { "epoch": 8.392215568862275, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1507, "step": 5606 }, { "epoch": 8.3937125748503, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1525, "step": 5607 }, { "epoch": 8.395209580838323, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1492, "step": 5608 }, { "epoch": 8.396706586826348, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1487, "step": 5609 }, { "epoch": 8.39820359281437, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1452, "step": 5610 }, { "epoch": 8.399700598802395, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1528, "step": 5611 }, { "epoch": 8.401197604790418, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1514, "step": 5612 }, { "epoch": 8.402694610778443, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1494, "step": 5613 }, { "epoch": 8.404191616766466, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1539, "step": 5614 }, { "epoch": 8.405688622754491, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1501, "step": 5615 }, { "epoch": 8.407185628742514, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1563, "step": 5616 }, { "epoch": 8.408682634730539, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1519, "step": 5617 }, { "epoch": 8.410179640718562, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1488, "step": 5618 }, { "epoch": 8.411676646706587, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1523, "step": 5619 }, { "epoch": 8.41317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1416, "step": 5620 }, { "epoch": 8.414670658682635, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1507, "step": 5621 }, { "epoch": 8.41616766467066, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1473, "step": 5622 }, { "epoch": 8.417664670658683, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1524, "step": 5623 }, { "epoch": 8.419161676646706, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1498, "step": 5624 }, { "epoch": 8.42065868263473, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1524, "step": 5625 }, { "epoch": 8.422155688622755, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1534, "step": 5626 }, { "epoch": 8.423652694610778, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1474, "step": 5627 }, { "epoch": 8.425149700598803, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1517, "step": 5628 }, { "epoch": 8.426646706586826, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1534, "step": 5629 }, { "epoch": 8.428143712574851, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1504, "step": 5630 }, { "epoch": 8.429640718562874, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1539, "step": 5631 }, { "epoch": 8.431137724550899, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1491, "step": 5632 }, { "epoch": 8.432634730538922, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1558, "step": 5633 }, { "epoch": 8.434131736526947, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1562, "step": 5634 }, { "epoch": 8.43562874251497, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1487, "step": 5635 }, { "epoch": 8.437125748502995, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1565, "step": 5636 }, { "epoch": 8.438622754491018, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1532, "step": 5637 }, { "epoch": 8.440119760479043, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1482, "step": 5638 }, { "epoch": 8.441616766467066, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.149, "step": 5639 }, { "epoch": 8.44311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1571, "step": 5640 }, { "epoch": 8.444610778443113, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1571, "step": 5641 }, { "epoch": 8.446107784431138, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1557, "step": 5642 }, { "epoch": 8.447604790419161, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1536, "step": 5643 }, { "epoch": 8.449101796407186, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1581, "step": 5644 }, { "epoch": 8.45059880239521, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1584, "step": 5645 }, { "epoch": 8.452095808383234, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1471, "step": 5646 }, { "epoch": 8.453592814371257, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1532, "step": 5647 }, { "epoch": 8.455089820359282, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1581, "step": 5648 }, { "epoch": 8.456586826347305, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1542, "step": 5649 }, { "epoch": 8.45808383233533, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1441, "step": 5650 }, { "epoch": 8.459580838323353, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1511, "step": 5651 }, { "epoch": 8.461077844311378, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1556, "step": 5652 }, { "epoch": 8.4625748502994, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.153, "step": 5653 }, { "epoch": 8.464071856287426, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1537, "step": 5654 }, { "epoch": 8.465568862275449, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5655 }, { "epoch": 8.467065868263473, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1428, "step": 5656 }, { "epoch": 8.468562874251496, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1452, "step": 5657 }, { "epoch": 8.470059880239521, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1479, "step": 5658 }, { "epoch": 8.471556886227544, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1569, "step": 5659 }, { "epoch": 8.47305389221557, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1533, "step": 5660 }, { "epoch": 8.474550898203592, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1509, "step": 5661 }, { "epoch": 8.476047904191617, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1473, "step": 5662 }, { "epoch": 8.47754491017964, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1553, "step": 5663 }, { "epoch": 8.479041916167665, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1464, "step": 5664 }, { "epoch": 8.480538922155688, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1563, "step": 5665 }, { "epoch": 8.482035928143713, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1521, "step": 5666 }, { "epoch": 8.483532934131736, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1492, "step": 5667 }, { "epoch": 8.48502994011976, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.151, "step": 5668 }, { "epoch": 8.486526946107784, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1552, "step": 5669 }, { "epoch": 8.488023952095809, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1397, "step": 5670 }, { "epoch": 8.489520958083832, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1578, "step": 5671 }, { "epoch": 8.491017964071856, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1505, "step": 5672 }, { "epoch": 8.49251497005988, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1484, "step": 5673 }, { "epoch": 8.494011976047904, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1488, "step": 5674 }, { "epoch": 8.495508982035927, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.151, "step": 5675 }, { "epoch": 8.497005988023952, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1497, "step": 5676 }, { "epoch": 8.498502994011975, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1541, "step": 5677 }, { "epoch": 8.5, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1582, "step": 5678 }, { "epoch": 8.501497005988025, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1533, "step": 5679 }, { "epoch": 8.502994011976048, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1493, "step": 5680 }, { "epoch": 8.504491017964071, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1544, "step": 5681 }, { "epoch": 8.505988023952096, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1498, "step": 5682 }, { "epoch": 8.50748502994012, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1507, "step": 5683 }, { "epoch": 8.508982035928144, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1444, "step": 5684 }, { "epoch": 8.510479041916168, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1445, "step": 5685 }, { "epoch": 8.511976047904191, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1501, "step": 5686 }, { "epoch": 8.513473053892216, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1503, "step": 5687 }, { "epoch": 8.51497005988024, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1521, "step": 5688 }, { "epoch": 8.516467065868264, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1551, "step": 5689 }, { "epoch": 8.517964071856287, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1501, "step": 5690 }, { "epoch": 8.519461077844312, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1469, "step": 5691 }, { "epoch": 8.520958083832335, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1495, "step": 5692 }, { "epoch": 8.52245508982036, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1549, "step": 5693 }, { "epoch": 8.523952095808383, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.154, "step": 5694 }, { "epoch": 8.525449101796408, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1568, "step": 5695 }, { "epoch": 8.52694610778443, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1582, "step": 5696 }, { "epoch": 8.528443113772456, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1446, "step": 5697 }, { "epoch": 8.529940119760479, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1585, "step": 5698 }, { "epoch": 8.531437125748504, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1491, "step": 5699 }, { "epoch": 8.532934131736527, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1443, "step": 5700 }, { "epoch": 8.534431137724551, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1557, "step": 5701 }, { "epoch": 8.535928143712574, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1503, "step": 5702 }, { "epoch": 8.5374251497006, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1495, "step": 5703 }, { "epoch": 8.538922155688622, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1521, "step": 5704 }, { "epoch": 8.540419161676647, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1544, "step": 5705 }, { "epoch": 8.54191616766467, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1512, "step": 5706 }, { "epoch": 8.543413173652695, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1512, "step": 5707 }, { "epoch": 8.544910179640718, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1498, "step": 5708 }, { "epoch": 8.546407185628743, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.147, "step": 5709 }, { "epoch": 8.547904191616766, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1456, "step": 5710 }, { "epoch": 8.54940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1486, "step": 5711 }, { "epoch": 8.550898203592814, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1469, "step": 5712 }, { "epoch": 8.552395209580839, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.155, "step": 5713 }, { "epoch": 8.553892215568862, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1618, "step": 5714 }, { "epoch": 8.555389221556887, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1589, "step": 5715 }, { "epoch": 8.55688622754491, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1527, "step": 5716 }, { "epoch": 8.558383233532934, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.148, "step": 5717 }, { "epoch": 8.559880239520957, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1568, "step": 5718 }, { "epoch": 8.561377245508982, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1543, "step": 5719 }, { "epoch": 8.562874251497005, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1536, "step": 5720 }, { "epoch": 8.56437125748503, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1504, "step": 5721 }, { "epoch": 8.565868263473053, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.15, "step": 5722 }, { "epoch": 8.567365269461078, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.154, "step": 5723 }, { "epoch": 8.568862275449101, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1512, "step": 5724 }, { "epoch": 8.570359281437126, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1513, "step": 5725 }, { "epoch": 8.571856287425149, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1517, "step": 5726 }, { "epoch": 8.573353293413174, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1509, "step": 5727 }, { "epoch": 8.574850299401197, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.153, "step": 5728 }, { "epoch": 8.576347305389222, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1501, "step": 5729 }, { "epoch": 8.577844311377245, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1547, "step": 5730 }, { "epoch": 8.57934131736527, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1552, "step": 5731 }, { "epoch": 8.580838323353294, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1497, "step": 5732 }, { "epoch": 8.582335329341317, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5733 }, { "epoch": 8.58383233532934, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.151, "step": 5734 }, { "epoch": 8.585329341317365, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1466, "step": 5735 }, { "epoch": 8.58682634730539, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1453, "step": 5736 }, { "epoch": 8.588323353293413, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1446, "step": 5737 }, { "epoch": 8.589820359281438, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.149, "step": 5738 }, { "epoch": 8.591317365269461, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1565, "step": 5739 }, { "epoch": 8.592814371257486, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1449, "step": 5740 }, { "epoch": 8.594311377245509, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1419, "step": 5741 }, { "epoch": 8.595808383233534, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1559, "step": 5742 }, { "epoch": 8.597305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1506, "step": 5743 }, { "epoch": 8.598802395209582, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1542, "step": 5744 }, { "epoch": 8.600299401197605, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1483, "step": 5745 }, { "epoch": 8.60179640718563, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1474, "step": 5746 }, { "epoch": 8.603293413173652, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1457, "step": 5747 }, { "epoch": 8.604790419161677, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5748 }, { "epoch": 8.6062874251497, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1472, "step": 5749 }, { "epoch": 8.607784431137725, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1496, "step": 5750 }, { "epoch": 8.609281437125748, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1503, "step": 5751 }, { "epoch": 8.610778443113773, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.142, "step": 5752 }, { "epoch": 8.612275449101796, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1484, "step": 5753 }, { "epoch": 8.613772455089821, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.156, "step": 5754 }, { "epoch": 8.615269461077844, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1499, "step": 5755 }, { "epoch": 8.616766467065869, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1475, "step": 5756 }, { "epoch": 8.618263473053892, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1487, "step": 5757 }, { "epoch": 8.619760479041917, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1507, "step": 5758 }, { "epoch": 8.62125748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1488, "step": 5759 }, { "epoch": 8.622754491017965, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1506, "step": 5760 }, { "epoch": 8.624251497005988, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1508, "step": 5761 }, { "epoch": 8.625748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1503, "step": 5762 }, { "epoch": 8.627245508982035, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1467, "step": 5763 }, { "epoch": 8.62874251497006, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1567, "step": 5764 }, { "epoch": 8.630239520958083, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1519, "step": 5765 }, { "epoch": 8.631736526946108, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1489, "step": 5766 }, { "epoch": 8.633233532934131, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1604, "step": 5767 }, { "epoch": 8.634730538922156, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1516, "step": 5768 }, { "epoch": 8.636227544910179, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1492, "step": 5769 }, { "epoch": 8.637724550898204, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.148, "step": 5770 }, { "epoch": 8.639221556886227, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1454, "step": 5771 }, { "epoch": 8.640718562874252, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1461, "step": 5772 }, { "epoch": 8.642215568862275, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1507, "step": 5773 }, { "epoch": 8.6437125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1552, "step": 5774 }, { "epoch": 8.645209580838323, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1455, "step": 5775 }, { "epoch": 8.646706586826348, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1466, "step": 5776 }, { "epoch": 8.64820359281437, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1539, "step": 5777 }, { "epoch": 8.649700598802395, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1553, "step": 5778 }, { "epoch": 8.651197604790418, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 5779 }, { "epoch": 8.652694610778443, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1518, "step": 5780 }, { "epoch": 8.654191616766466, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1473, "step": 5781 }, { "epoch": 8.655688622754491, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1477, "step": 5782 }, { "epoch": 8.657185628742514, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1498, "step": 5783 }, { "epoch": 8.658682634730539, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1515, "step": 5784 }, { "epoch": 8.660179640718562, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 5785 }, { "epoch": 8.661676646706587, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.15, "step": 5786 }, { "epoch": 8.66317365269461, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.158, "step": 5787 }, { "epoch": 8.664670658682635, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.15, "step": 5788 }, { "epoch": 8.66616766467066, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.15, "step": 5789 }, { "epoch": 8.667664670658683, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1457, "step": 5790 }, { "epoch": 8.669161676646706, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1472, "step": 5791 }, { "epoch": 8.67065868263473, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5792 }, { "epoch": 8.672155688622755, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1501, "step": 5793 }, { "epoch": 8.673652694610778, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1484, "step": 5794 }, { "epoch": 8.675149700598803, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1516, "step": 5795 }, { "epoch": 8.676646706586826, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1515, "step": 5796 }, { "epoch": 8.678143712574851, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.155, "step": 5797 }, { "epoch": 8.679640718562874, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1472, "step": 5798 }, { "epoch": 8.681137724550899, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.155, "step": 5799 }, { "epoch": 8.682634730538922, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.145, "step": 5800 }, { "epoch": 8.684131736526947, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1572, "step": 5801 }, { "epoch": 8.68562874251497, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1469, "step": 5802 }, { "epoch": 8.687125748502995, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1439, "step": 5803 }, { "epoch": 8.688622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1436, "step": 5804 }, { "epoch": 8.690119760479043, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1466, "step": 5805 }, { "epoch": 8.691616766467066, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.153, "step": 5806 }, { "epoch": 8.69311377245509, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1537, "step": 5807 }, { "epoch": 8.694610778443113, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1456, "step": 5808 }, { "epoch": 8.696107784431138, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1486, "step": 5809 }, { "epoch": 8.697604790419161, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1407, "step": 5810 }, { "epoch": 8.699101796407186, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1515, "step": 5811 }, { "epoch": 8.70059880239521, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1511, "step": 5812 }, { "epoch": 8.702095808383234, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1521, "step": 5813 }, { "epoch": 8.703592814371257, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1502, "step": 5814 }, { "epoch": 8.705089820359282, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1508, "step": 5815 }, { "epoch": 8.706586826347305, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.146, "step": 5816 }, { "epoch": 8.70808383233533, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1452, "step": 5817 }, { "epoch": 8.709580838323353, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1532, "step": 5818 }, { "epoch": 8.711077844311378, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1482, "step": 5819 }, { "epoch": 8.7125748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1485, "step": 5820 }, { "epoch": 8.714071856287426, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.157, "step": 5821 }, { "epoch": 8.715568862275449, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.155, "step": 5822 }, { "epoch": 8.717065868263473, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1456, "step": 5823 }, { "epoch": 8.718562874251496, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.15, "step": 5824 }, { "epoch": 8.720059880239521, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1495, "step": 5825 }, { "epoch": 8.721556886227544, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.15, "step": 5826 }, { "epoch": 8.72305389221557, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1487, "step": 5827 }, { "epoch": 8.724550898203592, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1496, "step": 5828 }, { "epoch": 8.726047904191617, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1447, "step": 5829 }, { "epoch": 8.72754491017964, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1494, "step": 5830 }, { "epoch": 8.729041916167665, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1492, "step": 5831 }, { "epoch": 8.730538922155688, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1538, "step": 5832 }, { "epoch": 8.732035928143713, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1534, "step": 5833 }, { "epoch": 8.733532934131736, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1464, "step": 5834 }, { "epoch": 8.73502994011976, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1423, "step": 5835 }, { "epoch": 8.736526946107784, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1531, "step": 5836 }, { "epoch": 8.738023952095809, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1541, "step": 5837 }, { "epoch": 8.739520958083832, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5838 }, { "epoch": 8.741017964071856, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1404, "step": 5839 }, { "epoch": 8.74251497005988, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1575, "step": 5840 }, { "epoch": 8.744011976047904, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1534, "step": 5841 }, { "epoch": 8.745508982035929, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1514, "step": 5842 }, { "epoch": 8.747005988023952, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1451, "step": 5843 }, { "epoch": 8.748502994011975, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1469, "step": 5844 }, { "epoch": 8.75, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1486, "step": 5845 }, { "epoch": 8.751497005988025, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1485, "step": 5846 }, { "epoch": 8.752994011976048, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1507, "step": 5847 }, { "epoch": 8.754491017964071, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1491, "step": 5848 }, { "epoch": 8.755988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1504, "step": 5849 }, { "epoch": 8.75748502994012, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1456, "step": 5850 }, { "epoch": 8.758982035928144, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1499, "step": 5851 }, { "epoch": 8.760479041916168, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1556, "step": 5852 }, { "epoch": 8.761976047904191, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1484, "step": 5853 }, { "epoch": 8.763473053892216, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1536, "step": 5854 }, { "epoch": 8.76497005988024, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1485, "step": 5855 }, { "epoch": 8.766467065868264, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1493, "step": 5856 }, { "epoch": 8.767964071856287, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1511, "step": 5857 }, { "epoch": 8.769461077844312, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1478, "step": 5858 }, { "epoch": 8.770958083832335, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1536, "step": 5859 }, { "epoch": 8.77245508982036, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1544, "step": 5860 }, { "epoch": 8.773952095808383, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1555, "step": 5861 }, { "epoch": 8.775449101796408, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1465, "step": 5862 }, { "epoch": 8.77694610778443, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.158, "step": 5863 }, { "epoch": 8.778443113772456, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1471, "step": 5864 }, { "epoch": 8.779940119760479, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.1502, "step": 5865 }, { "epoch": 8.781437125748504, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1498, "step": 5866 }, { "epoch": 8.782934131736527, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5867 }, { "epoch": 8.784431137724551, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.153, "step": 5868 }, { "epoch": 8.785928143712574, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1461, "step": 5869 }, { "epoch": 8.7874251497006, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1426, "step": 5870 }, { "epoch": 8.788922155688622, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1562, "step": 5871 }, { "epoch": 8.790419161676647, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1555, "step": 5872 }, { "epoch": 8.79191616766467, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.1455, "step": 5873 }, { "epoch": 8.793413173652695, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1514, "step": 5874 }, { "epoch": 8.794910179640718, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1552, "step": 5875 }, { "epoch": 8.796407185628743, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1485, "step": 5876 }, { "epoch": 8.797904191616766, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1542, "step": 5877 }, { "epoch": 8.79940119760479, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1525, "step": 5878 }, { "epoch": 8.800898203592814, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1472, "step": 5879 }, { "epoch": 8.802395209580839, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1497, "step": 5880 }, { "epoch": 8.803892215568862, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1531, "step": 5881 }, { "epoch": 8.805389221556887, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1555, "step": 5882 }, { "epoch": 8.80688622754491, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1483, "step": 5883 }, { "epoch": 8.808383233532934, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1504, "step": 5884 }, { "epoch": 8.809880239520957, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1491, "step": 5885 }, { "epoch": 8.811377245508982, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1511, "step": 5886 }, { "epoch": 8.812874251497005, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1457, "step": 5887 }, { "epoch": 8.81437125748503, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1436, "step": 5888 }, { "epoch": 8.815868263473053, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1566, "step": 5889 }, { "epoch": 8.817365269461078, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1491, "step": 5890 }, { "epoch": 8.818862275449101, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1493, "step": 5891 }, { "epoch": 8.820359281437126, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1512, "step": 5892 }, { "epoch": 8.821856287425149, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1588, "step": 5893 }, { "epoch": 8.823353293413174, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1448, "step": 5894 }, { "epoch": 8.824850299401197, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1486, "step": 5895 }, { "epoch": 8.826347305389222, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1494, "step": 5896 }, { "epoch": 8.827844311377245, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1469, "step": 5897 }, { "epoch": 8.82934131736527, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5898 }, { "epoch": 8.830838323353294, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.15, "step": 5899 }, { "epoch": 8.832335329341317, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1508, "step": 5900 }, { "epoch": 8.83383233532934, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1423, "step": 5901 }, { "epoch": 8.835329341317365, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1518, "step": 5902 }, { "epoch": 8.83682634730539, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1496, "step": 5903 }, { "epoch": 8.838323353293413, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1489, "step": 5904 }, { "epoch": 8.839820359281438, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.153, "step": 5905 }, { "epoch": 8.841317365269461, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1475, "step": 5906 }, { "epoch": 8.842814371257486, "grad_norm": 0.1787109375, "learning_rate": 0.0008, "loss": 1.1523, "step": 5907 }, { "epoch": 8.844311377245509, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1556, "step": 5908 }, { "epoch": 8.845808383233534, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1521, "step": 5909 }, { "epoch": 8.847305389221557, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.153, "step": 5910 }, { "epoch": 8.848802395209582, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.1468, "step": 5911 }, { "epoch": 8.850299401197605, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1502, "step": 5912 }, { "epoch": 8.85179640718563, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1572, "step": 5913 }, { "epoch": 8.853293413173652, "grad_norm": 0.1689453125, "learning_rate": 0.0008, "loss": 1.1499, "step": 5914 }, { "epoch": 8.854790419161677, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.1468, "step": 5915 }, { "epoch": 8.8562874251497, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1477, "step": 5916 }, { "epoch": 8.857784431137725, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.1491, "step": 5917 }, { "epoch": 8.859281437125748, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1539, "step": 5918 }, { "epoch": 8.860778443113773, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.157, "step": 5919 }, { "epoch": 8.862275449101796, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1505, "step": 5920 }, { "epoch": 8.863772455089821, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1522, "step": 5921 }, { "epoch": 8.865269461077844, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1479, "step": 5922 }, { "epoch": 8.866766467065869, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5923 }, { "epoch": 8.868263473053892, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1435, "step": 5924 }, { "epoch": 8.869760479041917, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1486, "step": 5925 }, { "epoch": 8.87125748502994, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1487, "step": 5926 }, { "epoch": 8.872754491017965, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.145, "step": 5927 }, { "epoch": 8.874251497005988, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1503, "step": 5928 }, { "epoch": 8.875748502994012, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1482, "step": 5929 }, { "epoch": 8.877245508982035, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1515, "step": 5930 }, { "epoch": 8.87874251497006, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1562, "step": 5931 }, { "epoch": 8.880239520958083, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1515, "step": 5932 }, { "epoch": 8.881736526946108, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.147, "step": 5933 }, { "epoch": 8.883233532934131, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1477, "step": 5934 }, { "epoch": 8.884730538922156, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.15, "step": 5935 }, { "epoch": 8.886227544910179, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.154, "step": 5936 }, { "epoch": 8.887724550898204, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1493, "step": 5937 }, { "epoch": 8.889221556886227, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1478, "step": 5938 }, { "epoch": 8.890718562874252, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1489, "step": 5939 }, { "epoch": 8.892215568862275, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1535, "step": 5940 }, { "epoch": 8.8937125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1527, "step": 5941 }, { "epoch": 8.895209580838323, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1529, "step": 5942 }, { "epoch": 8.896706586826348, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1513, "step": 5943 }, { "epoch": 8.89820359281437, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1531, "step": 5944 }, { "epoch": 8.899700598802395, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1522, "step": 5945 }, { "epoch": 8.901197604790418, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1452, "step": 5946 }, { "epoch": 8.902694610778443, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1465, "step": 5947 }, { "epoch": 8.904191616766466, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1519, "step": 5948 }, { "epoch": 8.905688622754491, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1397, "step": 5949 }, { "epoch": 8.907185628742514, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1467, "step": 5950 }, { "epoch": 8.908682634730539, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1563, "step": 5951 }, { "epoch": 8.910179640718562, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1566, "step": 5952 }, { "epoch": 8.911676646706587, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1518, "step": 5953 }, { "epoch": 8.91317365269461, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1534, "step": 5954 }, { "epoch": 8.914670658682635, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1558, "step": 5955 }, { "epoch": 8.91616766467066, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5956 }, { "epoch": 8.917664670658683, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1492, "step": 5957 }, { "epoch": 8.919161676646706, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1505, "step": 5958 }, { "epoch": 8.92065868263473, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1516, "step": 5959 }, { "epoch": 8.922155688622755, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1505, "step": 5960 }, { "epoch": 8.923652694610778, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1479, "step": 5961 }, { "epoch": 8.925149700598803, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 5962 }, { "epoch": 8.926646706586826, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1523, "step": 5963 }, { "epoch": 8.928143712574851, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1463, "step": 5964 }, { "epoch": 8.929640718562874, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1444, "step": 5965 }, { "epoch": 8.931137724550899, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1476, "step": 5966 }, { "epoch": 8.932634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1451, "step": 5967 }, { "epoch": 8.934131736526947, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1482, "step": 5968 }, { "epoch": 8.93562874251497, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1501, "step": 5969 }, { "epoch": 8.937125748502995, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1475, "step": 5970 }, { "epoch": 8.938622754491018, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5971 }, { "epoch": 8.940119760479043, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1521, "step": 5972 }, { "epoch": 8.941616766467066, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1555, "step": 5973 }, { "epoch": 8.94311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1535, "step": 5974 }, { "epoch": 8.944610778443113, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1466, "step": 5975 }, { "epoch": 8.946107784431138, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1508, "step": 5976 }, { "epoch": 8.947604790419161, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1509, "step": 5977 }, { "epoch": 8.949101796407186, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1481, "step": 5978 }, { "epoch": 8.95059880239521, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1424, "step": 5979 }, { "epoch": 8.952095808383234, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.144, "step": 5980 }, { "epoch": 8.953592814371257, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1466, "step": 5981 }, { "epoch": 8.955089820359282, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1504, "step": 5982 }, { "epoch": 8.956586826347305, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.16, "step": 5983 }, { "epoch": 8.95808383233533, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1553, "step": 5984 }, { "epoch": 8.959580838323353, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1506, "step": 5985 }, { "epoch": 8.961077844311378, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1524, "step": 5986 }, { "epoch": 8.9625748502994, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1522, "step": 5987 }, { "epoch": 8.964071856287426, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1478, "step": 5988 }, { "epoch": 8.965568862275449, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1485, "step": 5989 }, { "epoch": 8.967065868263473, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.155, "step": 5990 }, { "epoch": 8.968562874251496, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1452, "step": 5991 }, { "epoch": 8.970059880239521, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1563, "step": 5992 }, { "epoch": 8.971556886227544, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1424, "step": 5993 }, { "epoch": 8.97305389221557, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1537, "step": 5994 }, { "epoch": 8.974550898203592, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1447, "step": 5995 }, { "epoch": 8.976047904191617, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1496, "step": 5996 }, { "epoch": 8.97754491017964, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1554, "step": 5997 }, { "epoch": 8.979041916167665, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1557, "step": 5998 }, { "epoch": 8.980538922155688, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1521, "step": 5999 }, { "epoch": 8.982035928143713, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6000 }, { "epoch": 8.983532934131736, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1517, "step": 6001 }, { "epoch": 8.98502994011976, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.153, "step": 6002 }, { "epoch": 8.986526946107784, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1506, "step": 6003 }, { "epoch": 8.988023952095809, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1492, "step": 6004 }, { "epoch": 8.989520958083832, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.145, "step": 6005 }, { "epoch": 8.991017964071856, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1445, "step": 6006 }, { "epoch": 8.99251497005988, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.149, "step": 6007 }, { "epoch": 8.994011976047904, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1478, "step": 6008 }, { "epoch": 8.995508982035929, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1389, "step": 6009 }, { "epoch": 8.997005988023952, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1459, "step": 6010 }, { "epoch": 8.998502994011975, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1472, "step": 6011 }, { "epoch": 9.0, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1453, "step": 6012 }, { "epoch": 9.001497005988025, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1485, "step": 6013 }, { "epoch": 9.002994011976048, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1527, "step": 6014 }, { "epoch": 9.004491017964073, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1485, "step": 6015 }, { "epoch": 9.005988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1451, "step": 6016 }, { "epoch": 9.00748502994012, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1446, "step": 6017 }, { "epoch": 9.008982035928144, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1446, "step": 6018 }, { "epoch": 9.010479041916168, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1507, "step": 6019 }, { "epoch": 9.011976047904191, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1513, "step": 6020 }, { "epoch": 9.013473053892216, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1475, "step": 6021 }, { "epoch": 9.01497005988024, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 6022 }, { "epoch": 9.016467065868264, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1495, "step": 6023 }, { "epoch": 9.017964071856287, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1563, "step": 6024 }, { "epoch": 9.019461077844312, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1521, "step": 6025 }, { "epoch": 9.020958083832335, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1486, "step": 6026 }, { "epoch": 9.02245508982036, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1519, "step": 6027 }, { "epoch": 9.023952095808383, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1488, "step": 6028 }, { "epoch": 9.025449101796408, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1462, "step": 6029 }, { "epoch": 9.02694610778443, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1444, "step": 6030 }, { "epoch": 9.028443113772456, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1512, "step": 6031 }, { "epoch": 9.029940119760479, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6032 }, { "epoch": 9.031437125748504, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1503, "step": 6033 }, { "epoch": 9.032934131736527, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1481, "step": 6034 }, { "epoch": 9.034431137724551, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6035 }, { "epoch": 9.035928143712574, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1539, "step": 6036 }, { "epoch": 9.0374251497006, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1437, "step": 6037 }, { "epoch": 9.038922155688622, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6038 }, { "epoch": 9.040419161676647, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1558, "step": 6039 }, { "epoch": 9.04191616766467, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1449, "step": 6040 }, { "epoch": 9.043413173652695, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1496, "step": 6041 }, { "epoch": 9.044910179640718, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.146, "step": 6042 }, { "epoch": 9.046407185628743, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1496, "step": 6043 }, { "epoch": 9.047904191616766, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1468, "step": 6044 }, { "epoch": 9.04940119760479, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1504, "step": 6045 }, { "epoch": 9.050898203592814, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1533, "step": 6046 }, { "epoch": 9.052395209580839, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1485, "step": 6047 }, { "epoch": 9.053892215568862, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1496, "step": 6048 }, { "epoch": 9.055389221556887, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1497, "step": 6049 }, { "epoch": 9.05688622754491, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1509, "step": 6050 }, { "epoch": 9.058383233532934, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1478, "step": 6051 }, { "epoch": 9.059880239520957, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1359, "step": 6052 }, { "epoch": 9.061377245508982, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1521, "step": 6053 }, { "epoch": 9.062874251497005, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1528, "step": 6054 }, { "epoch": 9.06437125748503, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1458, "step": 6055 }, { "epoch": 9.065868263473053, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1457, "step": 6056 }, { "epoch": 9.067365269461078, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1464, "step": 6057 }, { "epoch": 9.068862275449101, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1456, "step": 6058 }, { "epoch": 9.070359281437126, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1485, "step": 6059 }, { "epoch": 9.071856287425149, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1493, "step": 6060 }, { "epoch": 9.073353293413174, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1427, "step": 6061 }, { "epoch": 9.074850299401197, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1497, "step": 6062 }, { "epoch": 9.076347305389222, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1583, "step": 6063 }, { "epoch": 9.077844311377245, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.145, "step": 6064 }, { "epoch": 9.07934131736527, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1447, "step": 6065 }, { "epoch": 9.080838323353293, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1438, "step": 6066 }, { "epoch": 9.082335329341317, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1545, "step": 6067 }, { "epoch": 9.08383233532934, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1534, "step": 6068 }, { "epoch": 9.085329341317365, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1601, "step": 6069 }, { "epoch": 9.08682634730539, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1495, "step": 6070 }, { "epoch": 9.088323353293413, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.1472, "step": 6071 }, { "epoch": 9.089820359281438, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1492, "step": 6072 }, { "epoch": 9.091317365269461, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1494, "step": 6073 }, { "epoch": 9.092814371257486, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1547, "step": 6074 }, { "epoch": 9.094311377245509, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1479, "step": 6075 }, { "epoch": 9.095808383233534, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1526, "step": 6076 }, { "epoch": 9.097305389221557, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1548, "step": 6077 }, { "epoch": 9.098802395209582, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1477, "step": 6078 }, { "epoch": 9.100299401197605, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1405, "step": 6079 }, { "epoch": 9.10179640718563, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1546, "step": 6080 }, { "epoch": 9.103293413173652, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1466, "step": 6081 }, { "epoch": 9.104790419161677, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1502, "step": 6082 }, { "epoch": 9.1062874251497, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1453, "step": 6083 }, { "epoch": 9.107784431137725, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1468, "step": 6084 }, { "epoch": 9.109281437125748, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1506, "step": 6085 }, { "epoch": 9.110778443113773, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1477, "step": 6086 }, { "epoch": 9.112275449101796, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 6087 }, { "epoch": 9.113772455089821, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1436, "step": 6088 }, { "epoch": 9.115269461077844, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1593, "step": 6089 }, { "epoch": 9.116766467065869, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1517, "step": 6090 }, { "epoch": 9.118263473053892, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.142, "step": 6091 }, { "epoch": 9.119760479041917, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1557, "step": 6092 }, { "epoch": 9.12125748502994, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1469, "step": 6093 }, { "epoch": 9.122754491017965, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1515, "step": 6094 }, { "epoch": 9.124251497005988, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1503, "step": 6095 }, { "epoch": 9.125748502994012, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1503, "step": 6096 }, { "epoch": 9.127245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1435, "step": 6097 }, { "epoch": 9.12874251497006, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1541, "step": 6098 }, { "epoch": 9.130239520958083, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1507, "step": 6099 }, { "epoch": 9.131736526946108, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1433, "step": 6100 }, { "epoch": 9.133233532934131, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.153, "step": 6101 }, { "epoch": 9.134730538922156, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1438, "step": 6102 }, { "epoch": 9.136227544910179, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1488, "step": 6103 }, { "epoch": 9.137724550898204, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1498, "step": 6104 }, { "epoch": 9.139221556886227, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1462, "step": 6105 }, { "epoch": 9.140718562874252, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1578, "step": 6106 }, { "epoch": 9.142215568862275, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1524, "step": 6107 }, { "epoch": 9.1437125748503, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1556, "step": 6108 }, { "epoch": 9.145209580838323, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1516, "step": 6109 }, { "epoch": 9.146706586826348, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1469, "step": 6110 }, { "epoch": 9.14820359281437, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1515, "step": 6111 }, { "epoch": 9.149700598802395, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1535, "step": 6112 }, { "epoch": 9.151197604790418, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1524, "step": 6113 }, { "epoch": 9.152694610778443, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1414, "step": 6114 }, { "epoch": 9.154191616766466, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1481, "step": 6115 }, { "epoch": 9.155688622754491, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6116 }, { "epoch": 9.157185628742514, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1464, "step": 6117 }, { "epoch": 9.158682634730539, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1508, "step": 6118 }, { "epoch": 9.160179640718562, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1531, "step": 6119 }, { "epoch": 9.161676646706587, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1477, "step": 6120 }, { "epoch": 9.16317365269461, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1398, "step": 6121 }, { "epoch": 9.164670658682635, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1506, "step": 6122 }, { "epoch": 9.16616766467066, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1487, "step": 6123 }, { "epoch": 9.167664670658683, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6124 }, { "epoch": 9.169161676646707, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1426, "step": 6125 }, { "epoch": 9.17065868263473, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1558, "step": 6126 }, { "epoch": 9.172155688622755, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1531, "step": 6127 }, { "epoch": 9.173652694610778, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.1492, "step": 6128 }, { "epoch": 9.175149700598803, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1495, "step": 6129 }, { "epoch": 9.176646706586826, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1531, "step": 6130 }, { "epoch": 9.178143712574851, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6131 }, { "epoch": 9.179640718562874, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1423, "step": 6132 }, { "epoch": 9.181137724550899, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1551, "step": 6133 }, { "epoch": 9.182634730538922, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.151, "step": 6134 }, { "epoch": 9.184131736526947, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1463, "step": 6135 }, { "epoch": 9.18562874251497, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1511, "step": 6136 }, { "epoch": 9.187125748502995, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1496, "step": 6137 }, { "epoch": 9.188622754491018, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1531, "step": 6138 }, { "epoch": 9.190119760479043, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1485, "step": 6139 }, { "epoch": 9.191616766467066, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1455, "step": 6140 }, { "epoch": 9.19311377245509, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1535, "step": 6141 }, { "epoch": 9.194610778443113, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1547, "step": 6142 }, { "epoch": 9.196107784431138, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1555, "step": 6143 }, { "epoch": 9.197604790419161, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1548, "step": 6144 }, { "epoch": 9.199101796407186, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 6145 }, { "epoch": 9.20059880239521, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.142, "step": 6146 }, { "epoch": 9.202095808383234, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.151, "step": 6147 }, { "epoch": 9.203592814371257, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.145, "step": 6148 }, { "epoch": 9.205089820359282, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1504, "step": 6149 }, { "epoch": 9.206586826347305, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1462, "step": 6150 }, { "epoch": 9.20808383233533, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.146, "step": 6151 }, { "epoch": 9.209580838323353, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1485, "step": 6152 }, { "epoch": 9.211077844311378, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1422, "step": 6153 }, { "epoch": 9.2125748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1551, "step": 6154 }, { "epoch": 9.214071856287426, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1391, "step": 6155 }, { "epoch": 9.215568862275449, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1522, "step": 6156 }, { "epoch": 9.217065868263473, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1485, "step": 6157 }, { "epoch": 9.218562874251496, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1522, "step": 6158 }, { "epoch": 9.220059880239521, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1518, "step": 6159 }, { "epoch": 9.221556886227544, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1508, "step": 6160 }, { "epoch": 9.22305389221557, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1537, "step": 6161 }, { "epoch": 9.224550898203592, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6162 }, { "epoch": 9.226047904191617, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1412, "step": 6163 }, { "epoch": 9.22754491017964, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1472, "step": 6164 }, { "epoch": 9.229041916167665, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1537, "step": 6165 }, { "epoch": 9.230538922155688, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1488, "step": 6166 }, { "epoch": 9.232035928143713, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1476, "step": 6167 }, { "epoch": 9.233532934131736, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6168 }, { "epoch": 9.23502994011976, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1464, "step": 6169 }, { "epoch": 9.236526946107784, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1496, "step": 6170 }, { "epoch": 9.238023952095809, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1445, "step": 6171 }, { "epoch": 9.239520958083832, "grad_norm": 0.2099609375, "learning_rate": 0.0008, "loss": 1.1537, "step": 6172 }, { "epoch": 9.241017964071856, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1481, "step": 6173 }, { "epoch": 9.24251497005988, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 6174 }, { "epoch": 9.244011976047904, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1534, "step": 6175 }, { "epoch": 9.245508982035927, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1485, "step": 6176 }, { "epoch": 9.247005988023952, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1522, "step": 6177 }, { "epoch": 9.248502994011975, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1431, "step": 6178 }, { "epoch": 9.25, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1462, "step": 6179 }, { "epoch": 9.251497005988025, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1536, "step": 6180 }, { "epoch": 9.252994011976048, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1485, "step": 6181 }, { "epoch": 9.254491017964073, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1553, "step": 6182 }, { "epoch": 9.255988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1465, "step": 6183 }, { "epoch": 9.25748502994012, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1427, "step": 6184 }, { "epoch": 9.258982035928144, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1493, "step": 6185 }, { "epoch": 9.260479041916168, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1422, "step": 6186 }, { "epoch": 9.261976047904191, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1435, "step": 6187 }, { "epoch": 9.263473053892216, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1511, "step": 6188 }, { "epoch": 9.26497005988024, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1544, "step": 6189 }, { "epoch": 9.266467065868264, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1511, "step": 6190 }, { "epoch": 9.267964071856287, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1449, "step": 6191 }, { "epoch": 9.269461077844312, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1484, "step": 6192 }, { "epoch": 9.270958083832335, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1492, "step": 6193 }, { "epoch": 9.27245508982036, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6194 }, { "epoch": 9.273952095808383, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1513, "step": 6195 }, { "epoch": 9.275449101796408, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1488, "step": 6196 }, { "epoch": 9.27694610778443, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1411, "step": 6197 }, { "epoch": 9.278443113772456, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1511, "step": 6198 }, { "epoch": 9.279940119760479, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1477, "step": 6199 }, { "epoch": 9.281437125748504, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1549, "step": 6200 }, { "epoch": 9.282934131736527, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1515, "step": 6201 }, { "epoch": 9.284431137724551, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1481, "step": 6202 }, { "epoch": 9.285928143712574, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1454, "step": 6203 }, { "epoch": 9.2874251497006, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1426, "step": 6204 }, { "epoch": 9.288922155688622, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1602, "step": 6205 }, { "epoch": 9.290419161676647, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.149, "step": 6206 }, { "epoch": 9.29191616766467, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1453, "step": 6207 }, { "epoch": 9.293413173652695, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1485, "step": 6208 }, { "epoch": 9.294910179640718, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1493, "step": 6209 }, { "epoch": 9.296407185628743, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1463, "step": 6210 }, { "epoch": 9.297904191616766, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1479, "step": 6211 }, { "epoch": 9.29940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6212 }, { "epoch": 9.300898203592814, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6213 }, { "epoch": 9.302395209580839, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1465, "step": 6214 }, { "epoch": 9.303892215568862, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1468, "step": 6215 }, { "epoch": 9.305389221556887, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1491, "step": 6216 }, { "epoch": 9.30688622754491, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1428, "step": 6217 }, { "epoch": 9.308383233532934, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1479, "step": 6218 }, { "epoch": 9.309880239520957, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1478, "step": 6219 }, { "epoch": 9.311377245508982, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1536, "step": 6220 }, { "epoch": 9.312874251497005, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1415, "step": 6221 }, { "epoch": 9.31437125748503, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1467, "step": 6222 }, { "epoch": 9.315868263473053, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1457, "step": 6223 }, { "epoch": 9.317365269461078, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1522, "step": 6224 }, { "epoch": 9.318862275449101, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1539, "step": 6225 }, { "epoch": 9.320359281437126, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1517, "step": 6226 }, { "epoch": 9.321856287425149, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1487, "step": 6227 }, { "epoch": 9.323353293413174, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1491, "step": 6228 }, { "epoch": 9.324850299401197, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1478, "step": 6229 }, { "epoch": 9.326347305389222, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1505, "step": 6230 }, { "epoch": 9.327844311377245, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1514, "step": 6231 }, { "epoch": 9.32934131736527, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1453, "step": 6232 }, { "epoch": 9.330838323353294, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1507, "step": 6233 }, { "epoch": 9.332335329341317, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1529, "step": 6234 }, { "epoch": 9.33383233532934, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1397, "step": 6235 }, { "epoch": 9.335329341317365, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1535, "step": 6236 }, { "epoch": 9.33682634730539, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1495, "step": 6237 }, { "epoch": 9.338323353293413, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1478, "step": 6238 }, { "epoch": 9.339820359281438, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1497, "step": 6239 }, { "epoch": 9.341317365269461, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1484, "step": 6240 }, { "epoch": 9.342814371257486, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.147, "step": 6241 }, { "epoch": 9.344311377245509, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1451, "step": 6242 }, { "epoch": 9.345808383233534, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1477, "step": 6243 }, { "epoch": 9.347305389221557, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1499, "step": 6244 }, { "epoch": 9.348802395209582, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1466, "step": 6245 }, { "epoch": 9.350299401197605, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1492, "step": 6246 }, { "epoch": 9.35179640718563, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1438, "step": 6247 }, { "epoch": 9.353293413173652, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1467, "step": 6248 }, { "epoch": 9.354790419161677, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1483, "step": 6249 }, { "epoch": 9.3562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1483, "step": 6250 }, { "epoch": 9.357784431137725, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1441, "step": 6251 }, { "epoch": 9.359281437125748, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1523, "step": 6252 }, { "epoch": 9.360778443113773, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1521, "step": 6253 }, { "epoch": 9.362275449101796, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1538, "step": 6254 }, { "epoch": 9.363772455089821, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1507, "step": 6255 }, { "epoch": 9.365269461077844, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1498, "step": 6256 }, { "epoch": 9.366766467065869, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.149, "step": 6257 }, { "epoch": 9.368263473053892, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.148, "step": 6258 }, { "epoch": 9.369760479041917, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1528, "step": 6259 }, { "epoch": 9.37125748502994, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1461, "step": 6260 }, { "epoch": 9.372754491017965, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1476, "step": 6261 }, { "epoch": 9.374251497005988, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1452, "step": 6262 }, { "epoch": 9.375748502994012, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1547, "step": 6263 }, { "epoch": 9.377245508982035, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1467, "step": 6264 }, { "epoch": 9.37874251497006, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1509, "step": 6265 }, { "epoch": 9.380239520958083, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1447, "step": 6266 }, { "epoch": 9.381736526946108, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1507, "step": 6267 }, { "epoch": 9.383233532934131, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.152, "step": 6268 }, { "epoch": 9.384730538922156, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1499, "step": 6269 }, { "epoch": 9.386227544910179, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1463, "step": 6270 }, { "epoch": 9.387724550898204, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1497, "step": 6271 }, { "epoch": 9.389221556886227, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1492, "step": 6272 }, { "epoch": 9.390718562874252, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1525, "step": 6273 }, { "epoch": 9.392215568862275, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.149, "step": 6274 }, { "epoch": 9.3937125748503, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1516, "step": 6275 }, { "epoch": 9.395209580838323, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1434, "step": 6276 }, { "epoch": 9.396706586826348, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1528, "step": 6277 }, { "epoch": 9.39820359281437, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.153, "step": 6278 }, { "epoch": 9.399700598802395, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.144, "step": 6279 }, { "epoch": 9.401197604790418, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1465, "step": 6280 }, { "epoch": 9.402694610778443, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1483, "step": 6281 }, { "epoch": 9.404191616766466, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1474, "step": 6282 }, { "epoch": 9.405688622754491, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1426, "step": 6283 }, { "epoch": 9.407185628742514, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1397, "step": 6284 }, { "epoch": 9.408682634730539, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1473, "step": 6285 }, { "epoch": 9.410179640718562, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6286 }, { "epoch": 9.411676646706587, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1469, "step": 6287 }, { "epoch": 9.41317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1533, "step": 6288 }, { "epoch": 9.414670658682635, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6289 }, { "epoch": 9.41616766467066, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1432, "step": 6290 }, { "epoch": 9.417664670658683, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1464, "step": 6291 }, { "epoch": 9.419161676646706, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1503, "step": 6292 }, { "epoch": 9.42065868263473, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1536, "step": 6293 }, { "epoch": 9.422155688622755, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1446, "step": 6294 }, { "epoch": 9.423652694610778, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1566, "step": 6295 }, { "epoch": 9.425149700598803, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1492, "step": 6296 }, { "epoch": 9.426646706586826, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1499, "step": 6297 }, { "epoch": 9.428143712574851, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1528, "step": 6298 }, { "epoch": 9.429640718562874, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1472, "step": 6299 }, { "epoch": 9.431137724550899, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1465, "step": 6300 }, { "epoch": 9.432634730538922, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.143, "step": 6301 }, { "epoch": 9.434131736526947, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1432, "step": 6302 }, { "epoch": 9.43562874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1521, "step": 6303 }, { "epoch": 9.437125748502995, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1491, "step": 6304 }, { "epoch": 9.438622754491018, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1506, "step": 6305 }, { "epoch": 9.440119760479043, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1449, "step": 6306 }, { "epoch": 9.441616766467066, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1541, "step": 6307 }, { "epoch": 9.44311377245509, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1449, "step": 6308 }, { "epoch": 9.444610778443113, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.147, "step": 6309 }, { "epoch": 9.446107784431138, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.15, "step": 6310 }, { "epoch": 9.447604790419161, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1509, "step": 6311 }, { "epoch": 9.449101796407186, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1563, "step": 6312 }, { "epoch": 9.45059880239521, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1548, "step": 6313 }, { "epoch": 9.452095808383234, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1504, "step": 6314 }, { "epoch": 9.453592814371257, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1546, "step": 6315 }, { "epoch": 9.455089820359282, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1477, "step": 6316 }, { "epoch": 9.456586826347305, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1509, "step": 6317 }, { "epoch": 9.45808383233533, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6318 }, { "epoch": 9.459580838323353, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1504, "step": 6319 }, { "epoch": 9.461077844311378, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.15, "step": 6320 }, { "epoch": 9.4625748502994, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6321 }, { "epoch": 9.464071856287426, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1508, "step": 6322 }, { "epoch": 9.465568862275449, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1513, "step": 6323 }, { "epoch": 9.467065868263473, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1476, "step": 6324 }, { "epoch": 9.468562874251496, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1465, "step": 6325 }, { "epoch": 9.470059880239521, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1523, "step": 6326 }, { "epoch": 9.471556886227544, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1506, "step": 6327 }, { "epoch": 9.47305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1462, "step": 6328 }, { "epoch": 9.474550898203592, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1485, "step": 6329 }, { "epoch": 9.476047904191617, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1462, "step": 6330 }, { "epoch": 9.47754491017964, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6331 }, { "epoch": 9.479041916167665, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1451, "step": 6332 }, { "epoch": 9.480538922155688, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1432, "step": 6333 }, { "epoch": 9.482035928143713, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1495, "step": 6334 }, { "epoch": 9.483532934131736, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1561, "step": 6335 }, { "epoch": 9.48502994011976, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1462, "step": 6336 }, { "epoch": 9.486526946107784, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1494, "step": 6337 }, { "epoch": 9.488023952095809, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1448, "step": 6338 }, { "epoch": 9.489520958083832, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1425, "step": 6339 }, { "epoch": 9.491017964071856, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1476, "step": 6340 }, { "epoch": 9.49251497005988, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6341 }, { "epoch": 9.494011976047904, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1421, "step": 6342 }, { "epoch": 9.495508982035927, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1516, "step": 6343 }, { "epoch": 9.497005988023952, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1441, "step": 6344 }, { "epoch": 9.498502994011975, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1554, "step": 6345 }, { "epoch": 9.5, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1529, "step": 6346 }, { "epoch": 9.501497005988025, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1433, "step": 6347 }, { "epoch": 9.502994011976048, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1505, "step": 6348 }, { "epoch": 9.504491017964071, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1457, "step": 6349 }, { "epoch": 9.505988023952096, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1464, "step": 6350 }, { "epoch": 9.50748502994012, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1466, "step": 6351 }, { "epoch": 9.508982035928144, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1536, "step": 6352 }, { "epoch": 9.510479041916168, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.153, "step": 6353 }, { "epoch": 9.511976047904191, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1536, "step": 6354 }, { "epoch": 9.513473053892216, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1449, "step": 6355 }, { "epoch": 9.51497005988024, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1444, "step": 6356 }, { "epoch": 9.516467065868264, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1502, "step": 6357 }, { "epoch": 9.517964071856287, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.152, "step": 6358 }, { "epoch": 9.519461077844312, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1474, "step": 6359 }, { "epoch": 9.520958083832335, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1471, "step": 6360 }, { "epoch": 9.52245508982036, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6361 }, { "epoch": 9.523952095808383, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1496, "step": 6362 }, { "epoch": 9.525449101796408, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1407, "step": 6363 }, { "epoch": 9.52694610778443, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1426, "step": 6364 }, { "epoch": 9.528443113772456, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1491, "step": 6365 }, { "epoch": 9.529940119760479, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1514, "step": 6366 }, { "epoch": 9.531437125748504, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6367 }, { "epoch": 9.532934131736527, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1595, "step": 6368 }, { "epoch": 9.534431137724551, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1415, "step": 6369 }, { "epoch": 9.535928143712574, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1529, "step": 6370 }, { "epoch": 9.5374251497006, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1457, "step": 6371 }, { "epoch": 9.538922155688622, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6372 }, { "epoch": 9.540419161676647, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1424, "step": 6373 }, { "epoch": 9.54191616766467, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1481, "step": 6374 }, { "epoch": 9.543413173652695, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1487, "step": 6375 }, { "epoch": 9.544910179640718, "grad_norm": 0.19140625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6376 }, { "epoch": 9.546407185628743, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.1521, "step": 6377 }, { "epoch": 9.547904191616766, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1425, "step": 6378 }, { "epoch": 9.54940119760479, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1528, "step": 6379 }, { "epoch": 9.550898203592814, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1486, "step": 6380 }, { "epoch": 9.552395209580839, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.1534, "step": 6381 }, { "epoch": 9.553892215568862, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1472, "step": 6382 }, { "epoch": 9.555389221556887, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1469, "step": 6383 }, { "epoch": 9.55688622754491, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.141, "step": 6384 }, { "epoch": 9.558383233532934, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1543, "step": 6385 }, { "epoch": 9.559880239520957, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1479, "step": 6386 }, { "epoch": 9.561377245508982, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1446, "step": 6387 }, { "epoch": 9.562874251497005, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.151, "step": 6388 }, { "epoch": 9.56437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.149, "step": 6389 }, { "epoch": 9.565868263473053, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1507, "step": 6390 }, { "epoch": 9.567365269461078, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1458, "step": 6391 }, { "epoch": 9.568862275449101, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.139, "step": 6392 }, { "epoch": 9.570359281437126, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1511, "step": 6393 }, { "epoch": 9.571856287425149, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1516, "step": 6394 }, { "epoch": 9.573353293413174, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1533, "step": 6395 }, { "epoch": 9.574850299401197, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1472, "step": 6396 }, { "epoch": 9.576347305389222, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1505, "step": 6397 }, { "epoch": 9.577844311377245, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1492, "step": 6398 }, { "epoch": 9.57934131736527, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.144, "step": 6399 }, { "epoch": 9.580838323353294, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1504, "step": 6400 }, { "epoch": 9.582335329341317, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1561, "step": 6401 }, { "epoch": 9.58383233532934, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1478, "step": 6402 }, { "epoch": 9.585329341317365, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.149, "step": 6403 }, { "epoch": 9.58682634730539, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1508, "step": 6404 }, { "epoch": 9.588323353293413, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1493, "step": 6405 }, { "epoch": 9.589820359281438, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1529, "step": 6406 }, { "epoch": 9.591317365269461, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1547, "step": 6407 }, { "epoch": 9.592814371257486, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1502, "step": 6408 }, { "epoch": 9.594311377245509, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1451, "step": 6409 }, { "epoch": 9.595808383233534, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1503, "step": 6410 }, { "epoch": 9.597305389221557, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1466, "step": 6411 }, { "epoch": 9.598802395209582, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1479, "step": 6412 }, { "epoch": 9.600299401197605, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1509, "step": 6413 }, { "epoch": 9.60179640718563, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1454, "step": 6414 }, { "epoch": 9.603293413173652, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1494, "step": 6415 }, { "epoch": 9.604790419161677, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.151, "step": 6416 }, { "epoch": 9.6062874251497, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1486, "step": 6417 }, { "epoch": 9.607784431137725, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1536, "step": 6418 }, { "epoch": 9.609281437125748, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1529, "step": 6419 }, { "epoch": 9.610778443113773, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1573, "step": 6420 }, { "epoch": 9.612275449101796, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1429, "step": 6421 }, { "epoch": 9.613772455089821, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1494, "step": 6422 }, { "epoch": 9.615269461077844, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.148, "step": 6423 }, { "epoch": 9.616766467065869, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.147, "step": 6424 }, { "epoch": 9.618263473053892, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6425 }, { "epoch": 9.619760479041917, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1495, "step": 6426 }, { "epoch": 9.62125748502994, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1527, "step": 6427 }, { "epoch": 9.622754491017965, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.15, "step": 6428 }, { "epoch": 9.624251497005988, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6429 }, { "epoch": 9.625748502994012, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.142, "step": 6430 }, { "epoch": 9.627245508982035, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1427, "step": 6431 }, { "epoch": 9.62874251497006, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 6432 }, { "epoch": 9.630239520958083, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1528, "step": 6433 }, { "epoch": 9.631736526946108, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.15, "step": 6434 }, { "epoch": 9.633233532934131, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1508, "step": 6435 }, { "epoch": 9.634730538922156, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1469, "step": 6436 }, { "epoch": 9.636227544910179, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1424, "step": 6437 }, { "epoch": 9.637724550898204, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1496, "step": 6438 }, { "epoch": 9.639221556886227, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1514, "step": 6439 }, { "epoch": 9.640718562874252, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6440 }, { "epoch": 9.642215568862275, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1506, "step": 6441 }, { "epoch": 9.6437125748503, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1533, "step": 6442 }, { "epoch": 9.645209580838323, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1448, "step": 6443 }, { "epoch": 9.646706586826348, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1524, "step": 6444 }, { "epoch": 9.64820359281437, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1497, "step": 6445 }, { "epoch": 9.649700598802395, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1524, "step": 6446 }, { "epoch": 9.651197604790418, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1475, "step": 6447 }, { "epoch": 9.652694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1482, "step": 6448 }, { "epoch": 9.654191616766466, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1499, "step": 6449 }, { "epoch": 9.655688622754491, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1554, "step": 6450 }, { "epoch": 9.657185628742514, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1495, "step": 6451 }, { "epoch": 9.658682634730539, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1481, "step": 6452 }, { "epoch": 9.660179640718562, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1449, "step": 6453 }, { "epoch": 9.661676646706587, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.143, "step": 6454 }, { "epoch": 9.66317365269461, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1506, "step": 6455 }, { "epoch": 9.664670658682635, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1495, "step": 6456 }, { "epoch": 9.66616766467066, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1534, "step": 6457 }, { "epoch": 9.667664670658683, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 6458 }, { "epoch": 9.669161676646706, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.147, "step": 6459 }, { "epoch": 9.67065868263473, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1502, "step": 6460 }, { "epoch": 9.672155688622755, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1504, "step": 6461 }, { "epoch": 9.673652694610778, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.145, "step": 6462 }, { "epoch": 9.675149700598803, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.148, "step": 6463 }, { "epoch": 9.676646706586826, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.154, "step": 6464 }, { "epoch": 9.678143712574851, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.15, "step": 6465 }, { "epoch": 9.679640718562874, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1403, "step": 6466 }, { "epoch": 9.681137724550899, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1453, "step": 6467 }, { "epoch": 9.682634730538922, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1516, "step": 6468 }, { "epoch": 9.684131736526947, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1468, "step": 6469 }, { "epoch": 9.68562874251497, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1531, "step": 6470 }, { "epoch": 9.687125748502995, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1536, "step": 6471 }, { "epoch": 9.688622754491018, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1474, "step": 6472 }, { "epoch": 9.690119760479043, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1506, "step": 6473 }, { "epoch": 9.691616766467066, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1452, "step": 6474 }, { "epoch": 9.69311377245509, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1529, "step": 6475 }, { "epoch": 9.694610778443113, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1455, "step": 6476 }, { "epoch": 9.696107784431138, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1497, "step": 6477 }, { "epoch": 9.697604790419161, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1486, "step": 6478 }, { "epoch": 9.699101796407186, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1455, "step": 6479 }, { "epoch": 9.70059880239521, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1456, "step": 6480 }, { "epoch": 9.702095808383234, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1513, "step": 6481 }, { "epoch": 9.703592814371257, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1539, "step": 6482 }, { "epoch": 9.705089820359282, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1419, "step": 6483 }, { "epoch": 9.706586826347305, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1447, "step": 6484 }, { "epoch": 9.70808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 6485 }, { "epoch": 9.709580838323353, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1497, "step": 6486 }, { "epoch": 9.711077844311378, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1444, "step": 6487 }, { "epoch": 9.7125748502994, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.152, "step": 6488 }, { "epoch": 9.714071856287426, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1462, "step": 6489 }, { "epoch": 9.715568862275449, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1524, "step": 6490 }, { "epoch": 9.717065868263473, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1506, "step": 6491 }, { "epoch": 9.718562874251496, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1446, "step": 6492 }, { "epoch": 9.720059880239521, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.152, "step": 6493 }, { "epoch": 9.721556886227544, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1498, "step": 6494 }, { "epoch": 9.72305389221557, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1547, "step": 6495 }, { "epoch": 9.724550898203592, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1498, "step": 6496 }, { "epoch": 9.726047904191617, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.15, "step": 6497 }, { "epoch": 9.72754491017964, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1546, "step": 6498 }, { "epoch": 9.729041916167665, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1518, "step": 6499 }, { "epoch": 9.730538922155688, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6500 }, { "epoch": 9.732035928143713, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.152, "step": 6501 }, { "epoch": 9.733532934131736, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1445, "step": 6502 }, { "epoch": 9.73502994011976, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1529, "step": 6503 }, { "epoch": 9.736526946107784, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1463, "step": 6504 }, { "epoch": 9.738023952095809, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1503, "step": 6505 }, { "epoch": 9.739520958083832, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1482, "step": 6506 }, { "epoch": 9.741017964071856, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1531, "step": 6507 }, { "epoch": 9.74251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1524, "step": 6508 }, { "epoch": 9.744011976047904, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1553, "step": 6509 }, { "epoch": 9.745508982035929, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1439, "step": 6510 }, { "epoch": 9.747005988023952, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1412, "step": 6511 }, { "epoch": 9.748502994011975, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1495, "step": 6512 }, { "epoch": 9.75, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6513 }, { "epoch": 9.751497005988025, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6514 }, { "epoch": 9.752994011976048, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1455, "step": 6515 }, { "epoch": 9.754491017964071, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1451, "step": 6516 }, { "epoch": 9.755988023952096, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1501, "step": 6517 }, { "epoch": 9.75748502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1465, "step": 6518 }, { "epoch": 9.758982035928144, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6519 }, { "epoch": 9.760479041916168, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1437, "step": 6520 }, { "epoch": 9.761976047904191, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6521 }, { "epoch": 9.763473053892216, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.146, "step": 6522 }, { "epoch": 9.76497005988024, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1453, "step": 6523 }, { "epoch": 9.766467065868264, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1504, "step": 6524 }, { "epoch": 9.767964071856287, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1496, "step": 6525 }, { "epoch": 9.769461077844312, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1426, "step": 6526 }, { "epoch": 9.770958083832335, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1439, "step": 6527 }, { "epoch": 9.77245508982036, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1399, "step": 6528 }, { "epoch": 9.773952095808383, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1392, "step": 6529 }, { "epoch": 9.775449101796408, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1467, "step": 6530 }, { "epoch": 9.77694610778443, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1453, "step": 6531 }, { "epoch": 9.778443113772456, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1436, "step": 6532 }, { "epoch": 9.779940119760479, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6533 }, { "epoch": 9.781437125748504, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1434, "step": 6534 }, { "epoch": 9.782934131736527, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1525, "step": 6535 }, { "epoch": 9.784431137724551, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1519, "step": 6536 }, { "epoch": 9.785928143712574, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1483, "step": 6537 }, { "epoch": 9.7874251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1547, "step": 6538 }, { "epoch": 9.788922155688622, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6539 }, { "epoch": 9.790419161676647, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1443, "step": 6540 }, { "epoch": 9.79191616766467, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1515, "step": 6541 }, { "epoch": 9.793413173652695, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1434, "step": 6542 }, { "epoch": 9.794910179640718, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1519, "step": 6543 }, { "epoch": 9.796407185628743, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1464, "step": 6544 }, { "epoch": 9.797904191616766, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1543, "step": 6545 }, { "epoch": 9.79940119760479, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1457, "step": 6546 }, { "epoch": 9.800898203592814, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6547 }, { "epoch": 9.802395209580839, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1461, "step": 6548 }, { "epoch": 9.803892215568862, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1492, "step": 6549 }, { "epoch": 9.805389221556887, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6550 }, { "epoch": 9.80688622754491, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1441, "step": 6551 }, { "epoch": 9.808383233532934, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1475, "step": 6552 }, { "epoch": 9.809880239520957, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1448, "step": 6553 }, { "epoch": 9.811377245508982, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1492, "step": 6554 }, { "epoch": 9.812874251497005, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1432, "step": 6555 }, { "epoch": 9.81437125748503, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1392, "step": 6556 }, { "epoch": 9.815868263473053, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1544, "step": 6557 }, { "epoch": 9.817365269461078, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.145, "step": 6558 }, { "epoch": 9.818862275449101, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1411, "step": 6559 }, { "epoch": 9.820359281437126, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1509, "step": 6560 }, { "epoch": 9.821856287425149, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1569, "step": 6561 }, { "epoch": 9.823353293413174, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1481, "step": 6562 }, { "epoch": 9.824850299401197, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.149, "step": 6563 }, { "epoch": 9.826347305389222, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1517, "step": 6564 }, { "epoch": 9.827844311377245, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1402, "step": 6565 }, { "epoch": 9.82934131736527, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1462, "step": 6566 }, { "epoch": 9.830838323353294, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1443, "step": 6567 }, { "epoch": 9.832335329341317, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6568 }, { "epoch": 9.83383233532934, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1503, "step": 6569 }, { "epoch": 9.835329341317365, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1428, "step": 6570 }, { "epoch": 9.83682634730539, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1446, "step": 6571 }, { "epoch": 9.838323353293413, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1527, "step": 6572 }, { "epoch": 9.839820359281438, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1519, "step": 6573 }, { "epoch": 9.841317365269461, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1401, "step": 6574 }, { "epoch": 9.842814371257486, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.146, "step": 6575 }, { "epoch": 9.844311377245509, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1456, "step": 6576 }, { "epoch": 9.845808383233534, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1503, "step": 6577 }, { "epoch": 9.847305389221557, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1461, "step": 6578 }, { "epoch": 9.848802395209582, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.156, "step": 6579 }, { "epoch": 9.850299401197605, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6580 }, { "epoch": 9.85179640718563, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.153, "step": 6581 }, { "epoch": 9.853293413173652, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1416, "step": 6582 }, { "epoch": 9.854790419161677, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.147, "step": 6583 }, { "epoch": 9.8562874251497, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1441, "step": 6584 }, { "epoch": 9.857784431137725, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1558, "step": 6585 }, { "epoch": 9.859281437125748, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1487, "step": 6586 }, { "epoch": 9.860778443113773, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1418, "step": 6587 }, { "epoch": 9.862275449101796, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1469, "step": 6588 }, { "epoch": 9.863772455089821, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1525, "step": 6589 }, { "epoch": 9.865269461077844, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1485, "step": 6590 }, { "epoch": 9.866766467065869, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1449, "step": 6591 }, { "epoch": 9.868263473053892, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.146, "step": 6592 }, { "epoch": 9.869760479041917, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1475, "step": 6593 }, { "epoch": 9.87125748502994, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1427, "step": 6594 }, { "epoch": 9.872754491017965, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1513, "step": 6595 }, { "epoch": 9.874251497005988, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1521, "step": 6596 }, { "epoch": 9.875748502994012, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1469, "step": 6597 }, { "epoch": 9.877245508982035, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.1529, "step": 6598 }, { "epoch": 9.87874251497006, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1485, "step": 6599 }, { "epoch": 9.880239520958083, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6600 }, { "epoch": 9.881736526946108, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1455, "step": 6601 }, { "epoch": 9.883233532934131, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1468, "step": 6602 }, { "epoch": 9.884730538922156, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1513, "step": 6603 }, { "epoch": 9.886227544910179, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1471, "step": 6604 }, { "epoch": 9.887724550898204, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1467, "step": 6605 }, { "epoch": 9.889221556886227, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1416, "step": 6606 }, { "epoch": 9.890718562874252, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1488, "step": 6607 }, { "epoch": 9.892215568862275, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1528, "step": 6608 }, { "epoch": 9.8937125748503, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1397, "step": 6609 }, { "epoch": 9.895209580838323, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1397, "step": 6610 }, { "epoch": 9.896706586826348, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1445, "step": 6611 }, { "epoch": 9.89820359281437, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1435, "step": 6612 }, { "epoch": 9.899700598802395, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1458, "step": 6613 }, { "epoch": 9.901197604790418, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1478, "step": 6614 }, { "epoch": 9.902694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1496, "step": 6615 }, { "epoch": 9.904191616766466, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6616 }, { "epoch": 9.905688622754491, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1438, "step": 6617 }, { "epoch": 9.907185628742514, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1436, "step": 6618 }, { "epoch": 9.908682634730539, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1451, "step": 6619 }, { "epoch": 9.910179640718562, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1501, "step": 6620 }, { "epoch": 9.911676646706587, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.149, "step": 6621 }, { "epoch": 9.91317365269461, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1426, "step": 6622 }, { "epoch": 9.914670658682635, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1461, "step": 6623 }, { "epoch": 9.91616766467066, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1422, "step": 6624 }, { "epoch": 9.917664670658683, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1383, "step": 6625 }, { "epoch": 9.919161676646706, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1507, "step": 6626 }, { "epoch": 9.92065868263473, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.149, "step": 6627 }, { "epoch": 9.922155688622755, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1458, "step": 6628 }, { "epoch": 9.923652694610778, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6629 }, { "epoch": 9.925149700598803, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1522, "step": 6630 }, { "epoch": 9.926646706586826, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1494, "step": 6631 }, { "epoch": 9.928143712574851, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1499, "step": 6632 }, { "epoch": 9.929640718562874, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1543, "step": 6633 }, { "epoch": 9.931137724550899, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1469, "step": 6634 }, { "epoch": 9.932634730538922, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1521, "step": 6635 }, { "epoch": 9.934131736526947, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1494, "step": 6636 }, { "epoch": 9.93562874251497, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1475, "step": 6637 }, { "epoch": 9.937125748502995, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1461, "step": 6638 }, { "epoch": 9.938622754491018, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.139, "step": 6639 }, { "epoch": 9.940119760479043, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1503, "step": 6640 }, { "epoch": 9.941616766467066, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1446, "step": 6641 }, { "epoch": 9.94311377245509, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1512, "step": 6642 }, { "epoch": 9.944610778443113, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1475, "step": 6643 }, { "epoch": 9.946107784431138, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1455, "step": 6644 }, { "epoch": 9.947604790419161, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.149, "step": 6645 }, { "epoch": 9.949101796407186, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6646 }, { "epoch": 9.95059880239521, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6647 }, { "epoch": 9.952095808383234, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1515, "step": 6648 }, { "epoch": 9.953592814371257, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1437, "step": 6649 }, { "epoch": 9.955089820359282, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1392, "step": 6650 }, { "epoch": 9.956586826347305, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1523, "step": 6651 }, { "epoch": 9.95808383233533, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1506, "step": 6652 }, { "epoch": 9.959580838323353, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1437, "step": 6653 }, { "epoch": 9.961077844311378, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1458, "step": 6654 }, { "epoch": 9.9625748502994, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1505, "step": 6655 }, { "epoch": 9.964071856287426, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1478, "step": 6656 }, { "epoch": 9.965568862275449, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6657 }, { "epoch": 9.967065868263473, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1443, "step": 6658 }, { "epoch": 9.968562874251496, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1525, "step": 6659 }, { "epoch": 9.970059880239521, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1514, "step": 6660 }, { "epoch": 9.971556886227544, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1445, "step": 6661 }, { "epoch": 9.97305389221557, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1457, "step": 6662 }, { "epoch": 9.974550898203592, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1431, "step": 6663 }, { "epoch": 9.976047904191617, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1464, "step": 6664 }, { "epoch": 9.97754491017964, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1487, "step": 6665 }, { "epoch": 9.979041916167665, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1377, "step": 6666 }, { "epoch": 9.980538922155688, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1529, "step": 6667 }, { "epoch": 9.982035928143713, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6668 }, { "epoch": 9.983532934131736, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1427, "step": 6669 }, { "epoch": 9.98502994011976, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1426, "step": 6670 }, { "epoch": 9.986526946107784, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1406, "step": 6671 }, { "epoch": 9.988023952095809, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6672 }, { "epoch": 9.989520958083832, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1507, "step": 6673 }, { "epoch": 9.991017964071856, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1489, "step": 6674 }, { "epoch": 9.99251497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1475, "step": 6675 }, { "epoch": 9.994011976047904, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1514, "step": 6676 }, { "epoch": 9.995508982035929, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1473, "step": 6677 }, { "epoch": 9.997005988023952, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1484, "step": 6678 }, { "epoch": 9.998502994011975, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.145, "step": 6679 }, { "epoch": 10.0, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1508, "step": 6680 }, { "epoch": 10.001497005988025, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1464, "step": 6681 }, { "epoch": 10.002994011976048, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 6682 }, { "epoch": 10.004491017964073, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1439, "step": 6683 }, { "epoch": 10.005988023952096, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1412, "step": 6684 }, { "epoch": 10.00748502994012, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1464, "step": 6685 }, { "epoch": 10.008982035928144, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1431, "step": 6686 }, { "epoch": 10.010479041916168, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1457, "step": 6687 }, { "epoch": 10.011976047904191, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1471, "step": 6688 }, { "epoch": 10.013473053892216, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.147, "step": 6689 }, { "epoch": 10.01497005988024, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1474, "step": 6690 }, { "epoch": 10.016467065868264, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1436, "step": 6691 }, { "epoch": 10.017964071856287, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1505, "step": 6692 }, { "epoch": 10.019461077844312, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1372, "step": 6693 }, { "epoch": 10.020958083832335, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1507, "step": 6694 }, { "epoch": 10.02245508982036, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1481, "step": 6695 }, { "epoch": 10.023952095808383, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 6696 }, { "epoch": 10.025449101796408, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1417, "step": 6697 }, { "epoch": 10.02694610778443, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1416, "step": 6698 }, { "epoch": 10.028443113772456, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1417, "step": 6699 }, { "epoch": 10.029940119760479, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1425, "step": 6700 }, { "epoch": 10.031437125748504, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1527, "step": 6701 }, { "epoch": 10.032934131736527, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1449, "step": 6702 }, { "epoch": 10.034431137724551, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1462, "step": 6703 }, { "epoch": 10.035928143712574, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1518, "step": 6704 }, { "epoch": 10.0374251497006, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1432, "step": 6705 }, { "epoch": 10.038922155688622, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1461, "step": 6706 }, { "epoch": 10.040419161676647, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1462, "step": 6707 }, { "epoch": 10.04191616766467, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1503, "step": 6708 }, { "epoch": 10.043413173652695, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1481, "step": 6709 }, { "epoch": 10.044910179640718, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.141, "step": 6710 }, { "epoch": 10.046407185628743, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1438, "step": 6711 }, { "epoch": 10.047904191616766, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1575, "step": 6712 }, { "epoch": 10.04940119760479, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1482, "step": 6713 }, { "epoch": 10.050898203592814, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.142, "step": 6714 }, { "epoch": 10.052395209580839, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6715 }, { "epoch": 10.053892215568862, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1563, "step": 6716 }, { "epoch": 10.055389221556887, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1493, "step": 6717 }, { "epoch": 10.05688622754491, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1474, "step": 6718 }, { "epoch": 10.058383233532934, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1445, "step": 6719 }, { "epoch": 10.059880239520957, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1456, "step": 6720 }, { "epoch": 10.061377245508982, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1479, "step": 6721 }, { "epoch": 10.062874251497005, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1466, "step": 6722 }, { "epoch": 10.06437125748503, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1412, "step": 6723 }, { "epoch": 10.065868263473053, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1419, "step": 6724 }, { "epoch": 10.067365269461078, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1439, "step": 6725 }, { "epoch": 10.068862275449101, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1401, "step": 6726 }, { "epoch": 10.070359281437126, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1452, "step": 6727 }, { "epoch": 10.071856287425149, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1511, "step": 6728 }, { "epoch": 10.073353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1511, "step": 6729 }, { "epoch": 10.074850299401197, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.149, "step": 6730 }, { "epoch": 10.076347305389222, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6731 }, { "epoch": 10.077844311377245, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1584, "step": 6732 }, { "epoch": 10.07934131736527, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1383, "step": 6733 }, { "epoch": 10.080838323353293, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1512, "step": 6734 }, { "epoch": 10.082335329341317, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.146, "step": 6735 }, { "epoch": 10.08383233532934, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1554, "step": 6736 }, { "epoch": 10.085329341317365, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1446, "step": 6737 }, { "epoch": 10.08682634730539, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1475, "step": 6738 }, { "epoch": 10.088323353293413, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6739 }, { "epoch": 10.089820359281438, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1487, "step": 6740 }, { "epoch": 10.091317365269461, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1488, "step": 6741 }, { "epoch": 10.092814371257486, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1368, "step": 6742 }, { "epoch": 10.094311377245509, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1445, "step": 6743 }, { "epoch": 10.095808383233534, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1476, "step": 6744 }, { "epoch": 10.097305389221557, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1478, "step": 6745 }, { "epoch": 10.098802395209582, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1381, "step": 6746 }, { "epoch": 10.100299401197605, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1421, "step": 6747 }, { "epoch": 10.10179640718563, "grad_norm": 0.049072265625, "learning_rate": 0.0008, "loss": 1.1513, "step": 6748 }, { "epoch": 10.103293413173652, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1393, "step": 6749 }, { "epoch": 10.104790419161677, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1485, "step": 6750 }, { "epoch": 10.1062874251497, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1517, "step": 6751 }, { "epoch": 10.107784431137725, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1514, "step": 6752 }, { "epoch": 10.109281437125748, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1469, "step": 6753 }, { "epoch": 10.110778443113773, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1511, "step": 6754 }, { "epoch": 10.112275449101796, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1447, "step": 6755 }, { "epoch": 10.113772455089821, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1442, "step": 6756 }, { "epoch": 10.115269461077844, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1439, "step": 6757 }, { "epoch": 10.116766467065869, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1422, "step": 6758 }, { "epoch": 10.118263473053892, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1417, "step": 6759 }, { "epoch": 10.119760479041917, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1448, "step": 6760 }, { "epoch": 10.12125748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.149, "step": 6761 }, { "epoch": 10.122754491017965, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1496, "step": 6762 }, { "epoch": 10.124251497005988, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1459, "step": 6763 }, { "epoch": 10.125748502994012, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1473, "step": 6764 }, { "epoch": 10.127245508982035, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1452, "step": 6765 }, { "epoch": 10.12874251497006, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 6766 }, { "epoch": 10.130239520958083, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1408, "step": 6767 }, { "epoch": 10.131736526946108, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1473, "step": 6768 }, { "epoch": 10.133233532934131, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6769 }, { "epoch": 10.134730538922156, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1453, "step": 6770 }, { "epoch": 10.136227544910179, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1524, "step": 6771 }, { "epoch": 10.137724550898204, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.145, "step": 6772 }, { "epoch": 10.139221556886227, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1444, "step": 6773 }, { "epoch": 10.140718562874252, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1531, "step": 6774 }, { "epoch": 10.142215568862275, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1484, "step": 6775 }, { "epoch": 10.1437125748503, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1494, "step": 6776 }, { "epoch": 10.145209580838323, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1478, "step": 6777 }, { "epoch": 10.146706586826348, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1453, "step": 6778 }, { "epoch": 10.14820359281437, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1401, "step": 6779 }, { "epoch": 10.149700598802395, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1524, "step": 6780 }, { "epoch": 10.151197604790418, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1487, "step": 6781 }, { "epoch": 10.152694610778443, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1432, "step": 6782 }, { "epoch": 10.154191616766466, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.153, "step": 6783 }, { "epoch": 10.155688622754491, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1483, "step": 6784 }, { "epoch": 10.157185628742514, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1409, "step": 6785 }, { "epoch": 10.158682634730539, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1497, "step": 6786 }, { "epoch": 10.160179640718562, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1495, "step": 6787 }, { "epoch": 10.161676646706587, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.147, "step": 6788 }, { "epoch": 10.16317365269461, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1488, "step": 6789 }, { "epoch": 10.164670658682635, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.1476, "step": 6790 }, { "epoch": 10.16616766467066, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1499, "step": 6791 }, { "epoch": 10.167664670658683, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.149, "step": 6792 }, { "epoch": 10.169161676646707, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.1456, "step": 6793 }, { "epoch": 10.17065868263473, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.142, "step": 6794 }, { "epoch": 10.172155688622755, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.146, "step": 6795 }, { "epoch": 10.173652694610778, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1441, "step": 6796 }, { "epoch": 10.175149700598803, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1451, "step": 6797 }, { "epoch": 10.176646706586826, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1466, "step": 6798 }, { "epoch": 10.178143712574851, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1486, "step": 6799 }, { "epoch": 10.179640718562874, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1467, "step": 6800 }, { "epoch": 10.181137724550899, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1472, "step": 6801 }, { "epoch": 10.182634730538922, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1467, "step": 6802 }, { "epoch": 10.184131736526947, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.154, "step": 6803 }, { "epoch": 10.18562874251497, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1449, "step": 6804 }, { "epoch": 10.187125748502995, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1418, "step": 6805 }, { "epoch": 10.188622754491018, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1448, "step": 6806 }, { "epoch": 10.190119760479043, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1416, "step": 6807 }, { "epoch": 10.191616766467066, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1439, "step": 6808 }, { "epoch": 10.19311377245509, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1534, "step": 6809 }, { "epoch": 10.194610778443113, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1484, "step": 6810 }, { "epoch": 10.196107784431138, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1499, "step": 6811 }, { "epoch": 10.197604790419161, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1454, "step": 6812 }, { "epoch": 10.199101796407186, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1462, "step": 6813 }, { "epoch": 10.20059880239521, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1482, "step": 6814 }, { "epoch": 10.202095808383234, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1458, "step": 6815 }, { "epoch": 10.203592814371257, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.151, "step": 6816 }, { "epoch": 10.205089820359282, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1435, "step": 6817 }, { "epoch": 10.206586826347305, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1454, "step": 6818 }, { "epoch": 10.20808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.147, "step": 6819 }, { "epoch": 10.209580838323353, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1428, "step": 6820 }, { "epoch": 10.211077844311378, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1447, "step": 6821 }, { "epoch": 10.2125748502994, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1473, "step": 6822 }, { "epoch": 10.214071856287426, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1444, "step": 6823 }, { "epoch": 10.215568862275449, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6824 }, { "epoch": 10.217065868263473, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1459, "step": 6825 }, { "epoch": 10.218562874251496, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1499, "step": 6826 }, { "epoch": 10.220059880239521, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.151, "step": 6827 }, { "epoch": 10.221556886227544, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1467, "step": 6828 }, { "epoch": 10.22305389221557, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1528, "step": 6829 }, { "epoch": 10.224550898203592, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1364, "step": 6830 }, { "epoch": 10.226047904191617, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1446, "step": 6831 }, { "epoch": 10.22754491017964, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1431, "step": 6832 }, { "epoch": 10.229041916167665, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1454, "step": 6833 }, { "epoch": 10.230538922155688, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.145, "step": 6834 }, { "epoch": 10.232035928143713, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1486, "step": 6835 }, { "epoch": 10.233532934131736, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1437, "step": 6836 }, { "epoch": 10.23502994011976, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1452, "step": 6837 }, { "epoch": 10.236526946107784, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1448, "step": 6838 }, { "epoch": 10.238023952095809, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1494, "step": 6839 }, { "epoch": 10.239520958083832, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1535, "step": 6840 }, { "epoch": 10.241017964071856, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1449, "step": 6841 }, { "epoch": 10.24251497005988, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1437, "step": 6842 }, { "epoch": 10.244011976047904, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1467, "step": 6843 }, { "epoch": 10.245508982035927, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1483, "step": 6844 }, { "epoch": 10.247005988023952, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1405, "step": 6845 }, { "epoch": 10.248502994011975, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6846 }, { "epoch": 10.25, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6847 }, { "epoch": 10.251497005988025, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1412, "step": 6848 }, { "epoch": 10.252994011976048, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.15, "step": 6849 }, { "epoch": 10.254491017964073, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1385, "step": 6850 }, { "epoch": 10.255988023952096, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1424, "step": 6851 }, { "epoch": 10.25748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.152, "step": 6852 }, { "epoch": 10.258982035928144, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6853 }, { "epoch": 10.260479041916168, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1439, "step": 6854 }, { "epoch": 10.261976047904191, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1445, "step": 6855 }, { "epoch": 10.263473053892216, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1438, "step": 6856 }, { "epoch": 10.26497005988024, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.142, "step": 6857 }, { "epoch": 10.266467065868264, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1515, "step": 6858 }, { "epoch": 10.267964071856287, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1478, "step": 6859 }, { "epoch": 10.269461077844312, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1546, "step": 6860 }, { "epoch": 10.270958083832335, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.145, "step": 6861 }, { "epoch": 10.27245508982036, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1438, "step": 6862 }, { "epoch": 10.273952095808383, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1389, "step": 6863 }, { "epoch": 10.275449101796408, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1472, "step": 6864 }, { "epoch": 10.27694610778443, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6865 }, { "epoch": 10.278443113772456, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.152, "step": 6866 }, { "epoch": 10.279940119760479, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1431, "step": 6867 }, { "epoch": 10.281437125748504, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1454, "step": 6868 }, { "epoch": 10.282934131736527, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1401, "step": 6869 }, { "epoch": 10.284431137724551, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6870 }, { "epoch": 10.285928143712574, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1513, "step": 6871 }, { "epoch": 10.2874251497006, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.145, "step": 6872 }, { "epoch": 10.288922155688622, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.14, "step": 6873 }, { "epoch": 10.290419161676647, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 6874 }, { "epoch": 10.29191616766467, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1505, "step": 6875 }, { "epoch": 10.293413173652695, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1518, "step": 6876 }, { "epoch": 10.294910179640718, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1509, "step": 6877 }, { "epoch": 10.296407185628743, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.143, "step": 6878 }, { "epoch": 10.297904191616766, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1418, "step": 6879 }, { "epoch": 10.29940119760479, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1519, "step": 6880 }, { "epoch": 10.300898203592814, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1508, "step": 6881 }, { "epoch": 10.302395209580839, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1531, "step": 6882 }, { "epoch": 10.303892215568862, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1499, "step": 6883 }, { "epoch": 10.305389221556887, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1417, "step": 6884 }, { "epoch": 10.30688622754491, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1386, "step": 6885 }, { "epoch": 10.308383233532934, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1396, "step": 6886 }, { "epoch": 10.309880239520957, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1515, "step": 6887 }, { "epoch": 10.311377245508982, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1448, "step": 6888 }, { "epoch": 10.312874251497005, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1396, "step": 6889 }, { "epoch": 10.31437125748503, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1448, "step": 6890 }, { "epoch": 10.315868263473053, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1459, "step": 6891 }, { "epoch": 10.317365269461078, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.145, "step": 6892 }, { "epoch": 10.318862275449101, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1465, "step": 6893 }, { "epoch": 10.320359281437126, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1536, "step": 6894 }, { "epoch": 10.321856287425149, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1463, "step": 6895 }, { "epoch": 10.323353293413174, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1442, "step": 6896 }, { "epoch": 10.324850299401197, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1423, "step": 6897 }, { "epoch": 10.326347305389222, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6898 }, { "epoch": 10.327844311377245, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6899 }, { "epoch": 10.32934131736527, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1488, "step": 6900 }, { "epoch": 10.330838323353294, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1513, "step": 6901 }, { "epoch": 10.332335329341317, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1456, "step": 6902 }, { "epoch": 10.33383233532934, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.152, "step": 6903 }, { "epoch": 10.335329341317365, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.142, "step": 6904 }, { "epoch": 10.33682634730539, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1491, "step": 6905 }, { "epoch": 10.338323353293413, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1463, "step": 6906 }, { "epoch": 10.339820359281438, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1457, "step": 6907 }, { "epoch": 10.341317365269461, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 6908 }, { "epoch": 10.342814371257486, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1475, "step": 6909 }, { "epoch": 10.344311377245509, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1545, "step": 6910 }, { "epoch": 10.345808383233534, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1414, "step": 6911 }, { "epoch": 10.347305389221557, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1449, "step": 6912 }, { "epoch": 10.348802395209582, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1457, "step": 6913 }, { "epoch": 10.350299401197605, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1425, "step": 6914 }, { "epoch": 10.35179640718563, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1493, "step": 6915 }, { "epoch": 10.353293413173652, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1494, "step": 6916 }, { "epoch": 10.354790419161677, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1454, "step": 6917 }, { "epoch": 10.3562874251497, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1396, "step": 6918 }, { "epoch": 10.357784431137725, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6919 }, { "epoch": 10.359281437125748, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1505, "step": 6920 }, { "epoch": 10.360778443113773, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1493, "step": 6921 }, { "epoch": 10.362275449101796, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1391, "step": 6922 }, { "epoch": 10.363772455089821, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6923 }, { "epoch": 10.365269461077844, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1419, "step": 6924 }, { "epoch": 10.366766467065869, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1419, "step": 6925 }, { "epoch": 10.368263473053892, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.14, "step": 6926 }, { "epoch": 10.369760479041917, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1411, "step": 6927 }, { "epoch": 10.37125748502994, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.141, "step": 6928 }, { "epoch": 10.372754491017965, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1442, "step": 6929 }, { "epoch": 10.374251497005988, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1389, "step": 6930 }, { "epoch": 10.375748502994012, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1356, "step": 6931 }, { "epoch": 10.377245508982035, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1518, "step": 6932 }, { "epoch": 10.37874251497006, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.151, "step": 6933 }, { "epoch": 10.380239520958083, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1443, "step": 6934 }, { "epoch": 10.381736526946108, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1515, "step": 6935 }, { "epoch": 10.383233532934131, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1513, "step": 6936 }, { "epoch": 10.384730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.147, "step": 6937 }, { "epoch": 10.386227544910179, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1477, "step": 6938 }, { "epoch": 10.387724550898204, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.151, "step": 6939 }, { "epoch": 10.389221556886227, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1454, "step": 6940 }, { "epoch": 10.390718562874252, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1413, "step": 6941 }, { "epoch": 10.392215568862275, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1432, "step": 6942 }, { "epoch": 10.3937125748503, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6943 }, { "epoch": 10.395209580838323, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1551, "step": 6944 }, { "epoch": 10.396706586826348, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1546, "step": 6945 }, { "epoch": 10.39820359281437, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1487, "step": 6946 }, { "epoch": 10.399700598802395, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1475, "step": 6947 }, { "epoch": 10.401197604790418, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1408, "step": 6948 }, { "epoch": 10.402694610778443, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1484, "step": 6949 }, { "epoch": 10.404191616766466, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1473, "step": 6950 }, { "epoch": 10.405688622754491, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1428, "step": 6951 }, { "epoch": 10.407185628742514, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1508, "step": 6952 }, { "epoch": 10.408682634730539, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1507, "step": 6953 }, { "epoch": 10.410179640718562, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.155, "step": 6954 }, { "epoch": 10.411676646706587, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6955 }, { "epoch": 10.41317365269461, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1424, "step": 6956 }, { "epoch": 10.414670658682635, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.139, "step": 6957 }, { "epoch": 10.41616766467066, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1452, "step": 6958 }, { "epoch": 10.417664670658683, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1435, "step": 6959 }, { "epoch": 10.419161676646706, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1442, "step": 6960 }, { "epoch": 10.42065868263473, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.151, "step": 6961 }, { "epoch": 10.422155688622755, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1423, "step": 6962 }, { "epoch": 10.423652694610778, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1451, "step": 6963 }, { "epoch": 10.425149700598803, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1393, "step": 6964 }, { "epoch": 10.426646706586826, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1423, "step": 6965 }, { "epoch": 10.428143712574851, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1515, "step": 6966 }, { "epoch": 10.429640718562874, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6967 }, { "epoch": 10.431137724550899, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1477, "step": 6968 }, { "epoch": 10.432634730538922, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1451, "step": 6969 }, { "epoch": 10.434131736526947, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1445, "step": 6970 }, { "epoch": 10.43562874251497, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1387, "step": 6971 }, { "epoch": 10.437125748502995, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1439, "step": 6972 }, { "epoch": 10.438622754491018, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1436, "step": 6973 }, { "epoch": 10.440119760479043, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1455, "step": 6974 }, { "epoch": 10.441616766467066, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1526, "step": 6975 }, { "epoch": 10.44311377245509, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1461, "step": 6976 }, { "epoch": 10.444610778443113, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1466, "step": 6977 }, { "epoch": 10.446107784431138, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1391, "step": 6978 }, { "epoch": 10.447604790419161, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1432, "step": 6979 }, { "epoch": 10.449101796407186, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1432, "step": 6980 }, { "epoch": 10.45059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.145, "step": 6981 }, { "epoch": 10.452095808383234, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1453, "step": 6982 }, { "epoch": 10.453592814371257, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1494, "step": 6983 }, { "epoch": 10.455089820359282, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1485, "step": 6984 }, { "epoch": 10.456586826347305, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1433, "step": 6985 }, { "epoch": 10.45808383233533, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.138, "step": 6986 }, { "epoch": 10.459580838323353, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1438, "step": 6987 }, { "epoch": 10.461077844311378, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1428, "step": 6988 }, { "epoch": 10.4625748502994, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1473, "step": 6989 }, { "epoch": 10.464071856287426, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1501, "step": 6990 }, { "epoch": 10.465568862275449, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.1496, "step": 6991 }, { "epoch": 10.467065868263473, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1479, "step": 6992 }, { "epoch": 10.468562874251496, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1461, "step": 6993 }, { "epoch": 10.470059880239521, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1388, "step": 6994 }, { "epoch": 10.471556886227544, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1572, "step": 6995 }, { "epoch": 10.47305389221557, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1441, "step": 6996 }, { "epoch": 10.474550898203592, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1408, "step": 6997 }, { "epoch": 10.476047904191617, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1436, "step": 6998 }, { "epoch": 10.47754491017964, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.138, "step": 6999 }, { "epoch": 10.479041916167665, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 7000 }, { "epoch": 10.480538922155688, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1498, "step": 7001 }, { "epoch": 10.482035928143713, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.14, "step": 7002 }, { "epoch": 10.483532934131736, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1432, "step": 7003 }, { "epoch": 10.48502994011976, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1471, "step": 7004 }, { "epoch": 10.486526946107784, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1468, "step": 7005 }, { "epoch": 10.488023952095809, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1512, "step": 7006 }, { "epoch": 10.489520958083832, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1449, "step": 7007 }, { "epoch": 10.491017964071856, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1484, "step": 7008 }, { "epoch": 10.49251497005988, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.149, "step": 7009 }, { "epoch": 10.494011976047904, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.14, "step": 7010 }, { "epoch": 10.495508982035927, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1456, "step": 7011 }, { "epoch": 10.497005988023952, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7012 }, { "epoch": 10.498502994011975, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 7013 }, { "epoch": 10.5, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1442, "step": 7014 }, { "epoch": 10.501497005988025, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.139, "step": 7015 }, { "epoch": 10.502994011976048, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.148, "step": 7016 }, { "epoch": 10.504491017964071, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1432, "step": 7017 }, { "epoch": 10.505988023952096, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1409, "step": 7018 }, { "epoch": 10.50748502994012, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1478, "step": 7019 }, { "epoch": 10.508982035928144, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1442, "step": 7020 }, { "epoch": 10.510479041916168, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1444, "step": 7021 }, { "epoch": 10.511976047904191, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1504, "step": 7022 }, { "epoch": 10.513473053892216, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1451, "step": 7023 }, { "epoch": 10.51497005988024, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1506, "step": 7024 }, { "epoch": 10.516467065868264, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1518, "step": 7025 }, { "epoch": 10.517964071856287, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1452, "step": 7026 }, { "epoch": 10.519461077844312, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1413, "step": 7027 }, { "epoch": 10.520958083832335, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1526, "step": 7028 }, { "epoch": 10.52245508982036, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1438, "step": 7029 }, { "epoch": 10.523952095808383, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1495, "step": 7030 }, { "epoch": 10.525449101796408, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1452, "step": 7031 }, { "epoch": 10.52694610778443, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1477, "step": 7032 }, { "epoch": 10.528443113772456, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1506, "step": 7033 }, { "epoch": 10.529940119760479, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1457, "step": 7034 }, { "epoch": 10.531437125748504, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.153, "step": 7035 }, { "epoch": 10.532934131736527, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7036 }, { "epoch": 10.534431137724551, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1438, "step": 7037 }, { "epoch": 10.535928143712574, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1429, "step": 7038 }, { "epoch": 10.5374251497006, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1401, "step": 7039 }, { "epoch": 10.538922155688622, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1499, "step": 7040 }, { "epoch": 10.540419161676647, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1463, "step": 7041 }, { "epoch": 10.54191616766467, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1445, "step": 7042 }, { "epoch": 10.543413173652695, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1499, "step": 7043 }, { "epoch": 10.544910179640718, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1432, "step": 7044 }, { "epoch": 10.546407185628743, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1441, "step": 7045 }, { "epoch": 10.547904191616766, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7046 }, { "epoch": 10.54940119760479, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.1376, "step": 7047 }, { "epoch": 10.550898203592814, "grad_norm": 0.1689453125, "learning_rate": 0.0008, "loss": 1.1445, "step": 7048 }, { "epoch": 10.552395209580839, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1392, "step": 7049 }, { "epoch": 10.553892215568862, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.141, "step": 7050 }, { "epoch": 10.555389221556887, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1396, "step": 7051 }, { "epoch": 10.55688622754491, "grad_norm": 0.26953125, "learning_rate": 0.0008, "loss": 1.1431, "step": 7052 }, { "epoch": 10.558383233532934, "grad_norm": 0.28125, "learning_rate": 0.0008, "loss": 1.1436, "step": 7053 }, { "epoch": 10.559880239520957, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1453, "step": 7054 }, { "epoch": 10.561377245508982, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1498, "step": 7055 }, { "epoch": 10.562874251497005, "grad_norm": 0.2109375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7056 }, { "epoch": 10.56437125748503, "grad_norm": 0.240234375, "learning_rate": 0.0008, "loss": 1.1476, "step": 7057 }, { "epoch": 10.565868263473053, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.1449, "step": 7058 }, { "epoch": 10.567365269461078, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1484, "step": 7059 }, { "epoch": 10.568862275449101, "grad_norm": 0.2490234375, "learning_rate": 0.0008, "loss": 1.15, "step": 7060 }, { "epoch": 10.570359281437126, "grad_norm": 0.22265625, "learning_rate": 0.0008, "loss": 1.1556, "step": 7061 }, { "epoch": 10.571856287425149, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1548, "step": 7062 }, { "epoch": 10.573353293413174, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1445, "step": 7063 }, { "epoch": 10.574850299401197, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1503, "step": 7064 }, { "epoch": 10.576347305389222, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1491, "step": 7065 }, { "epoch": 10.577844311377245, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1516, "step": 7066 }, { "epoch": 10.57934131736527, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.1539, "step": 7067 }, { "epoch": 10.580838323353294, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1531, "step": 7068 }, { "epoch": 10.582335329341317, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1469, "step": 7069 }, { "epoch": 10.58383233532934, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.1496, "step": 7070 }, { "epoch": 10.585329341317365, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1434, "step": 7071 }, { "epoch": 10.58682634730539, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1503, "step": 7072 }, { "epoch": 10.588323353293413, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7073 }, { "epoch": 10.589820359281438, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1445, "step": 7074 }, { "epoch": 10.591317365269461, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1451, "step": 7075 }, { "epoch": 10.592814371257486, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.15, "step": 7076 }, { "epoch": 10.594311377245509, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1413, "step": 7077 }, { "epoch": 10.595808383233534, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.147, "step": 7078 }, { "epoch": 10.597305389221557, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1495, "step": 7079 }, { "epoch": 10.598802395209582, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1446, "step": 7080 }, { "epoch": 10.600299401197605, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1443, "step": 7081 }, { "epoch": 10.60179640718563, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.14, "step": 7082 }, { "epoch": 10.603293413173652, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.141, "step": 7083 }, { "epoch": 10.604790419161677, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1508, "step": 7084 }, { "epoch": 10.6062874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1384, "step": 7085 }, { "epoch": 10.607784431137725, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1486, "step": 7086 }, { "epoch": 10.609281437125748, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1471, "step": 7087 }, { "epoch": 10.610778443113773, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1474, "step": 7088 }, { "epoch": 10.612275449101796, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 7089 }, { "epoch": 10.613772455089821, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1437, "step": 7090 }, { "epoch": 10.615269461077844, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1482, "step": 7091 }, { "epoch": 10.616766467065869, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1536, "step": 7092 }, { "epoch": 10.618263473053892, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1414, "step": 7093 }, { "epoch": 10.619760479041917, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1436, "step": 7094 }, { "epoch": 10.62125748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.143, "step": 7095 }, { "epoch": 10.622754491017965, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1537, "step": 7096 }, { "epoch": 10.624251497005988, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1526, "step": 7097 }, { "epoch": 10.625748502994012, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1506, "step": 7098 }, { "epoch": 10.627245508982035, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1474, "step": 7099 }, { "epoch": 10.62874251497006, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1501, "step": 7100 }, { "epoch": 10.630239520958083, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1412, "step": 7101 }, { "epoch": 10.631736526946108, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7102 }, { "epoch": 10.633233532934131, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1466, "step": 7103 }, { "epoch": 10.634730538922156, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1468, "step": 7104 }, { "epoch": 10.636227544910179, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1445, "step": 7105 }, { "epoch": 10.637724550898204, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1403, "step": 7106 }, { "epoch": 10.639221556886227, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1471, "step": 7107 }, { "epoch": 10.640718562874252, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1463, "step": 7108 }, { "epoch": 10.642215568862275, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1492, "step": 7109 }, { "epoch": 10.6437125748503, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1489, "step": 7110 }, { "epoch": 10.645209580838323, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1396, "step": 7111 }, { "epoch": 10.646706586826348, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.143, "step": 7112 }, { "epoch": 10.64820359281437, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1478, "step": 7113 }, { "epoch": 10.649700598802395, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1451, "step": 7114 }, { "epoch": 10.651197604790418, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1484, "step": 7115 }, { "epoch": 10.652694610778443, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1462, "step": 7116 }, { "epoch": 10.654191616766466, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1533, "step": 7117 }, { "epoch": 10.655688622754491, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1477, "step": 7118 }, { "epoch": 10.657185628742514, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1478, "step": 7119 }, { "epoch": 10.658682634730539, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1491, "step": 7120 }, { "epoch": 10.660179640718562, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1427, "step": 7121 }, { "epoch": 10.661676646706587, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1535, "step": 7122 }, { "epoch": 10.66317365269461, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.148, "step": 7123 }, { "epoch": 10.664670658682635, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1388, "step": 7124 }, { "epoch": 10.66616766467066, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1454, "step": 7125 }, { "epoch": 10.667664670658683, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1458, "step": 7126 }, { "epoch": 10.669161676646706, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1493, "step": 7127 }, { "epoch": 10.67065868263473, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1451, "step": 7128 }, { "epoch": 10.672155688622755, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1357, "step": 7129 }, { "epoch": 10.673652694610778, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1449, "step": 7130 }, { "epoch": 10.675149700598803, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1415, "step": 7131 }, { "epoch": 10.676646706586826, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1428, "step": 7132 }, { "epoch": 10.678143712574851, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.142, "step": 7133 }, { "epoch": 10.679640718562874, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1471, "step": 7134 }, { "epoch": 10.681137724550899, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1405, "step": 7135 }, { "epoch": 10.682634730538922, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1475, "step": 7136 }, { "epoch": 10.684131736526947, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1434, "step": 7137 }, { "epoch": 10.68562874251497, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7138 }, { "epoch": 10.687125748502995, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1503, "step": 7139 }, { "epoch": 10.688622754491018, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1455, "step": 7140 }, { "epoch": 10.690119760479043, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1437, "step": 7141 }, { "epoch": 10.691616766467066, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1426, "step": 7142 }, { "epoch": 10.69311377245509, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.149, "step": 7143 }, { "epoch": 10.694610778443113, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1402, "step": 7144 }, { "epoch": 10.696107784431138, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1478, "step": 7145 }, { "epoch": 10.697604790419161, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1449, "step": 7146 }, { "epoch": 10.699101796407186, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1387, "step": 7147 }, { "epoch": 10.70059880239521, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1379, "step": 7148 }, { "epoch": 10.702095808383234, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1451, "step": 7149 }, { "epoch": 10.703592814371257, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1433, "step": 7150 }, { "epoch": 10.705089820359282, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1406, "step": 7151 }, { "epoch": 10.706586826347305, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.145, "step": 7152 }, { "epoch": 10.70808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1463, "step": 7153 }, { "epoch": 10.709580838323353, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1442, "step": 7154 }, { "epoch": 10.711077844311378, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1458, "step": 7155 }, { "epoch": 10.7125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1431, "step": 7156 }, { "epoch": 10.714071856287426, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1446, "step": 7157 }, { "epoch": 10.715568862275449, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7158 }, { "epoch": 10.717065868263473, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1498, "step": 7159 }, { "epoch": 10.718562874251496, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1448, "step": 7160 }, { "epoch": 10.720059880239521, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1415, "step": 7161 }, { "epoch": 10.721556886227544, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1472, "step": 7162 }, { "epoch": 10.72305389221557, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1448, "step": 7163 }, { "epoch": 10.724550898203592, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1464, "step": 7164 }, { "epoch": 10.726047904191617, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1449, "step": 7165 }, { "epoch": 10.72754491017964, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1476, "step": 7166 }, { "epoch": 10.729041916167665, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1434, "step": 7167 }, { "epoch": 10.730538922155688, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1437, "step": 7168 }, { "epoch": 10.732035928143713, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1471, "step": 7169 }, { "epoch": 10.733532934131736, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1467, "step": 7170 }, { "epoch": 10.73502994011976, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1498, "step": 7171 }, { "epoch": 10.736526946107784, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1464, "step": 7172 }, { "epoch": 10.738023952095809, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.147, "step": 7173 }, { "epoch": 10.739520958083832, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1513, "step": 7174 }, { "epoch": 10.741017964071856, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1428, "step": 7175 }, { "epoch": 10.74251497005988, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1466, "step": 7176 }, { "epoch": 10.744011976047904, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1447, "step": 7177 }, { "epoch": 10.745508982035929, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1501, "step": 7178 }, { "epoch": 10.747005988023952, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.146, "step": 7179 }, { "epoch": 10.748502994011975, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.145, "step": 7180 }, { "epoch": 10.75, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.142, "step": 7181 }, { "epoch": 10.751497005988025, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.1461, "step": 7182 }, { "epoch": 10.752994011976048, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1452, "step": 7183 }, { "epoch": 10.754491017964071, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1468, "step": 7184 }, { "epoch": 10.755988023952096, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1471, "step": 7185 }, { "epoch": 10.75748502994012, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1472, "step": 7186 }, { "epoch": 10.758982035928144, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1452, "step": 7187 }, { "epoch": 10.760479041916168, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1444, "step": 7188 }, { "epoch": 10.761976047904191, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1504, "step": 7189 }, { "epoch": 10.763473053892216, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1413, "step": 7190 }, { "epoch": 10.76497005988024, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.146, "step": 7191 }, { "epoch": 10.766467065868264, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1426, "step": 7192 }, { "epoch": 10.767964071856287, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.151, "step": 7193 }, { "epoch": 10.769461077844312, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1447, "step": 7194 }, { "epoch": 10.770958083832335, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1495, "step": 7195 }, { "epoch": 10.77245508982036, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1507, "step": 7196 }, { "epoch": 10.773952095808383, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1462, "step": 7197 }, { "epoch": 10.775449101796408, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1417, "step": 7198 }, { "epoch": 10.77694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1485, "step": 7199 }, { "epoch": 10.778443113772456, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.143, "step": 7200 }, { "epoch": 10.779940119760479, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7201 }, { "epoch": 10.781437125748504, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1478, "step": 7202 }, { "epoch": 10.782934131736527, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1457, "step": 7203 }, { "epoch": 10.784431137724551, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1466, "step": 7204 }, { "epoch": 10.785928143712574, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 7205 }, { "epoch": 10.7874251497006, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1473, "step": 7206 }, { "epoch": 10.788922155688622, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1401, "step": 7207 }, { "epoch": 10.790419161676647, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1475, "step": 7208 }, { "epoch": 10.79191616766467, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1512, "step": 7209 }, { "epoch": 10.793413173652695, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7210 }, { "epoch": 10.794910179640718, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1426, "step": 7211 }, { "epoch": 10.796407185628743, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1373, "step": 7212 }, { "epoch": 10.797904191616766, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1454, "step": 7213 }, { "epoch": 10.79940119760479, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1442, "step": 7214 }, { "epoch": 10.800898203592814, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1408, "step": 7215 }, { "epoch": 10.802395209580839, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1462, "step": 7216 }, { "epoch": 10.803892215568862, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1468, "step": 7217 }, { "epoch": 10.805389221556887, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.144, "step": 7218 }, { "epoch": 10.80688622754491, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1487, "step": 7219 }, { "epoch": 10.808383233532934, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1416, "step": 7220 }, { "epoch": 10.809880239520957, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1472, "step": 7221 }, { "epoch": 10.811377245508982, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1464, "step": 7222 }, { "epoch": 10.812874251497005, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7223 }, { "epoch": 10.81437125748503, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1536, "step": 7224 }, { "epoch": 10.815868263473053, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1377, "step": 7225 }, { "epoch": 10.817365269461078, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1475, "step": 7226 }, { "epoch": 10.818862275449101, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1486, "step": 7227 }, { "epoch": 10.820359281437126, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1464, "step": 7228 }, { "epoch": 10.821856287425149, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1416, "step": 7229 }, { "epoch": 10.823353293413174, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1468, "step": 7230 }, { "epoch": 10.824850299401197, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1483, "step": 7231 }, { "epoch": 10.826347305389222, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1466, "step": 7232 }, { "epoch": 10.827844311377245, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1457, "step": 7233 }, { "epoch": 10.82934131736527, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1467, "step": 7234 }, { "epoch": 10.830838323353294, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1442, "step": 7235 }, { "epoch": 10.832335329341317, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1488, "step": 7236 }, { "epoch": 10.83383233532934, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 7237 }, { "epoch": 10.835329341317365, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 7238 }, { "epoch": 10.83682634730539, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1471, "step": 7239 }, { "epoch": 10.838323353293413, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1518, "step": 7240 }, { "epoch": 10.839820359281438, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1419, "step": 7241 }, { "epoch": 10.841317365269461, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.152, "step": 7242 }, { "epoch": 10.842814371257486, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1436, "step": 7243 }, { "epoch": 10.844311377245509, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1462, "step": 7244 }, { "epoch": 10.845808383233534, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1421, "step": 7245 }, { "epoch": 10.847305389221557, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1456, "step": 7246 }, { "epoch": 10.848802395209582, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1403, "step": 7247 }, { "epoch": 10.850299401197605, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1393, "step": 7248 }, { "epoch": 10.85179640718563, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1455, "step": 7249 }, { "epoch": 10.853293413173652, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7250 }, { "epoch": 10.854790419161677, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1445, "step": 7251 }, { "epoch": 10.8562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1451, "step": 7252 }, { "epoch": 10.857784431137725, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1415, "step": 7253 }, { "epoch": 10.859281437125748, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1396, "step": 7254 }, { "epoch": 10.860778443113773, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1447, "step": 7255 }, { "epoch": 10.862275449101796, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.138, "step": 7256 }, { "epoch": 10.863772455089821, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1492, "step": 7257 }, { "epoch": 10.865269461077844, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1456, "step": 7258 }, { "epoch": 10.866766467065869, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1336, "step": 7259 }, { "epoch": 10.868263473053892, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1498, "step": 7260 }, { "epoch": 10.869760479041917, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1424, "step": 7261 }, { "epoch": 10.87125748502994, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1471, "step": 7262 }, { "epoch": 10.872754491017965, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1517, "step": 7263 }, { "epoch": 10.874251497005988, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1442, "step": 7264 }, { "epoch": 10.875748502994012, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1453, "step": 7265 }, { "epoch": 10.877245508982035, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1462, "step": 7266 }, { "epoch": 10.87874251497006, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1414, "step": 7267 }, { "epoch": 10.880239520958083, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1448, "step": 7268 }, { "epoch": 10.881736526946108, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 7269 }, { "epoch": 10.883233532934131, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.148, "step": 7270 }, { "epoch": 10.884730538922156, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1405, "step": 7271 }, { "epoch": 10.886227544910179, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1424, "step": 7272 }, { "epoch": 10.887724550898204, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1484, "step": 7273 }, { "epoch": 10.889221556886227, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1451, "step": 7274 }, { "epoch": 10.890718562874252, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1469, "step": 7275 }, { "epoch": 10.892215568862275, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1415, "step": 7276 }, { "epoch": 10.8937125748503, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1467, "step": 7277 }, { "epoch": 10.895209580838323, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1455, "step": 7278 }, { "epoch": 10.896706586826348, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1466, "step": 7279 }, { "epoch": 10.89820359281437, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1509, "step": 7280 }, { "epoch": 10.899700598802395, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1453, "step": 7281 }, { "epoch": 10.901197604790418, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1495, "step": 7282 }, { "epoch": 10.902694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1442, "step": 7283 }, { "epoch": 10.904191616766466, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7284 }, { "epoch": 10.905688622754491, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1469, "step": 7285 }, { "epoch": 10.907185628742514, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.144, "step": 7286 }, { "epoch": 10.908682634730539, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1445, "step": 7287 }, { "epoch": 10.910179640718562, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1441, "step": 7288 }, { "epoch": 10.911676646706587, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1513, "step": 7289 }, { "epoch": 10.91317365269461, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1395, "step": 7290 }, { "epoch": 10.914670658682635, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1465, "step": 7291 }, { "epoch": 10.91616766467066, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.145, "step": 7292 }, { "epoch": 10.917664670658683, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1415, "step": 7293 }, { "epoch": 10.919161676646706, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 7294 }, { "epoch": 10.92065868263473, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1402, "step": 7295 }, { "epoch": 10.922155688622755, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1419, "step": 7296 }, { "epoch": 10.923652694610778, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.15, "step": 7297 }, { "epoch": 10.925149700598803, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 7298 }, { "epoch": 10.926646706586826, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1424, "step": 7299 }, { "epoch": 10.928143712574851, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1536, "step": 7300 }, { "epoch": 10.929640718562874, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1485, "step": 7301 }, { "epoch": 10.931137724550899, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1436, "step": 7302 }, { "epoch": 10.932634730538922, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1474, "step": 7303 }, { "epoch": 10.934131736526947, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1453, "step": 7304 }, { "epoch": 10.93562874251497, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1461, "step": 7305 }, { "epoch": 10.937125748502995, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1429, "step": 7306 }, { "epoch": 10.938622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1468, "step": 7307 }, { "epoch": 10.940119760479043, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1454, "step": 7308 }, { "epoch": 10.941616766467066, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7309 }, { "epoch": 10.94311377245509, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.142, "step": 7310 }, { "epoch": 10.944610778443113, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1442, "step": 7311 }, { "epoch": 10.946107784431138, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.139, "step": 7312 }, { "epoch": 10.947604790419161, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1474, "step": 7313 }, { "epoch": 10.949101796407186, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1478, "step": 7314 }, { "epoch": 10.95059880239521, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1513, "step": 7315 }, { "epoch": 10.952095808383234, "grad_norm": 0.04931640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 7316 }, { "epoch": 10.953592814371257, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1426, "step": 7317 }, { "epoch": 10.955089820359282, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1396, "step": 7318 }, { "epoch": 10.956586826347305, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1419, "step": 7319 }, { "epoch": 10.95808383233533, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1459, "step": 7320 }, { "epoch": 10.959580838323353, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1437, "step": 7321 }, { "epoch": 10.961077844311378, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1476, "step": 7322 }, { "epoch": 10.9625748502994, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.145, "step": 7323 }, { "epoch": 10.964071856287426, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1483, "step": 7324 }, { "epoch": 10.965568862275449, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.143, "step": 7325 }, { "epoch": 10.967065868263473, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1456, "step": 7326 }, { "epoch": 10.968562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1455, "step": 7327 }, { "epoch": 10.970059880239521, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1376, "step": 7328 }, { "epoch": 10.971556886227544, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.143, "step": 7329 }, { "epoch": 10.97305389221557, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1394, "step": 7330 }, { "epoch": 10.974550898203592, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.141, "step": 7331 }, { "epoch": 10.976047904191617, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1466, "step": 7332 }, { "epoch": 10.97754491017964, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1469, "step": 7333 }, { "epoch": 10.979041916167665, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1514, "step": 7334 }, { "epoch": 10.980538922155688, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1397, "step": 7335 }, { "epoch": 10.982035928143713, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1378, "step": 7336 }, { "epoch": 10.983532934131736, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.142, "step": 7337 }, { "epoch": 10.98502994011976, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1447, "step": 7338 }, { "epoch": 10.986526946107784, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1484, "step": 7339 }, { "epoch": 10.988023952095809, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1473, "step": 7340 }, { "epoch": 10.989520958083832, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1475, "step": 7341 }, { "epoch": 10.991017964071856, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1426, "step": 7342 }, { "epoch": 10.99251497005988, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1407, "step": 7343 }, { "epoch": 10.994011976047904, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1458, "step": 7344 }, { "epoch": 10.995508982035929, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1433, "step": 7345 }, { "epoch": 10.997005988023952, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1411, "step": 7346 }, { "epoch": 10.998502994011975, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1384, "step": 7347 }, { "epoch": 11.0, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7348 }, { "epoch": 11.001497005988025, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1442, "step": 7349 }, { "epoch": 11.002994011976048, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1447, "step": 7350 }, { "epoch": 11.004491017964073, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1396, "step": 7351 }, { "epoch": 11.005988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1493, "step": 7352 }, { "epoch": 11.00748502994012, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1477, "step": 7353 }, { "epoch": 11.008982035928144, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1402, "step": 7354 }, { "epoch": 11.010479041916168, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1495, "step": 7355 }, { "epoch": 11.011976047904191, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 7356 }, { "epoch": 11.013473053892216, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1416, "step": 7357 }, { "epoch": 11.01497005988024, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.143, "step": 7358 }, { "epoch": 11.016467065868264, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.141, "step": 7359 }, { "epoch": 11.017964071856287, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 7360 }, { "epoch": 11.019461077844312, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1466, "step": 7361 }, { "epoch": 11.020958083832335, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 7362 }, { "epoch": 11.02245508982036, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1492, "step": 7363 }, { "epoch": 11.023952095808383, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.141, "step": 7364 }, { "epoch": 11.025449101796408, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1469, "step": 7365 }, { "epoch": 11.02694610778443, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1447, "step": 7366 }, { "epoch": 11.028443113772456, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1405, "step": 7367 }, { "epoch": 11.029940119760479, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1435, "step": 7368 }, { "epoch": 11.031437125748504, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1386, "step": 7369 }, { "epoch": 11.032934131736527, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1501, "step": 7370 }, { "epoch": 11.034431137724551, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1431, "step": 7371 }, { "epoch": 11.035928143712574, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1369, "step": 7372 }, { "epoch": 11.0374251497006, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1437, "step": 7373 }, { "epoch": 11.038922155688622, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1417, "step": 7374 }, { "epoch": 11.040419161676647, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1365, "step": 7375 }, { "epoch": 11.04191616766467, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1436, "step": 7376 }, { "epoch": 11.043413173652695, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1462, "step": 7377 }, { "epoch": 11.044910179640718, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1412, "step": 7378 }, { "epoch": 11.046407185628743, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1425, "step": 7379 }, { "epoch": 11.047904191616766, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.146, "step": 7380 }, { "epoch": 11.04940119760479, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1526, "step": 7381 }, { "epoch": 11.050898203592814, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1386, "step": 7382 }, { "epoch": 11.052395209580839, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1376, "step": 7383 }, { "epoch": 11.053892215568862, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1526, "step": 7384 }, { "epoch": 11.055389221556887, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 7385 }, { "epoch": 11.05688622754491, "grad_norm": 0.049560546875, "learning_rate": 0.0008, "loss": 1.146, "step": 7386 }, { "epoch": 11.058383233532934, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1452, "step": 7387 }, { "epoch": 11.059880239520957, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.136, "step": 7388 }, { "epoch": 11.061377245508982, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1465, "step": 7389 }, { "epoch": 11.062874251497005, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1463, "step": 7390 }, { "epoch": 11.06437125748503, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.151, "step": 7391 }, { "epoch": 11.065868263473053, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1463, "step": 7392 }, { "epoch": 11.067365269461078, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1458, "step": 7393 }, { "epoch": 11.068862275449101, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.15, "step": 7394 }, { "epoch": 11.070359281437126, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1432, "step": 7395 }, { "epoch": 11.071856287425149, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1456, "step": 7396 }, { "epoch": 11.073353293413174, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.141, "step": 7397 }, { "epoch": 11.074850299401197, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 7398 }, { "epoch": 11.076347305389222, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1408, "step": 7399 }, { "epoch": 11.077844311377245, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1428, "step": 7400 }, { "epoch": 11.07934131736527, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1494, "step": 7401 }, { "epoch": 11.080838323353293, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1455, "step": 7402 }, { "epoch": 11.082335329341317, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1502, "step": 7403 }, { "epoch": 11.08383233532934, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1481, "step": 7404 }, { "epoch": 11.085329341317365, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1426, "step": 7405 }, { "epoch": 11.08682634730539, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1448, "step": 7406 }, { "epoch": 11.088323353293413, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.147, "step": 7407 }, { "epoch": 11.089820359281438, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1443, "step": 7408 }, { "epoch": 11.091317365269461, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1495, "step": 7409 }, { "epoch": 11.092814371257486, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1446, "step": 7410 }, { "epoch": 11.094311377245509, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1468, "step": 7411 }, { "epoch": 11.095808383233534, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1382, "step": 7412 }, { "epoch": 11.097305389221557, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1411, "step": 7413 }, { "epoch": 11.098802395209582, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1494, "step": 7414 }, { "epoch": 11.100299401197605, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1432, "step": 7415 }, { "epoch": 11.10179640718563, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1383, "step": 7416 }, { "epoch": 11.103293413173652, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1392, "step": 7417 }, { "epoch": 11.104790419161677, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1453, "step": 7418 }, { "epoch": 11.1062874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1443, "step": 7419 }, { "epoch": 11.107784431137725, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1451, "step": 7420 }, { "epoch": 11.109281437125748, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1365, "step": 7421 }, { "epoch": 11.110778443113773, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1432, "step": 7422 }, { "epoch": 11.112275449101796, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1455, "step": 7423 }, { "epoch": 11.113772455089821, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1456, "step": 7424 }, { "epoch": 11.115269461077844, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 7425 }, { "epoch": 11.116766467065869, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1517, "step": 7426 }, { "epoch": 11.118263473053892, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7427 }, { "epoch": 11.119760479041917, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1443, "step": 7428 }, { "epoch": 11.12125748502994, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1442, "step": 7429 }, { "epoch": 11.122754491017965, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1427, "step": 7430 }, { "epoch": 11.124251497005988, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 7431 }, { "epoch": 11.125748502994012, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1475, "step": 7432 }, { "epoch": 11.127245508982035, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1479, "step": 7433 }, { "epoch": 11.12874251497006, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1435, "step": 7434 }, { "epoch": 11.130239520958083, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1479, "step": 7435 }, { "epoch": 11.131736526946108, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1397, "step": 7436 }, { "epoch": 11.133233532934131, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.145, "step": 7437 }, { "epoch": 11.134730538922156, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1484, "step": 7438 }, { "epoch": 11.136227544910179, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1442, "step": 7439 }, { "epoch": 11.137724550898204, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1472, "step": 7440 }, { "epoch": 11.139221556886227, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1443, "step": 7441 }, { "epoch": 11.140718562874252, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1455, "step": 7442 }, { "epoch": 11.142215568862275, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1423, "step": 7443 }, { "epoch": 11.1437125748503, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1487, "step": 7444 }, { "epoch": 11.145209580838323, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1515, "step": 7445 }, { "epoch": 11.146706586826348, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 7446 }, { "epoch": 11.14820359281437, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1477, "step": 7447 }, { "epoch": 11.149700598802395, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1421, "step": 7448 }, { "epoch": 11.151197604790418, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1368, "step": 7449 }, { "epoch": 11.152694610778443, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1481, "step": 7450 }, { "epoch": 11.154191616766466, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.1481, "step": 7451 }, { "epoch": 11.155688622754491, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1433, "step": 7452 }, { "epoch": 11.157185628742514, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1414, "step": 7453 }, { "epoch": 11.158682634730539, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7454 }, { "epoch": 11.160179640718562, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1533, "step": 7455 }, { "epoch": 11.161676646706587, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.147, "step": 7456 }, { "epoch": 11.16317365269461, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1459, "step": 7457 }, { "epoch": 11.164670658682635, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 7458 }, { "epoch": 11.16616766467066, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1556, "step": 7459 }, { "epoch": 11.167664670658683, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1391, "step": 7460 }, { "epoch": 11.169161676646707, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1461, "step": 7461 }, { "epoch": 11.17065868263473, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1474, "step": 7462 }, { "epoch": 11.172155688622755, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.144, "step": 7463 }, { "epoch": 11.173652694610778, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1422, "step": 7464 }, { "epoch": 11.175149700598803, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1464, "step": 7465 }, { "epoch": 11.176646706586826, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1424, "step": 7466 }, { "epoch": 11.178143712574851, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1493, "step": 7467 }, { "epoch": 11.179640718562874, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1441, "step": 7468 }, { "epoch": 11.181137724550899, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 7469 }, { "epoch": 11.182634730538922, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1388, "step": 7470 }, { "epoch": 11.184131736526947, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1506, "step": 7471 }, { "epoch": 11.18562874251497, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1477, "step": 7472 }, { "epoch": 11.187125748502995, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1452, "step": 7473 }, { "epoch": 11.188622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 7474 }, { "epoch": 11.190119760479043, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 7475 }, { "epoch": 11.191616766467066, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1515, "step": 7476 }, { "epoch": 11.19311377245509, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1436, "step": 7477 }, { "epoch": 11.194610778443113, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1475, "step": 7478 }, { "epoch": 11.196107784431138, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1406, "step": 7479 }, { "epoch": 11.197604790419161, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7480 }, { "epoch": 11.199101796407186, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1486, "step": 7481 }, { "epoch": 11.20059880239521, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1442, "step": 7482 }, { "epoch": 11.202095808383234, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1428, "step": 7483 }, { "epoch": 11.203592814371257, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1441, "step": 7484 }, { "epoch": 11.205089820359282, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1391, "step": 7485 }, { "epoch": 11.206586826347305, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.15, "step": 7486 }, { "epoch": 11.20808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1425, "step": 7487 }, { "epoch": 11.209580838323353, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1404, "step": 7488 }, { "epoch": 11.211077844311378, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1447, "step": 7489 }, { "epoch": 11.2125748502994, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1471, "step": 7490 }, { "epoch": 11.214071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1443, "step": 7491 }, { "epoch": 11.215568862275449, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1417, "step": 7492 }, { "epoch": 11.217065868263473, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1419, "step": 7493 }, { "epoch": 11.218562874251496, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1367, "step": 7494 }, { "epoch": 11.220059880239521, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1523, "step": 7495 }, { "epoch": 11.221556886227544, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1427, "step": 7496 }, { "epoch": 11.22305389221557, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1453, "step": 7497 }, { "epoch": 11.224550898203592, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1469, "step": 7498 }, { "epoch": 11.226047904191617, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1459, "step": 7499 }, { "epoch": 11.22754491017964, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.146, "step": 7500 }, { "epoch": 11.229041916167665, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1409, "step": 7501 }, { "epoch": 11.230538922155688, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1408, "step": 7502 }, { "epoch": 11.232035928143713, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.141, "step": 7503 }, { "epoch": 11.233532934131736, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1366, "step": 7504 }, { "epoch": 11.23502994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.145, "step": 7505 }, { "epoch": 11.236526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1445, "step": 7506 }, { "epoch": 11.238023952095809, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.145, "step": 7507 }, { "epoch": 11.239520958083832, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1541, "step": 7508 }, { "epoch": 11.241017964071856, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 7509 }, { "epoch": 11.24251497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1449, "step": 7510 }, { "epoch": 11.244011976047904, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1494, "step": 7511 }, { "epoch": 11.245508982035927, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1387, "step": 7512 }, { "epoch": 11.247005988023952, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1489, "step": 7513 }, { "epoch": 11.248502994011975, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1505, "step": 7514 }, { "epoch": 11.25, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 7515 }, { "epoch": 11.251497005988025, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1468, "step": 7516 }, { "epoch": 11.252994011976048, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1445, "step": 7517 }, { "epoch": 11.254491017964073, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1434, "step": 7518 }, { "epoch": 11.255988023952096, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 7519 }, { "epoch": 11.25748502994012, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1464, "step": 7520 }, { "epoch": 11.258982035928144, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1468, "step": 7521 }, { "epoch": 11.260479041916168, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1419, "step": 7522 }, { "epoch": 11.261976047904191, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1494, "step": 7523 }, { "epoch": 11.263473053892216, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1391, "step": 7524 }, { "epoch": 11.26497005988024, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1478, "step": 7525 }, { "epoch": 11.266467065868264, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.148, "step": 7526 }, { "epoch": 11.267964071856287, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1334, "step": 7527 }, { "epoch": 11.269461077844312, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1419, "step": 7528 }, { "epoch": 11.270958083832335, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1465, "step": 7529 }, { "epoch": 11.27245508982036, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.145, "step": 7530 }, { "epoch": 11.273952095808383, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.141, "step": 7531 }, { "epoch": 11.275449101796408, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1441, "step": 7532 }, { "epoch": 11.27694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1397, "step": 7533 }, { "epoch": 11.278443113772456, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1448, "step": 7534 }, { "epoch": 11.279940119760479, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1437, "step": 7535 }, { "epoch": 11.281437125748504, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 7536 }, { "epoch": 11.282934131736527, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1429, "step": 7537 }, { "epoch": 11.284431137724551, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.144, "step": 7538 }, { "epoch": 11.285928143712574, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1434, "step": 7539 }, { "epoch": 11.2874251497006, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1424, "step": 7540 }, { "epoch": 11.288922155688622, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1461, "step": 7541 }, { "epoch": 11.290419161676647, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1451, "step": 7542 }, { "epoch": 11.29191616766467, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1435, "step": 7543 }, { "epoch": 11.293413173652695, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1532, "step": 7544 }, { "epoch": 11.294910179640718, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1496, "step": 7545 }, { "epoch": 11.296407185628743, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1403, "step": 7546 }, { "epoch": 11.297904191616766, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1388, "step": 7547 }, { "epoch": 11.29940119760479, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1409, "step": 7548 }, { "epoch": 11.300898203592814, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1445, "step": 7549 }, { "epoch": 11.302395209580839, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1406, "step": 7550 }, { "epoch": 11.303892215568862, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1506, "step": 7551 }, { "epoch": 11.305389221556887, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1368, "step": 7552 }, { "epoch": 11.30688622754491, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1415, "step": 7553 }, { "epoch": 11.308383233532934, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1457, "step": 7554 }, { "epoch": 11.309880239520957, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1405, "step": 7555 }, { "epoch": 11.311377245508982, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1348, "step": 7556 }, { "epoch": 11.312874251497005, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1385, "step": 7557 }, { "epoch": 11.31437125748503, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.139, "step": 7558 }, { "epoch": 11.315868263473053, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1418, "step": 7559 }, { "epoch": 11.317365269461078, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1437, "step": 7560 }, { "epoch": 11.318862275449101, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1361, "step": 7561 }, { "epoch": 11.320359281437126, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1451, "step": 7562 }, { "epoch": 11.321856287425149, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1477, "step": 7563 }, { "epoch": 11.323353293413174, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1439, "step": 7564 }, { "epoch": 11.324850299401197, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1456, "step": 7565 }, { "epoch": 11.326347305389222, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7566 }, { "epoch": 11.327844311377245, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1387, "step": 7567 }, { "epoch": 11.32934131736527, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1478, "step": 7568 }, { "epoch": 11.330838323353294, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 7569 }, { "epoch": 11.332335329341317, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1446, "step": 7570 }, { "epoch": 11.33383233532934, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1439, "step": 7571 }, { "epoch": 11.335329341317365, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1421, "step": 7572 }, { "epoch": 11.33682634730539, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1453, "step": 7573 }, { "epoch": 11.338323353293413, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1489, "step": 7574 }, { "epoch": 11.339820359281438, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1457, "step": 7575 }, { "epoch": 11.341317365269461, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.143, "step": 7576 }, { "epoch": 11.342814371257486, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1389, "step": 7577 }, { "epoch": 11.344311377245509, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1417, "step": 7578 }, { "epoch": 11.345808383233534, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1467, "step": 7579 }, { "epoch": 11.347305389221557, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1439, "step": 7580 }, { "epoch": 11.348802395209582, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1409, "step": 7581 }, { "epoch": 11.350299401197605, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1397, "step": 7582 }, { "epoch": 11.35179640718563, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.143, "step": 7583 }, { "epoch": 11.353293413173652, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1491, "step": 7584 }, { "epoch": 11.354790419161677, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1428, "step": 7585 }, { "epoch": 11.3562874251497, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1473, "step": 7586 }, { "epoch": 11.357784431137725, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1408, "step": 7587 }, { "epoch": 11.359281437125748, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1416, "step": 7588 }, { "epoch": 11.360778443113773, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.145, "step": 7589 }, { "epoch": 11.362275449101796, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1449, "step": 7590 }, { "epoch": 11.363772455089821, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1453, "step": 7591 }, { "epoch": 11.365269461077844, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1458, "step": 7592 }, { "epoch": 11.366766467065869, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1435, "step": 7593 }, { "epoch": 11.368263473053892, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1478, "step": 7594 }, { "epoch": 11.369760479041917, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1459, "step": 7595 }, { "epoch": 11.37125748502994, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1362, "step": 7596 }, { "epoch": 11.372754491017965, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1393, "step": 7597 }, { "epoch": 11.374251497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1496, "step": 7598 }, { "epoch": 11.375748502994012, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1496, "step": 7599 }, { "epoch": 11.377245508982035, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7600 }, { "epoch": 11.37874251497006, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1448, "step": 7601 }, { "epoch": 11.380239520958083, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 7602 }, { "epoch": 11.381736526946108, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1474, "step": 7603 }, { "epoch": 11.383233532934131, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1367, "step": 7604 }, { "epoch": 11.384730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1431, "step": 7605 }, { "epoch": 11.386227544910179, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1437, "step": 7606 }, { "epoch": 11.387724550898204, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7607 }, { "epoch": 11.389221556886227, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1422, "step": 7608 }, { "epoch": 11.390718562874252, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1437, "step": 7609 }, { "epoch": 11.392215568862275, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1502, "step": 7610 }, { "epoch": 11.3937125748503, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1469, "step": 7611 }, { "epoch": 11.395209580838323, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1408, "step": 7612 }, { "epoch": 11.396706586826348, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1381, "step": 7613 }, { "epoch": 11.39820359281437, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1408, "step": 7614 }, { "epoch": 11.399700598802395, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1484, "step": 7615 }, { "epoch": 11.401197604790418, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1441, "step": 7616 }, { "epoch": 11.402694610778443, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1415, "step": 7617 }, { "epoch": 11.404191616766466, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1464, "step": 7618 }, { "epoch": 11.405688622754491, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1436, "step": 7619 }, { "epoch": 11.407185628742514, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1391, "step": 7620 }, { "epoch": 11.408682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1406, "step": 7621 }, { "epoch": 11.410179640718562, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1396, "step": 7622 }, { "epoch": 11.411676646706587, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 7623 }, { "epoch": 11.41317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1442, "step": 7624 }, { "epoch": 11.414670658682635, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1386, "step": 7625 }, { "epoch": 11.41616766467066, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1409, "step": 7626 }, { "epoch": 11.417664670658683, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1415, "step": 7627 }, { "epoch": 11.419161676646706, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1414, "step": 7628 }, { "epoch": 11.42065868263473, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1471, "step": 7629 }, { "epoch": 11.422155688622755, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1464, "step": 7630 }, { "epoch": 11.423652694610778, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1386, "step": 7631 }, { "epoch": 11.425149700598803, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1339, "step": 7632 }, { "epoch": 11.426646706586826, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1363, "step": 7633 }, { "epoch": 11.428143712574851, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1391, "step": 7634 }, { "epoch": 11.429640718562874, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1475, "step": 7635 }, { "epoch": 11.431137724550899, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1475, "step": 7636 }, { "epoch": 11.432634730538922, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 7637 }, { "epoch": 11.434131736526947, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1412, "step": 7638 }, { "epoch": 11.43562874251497, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 7639 }, { "epoch": 11.437125748502995, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1435, "step": 7640 }, { "epoch": 11.438622754491018, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1392, "step": 7641 }, { "epoch": 11.440119760479043, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1445, "step": 7642 }, { "epoch": 11.441616766467066, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1426, "step": 7643 }, { "epoch": 11.44311377245509, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1459, "step": 7644 }, { "epoch": 11.444610778443113, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1422, "step": 7645 }, { "epoch": 11.446107784431138, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1415, "step": 7646 }, { "epoch": 11.447604790419161, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.145, "step": 7647 }, { "epoch": 11.449101796407186, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 7648 }, { "epoch": 11.45059880239521, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1403, "step": 7649 }, { "epoch": 11.452095808383234, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1416, "step": 7650 }, { "epoch": 11.453592814371257, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1461, "step": 7651 }, { "epoch": 11.455089820359282, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1468, "step": 7652 }, { "epoch": 11.456586826347305, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1477, "step": 7653 }, { "epoch": 11.45808383233533, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1507, "step": 7654 }, { "epoch": 11.459580838323353, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1478, "step": 7655 }, { "epoch": 11.461077844311378, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 7656 }, { "epoch": 11.4625748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1417, "step": 7657 }, { "epoch": 11.464071856287426, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1457, "step": 7658 }, { "epoch": 11.465568862275449, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1415, "step": 7659 }, { "epoch": 11.467065868263473, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1483, "step": 7660 }, { "epoch": 11.468562874251496, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1438, "step": 7661 }, { "epoch": 11.470059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.139, "step": 7662 }, { "epoch": 11.471556886227544, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1403, "step": 7663 }, { "epoch": 11.47305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1441, "step": 7664 }, { "epoch": 11.474550898203592, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1445, "step": 7665 }, { "epoch": 11.476047904191617, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1379, "step": 7666 }, { "epoch": 11.47754491017964, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1413, "step": 7667 }, { "epoch": 11.479041916167665, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1437, "step": 7668 }, { "epoch": 11.480538922155688, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 7669 }, { "epoch": 11.482035928143713, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1436, "step": 7670 }, { "epoch": 11.483532934131736, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1447, "step": 7671 }, { "epoch": 11.48502994011976, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1388, "step": 7672 }, { "epoch": 11.486526946107784, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1436, "step": 7673 }, { "epoch": 11.488023952095809, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1393, "step": 7674 }, { "epoch": 11.489520958083832, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1462, "step": 7675 }, { "epoch": 11.491017964071856, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1421, "step": 7676 }, { "epoch": 11.49251497005988, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1406, "step": 7677 }, { "epoch": 11.494011976047904, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 7678 }, { "epoch": 11.495508982035927, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1493, "step": 7679 }, { "epoch": 11.497005988023952, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1456, "step": 7680 }, { "epoch": 11.498502994011975, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1491, "step": 7681 }, { "epoch": 11.5, "grad_norm": 0.051025390625, "learning_rate": 0.0008, "loss": 1.1402, "step": 7682 }, { "epoch": 11.501497005988025, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1405, "step": 7683 }, { "epoch": 11.502994011976048, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1403, "step": 7684 }, { "epoch": 11.504491017964071, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.141, "step": 7685 }, { "epoch": 11.505988023952096, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1388, "step": 7686 }, { "epoch": 11.50748502994012, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1427, "step": 7687 }, { "epoch": 11.508982035928144, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1435, "step": 7688 }, { "epoch": 11.510479041916168, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1387, "step": 7689 }, { "epoch": 11.511976047904191, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1471, "step": 7690 }, { "epoch": 11.513473053892216, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.146, "step": 7691 }, { "epoch": 11.51497005988024, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1477, "step": 7692 }, { "epoch": 11.516467065868264, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1435, "step": 7693 }, { "epoch": 11.517964071856287, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1417, "step": 7694 }, { "epoch": 11.519461077844312, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1428, "step": 7695 }, { "epoch": 11.520958083832335, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.138, "step": 7696 }, { "epoch": 11.52245508982036, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.135, "step": 7697 }, { "epoch": 11.523952095808383, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1437, "step": 7698 }, { "epoch": 11.525449101796408, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1498, "step": 7699 }, { "epoch": 11.52694610778443, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1491, "step": 7700 }, { "epoch": 11.528443113772456, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1364, "step": 7701 }, { "epoch": 11.529940119760479, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1454, "step": 7702 }, { "epoch": 11.531437125748504, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1389, "step": 7703 }, { "epoch": 11.532934131736527, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1485, "step": 7704 }, { "epoch": 11.534431137724551, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.143, "step": 7705 }, { "epoch": 11.535928143712574, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1384, "step": 7706 }, { "epoch": 11.5374251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1403, "step": 7707 }, { "epoch": 11.538922155688622, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1481, "step": 7708 }, { "epoch": 11.540419161676647, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1391, "step": 7709 }, { "epoch": 11.54191616766467, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1467, "step": 7710 }, { "epoch": 11.543413173652695, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1412, "step": 7711 }, { "epoch": 11.544910179640718, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1452, "step": 7712 }, { "epoch": 11.546407185628743, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1438, "step": 7713 }, { "epoch": 11.547904191616766, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1455, "step": 7714 }, { "epoch": 11.54940119760479, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1397, "step": 7715 }, { "epoch": 11.550898203592814, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1436, "step": 7716 }, { "epoch": 11.552395209580839, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1456, "step": 7717 }, { "epoch": 11.553892215568862, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1409, "step": 7718 }, { "epoch": 11.555389221556887, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1402, "step": 7719 }, { "epoch": 11.55688622754491, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1458, "step": 7720 }, { "epoch": 11.558383233532934, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 7721 }, { "epoch": 11.559880239520957, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1382, "step": 7722 }, { "epoch": 11.561377245508982, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1451, "step": 7723 }, { "epoch": 11.562874251497005, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.147, "step": 7724 }, { "epoch": 11.56437125748503, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1411, "step": 7725 }, { "epoch": 11.565868263473053, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1519, "step": 7726 }, { "epoch": 11.567365269461078, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1456, "step": 7727 }, { "epoch": 11.568862275449101, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 7728 }, { "epoch": 11.570359281437126, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1397, "step": 7729 }, { "epoch": 11.571856287425149, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 7730 }, { "epoch": 11.573353293413174, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1419, "step": 7731 }, { "epoch": 11.574850299401197, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1466, "step": 7732 }, { "epoch": 11.576347305389222, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1418, "step": 7733 }, { "epoch": 11.577844311377245, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1487, "step": 7734 }, { "epoch": 11.57934131736527, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.142, "step": 7735 }, { "epoch": 11.580838323353294, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1363, "step": 7736 }, { "epoch": 11.582335329341317, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1384, "step": 7737 }, { "epoch": 11.58383233532934, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1452, "step": 7738 }, { "epoch": 11.585329341317365, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1361, "step": 7739 }, { "epoch": 11.58682634730539, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1329, "step": 7740 }, { "epoch": 11.588323353293413, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1477, "step": 7741 }, { "epoch": 11.589820359281438, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1378, "step": 7742 }, { "epoch": 11.591317365269461, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1458, "step": 7743 }, { "epoch": 11.592814371257486, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1477, "step": 7744 }, { "epoch": 11.594311377245509, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1374, "step": 7745 }, { "epoch": 11.595808383233534, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1431, "step": 7746 }, { "epoch": 11.597305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7747 }, { "epoch": 11.598802395209582, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1448, "step": 7748 }, { "epoch": 11.600299401197605, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1379, "step": 7749 }, { "epoch": 11.60179640718563, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1464, "step": 7750 }, { "epoch": 11.603293413173652, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1382, "step": 7751 }, { "epoch": 11.604790419161677, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1424, "step": 7752 }, { "epoch": 11.6062874251497, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1471, "step": 7753 }, { "epoch": 11.607784431137725, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1368, "step": 7754 }, { "epoch": 11.609281437125748, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1442, "step": 7755 }, { "epoch": 11.610778443113773, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1403, "step": 7756 }, { "epoch": 11.612275449101796, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1433, "step": 7757 }, { "epoch": 11.613772455089821, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.138, "step": 7758 }, { "epoch": 11.615269461077844, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1395, "step": 7759 }, { "epoch": 11.616766467065869, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1439, "step": 7760 }, { "epoch": 11.618263473053892, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7761 }, { "epoch": 11.619760479041917, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.142, "step": 7762 }, { "epoch": 11.62125748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1446, "step": 7763 }, { "epoch": 11.622754491017965, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.144, "step": 7764 }, { "epoch": 11.624251497005988, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1461, "step": 7765 }, { "epoch": 11.625748502994012, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1524, "step": 7766 }, { "epoch": 11.627245508982035, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1465, "step": 7767 }, { "epoch": 11.62874251497006, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1477, "step": 7768 }, { "epoch": 11.630239520958083, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1453, "step": 7769 }, { "epoch": 11.631736526946108, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1473, "step": 7770 }, { "epoch": 11.633233532934131, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1377, "step": 7771 }, { "epoch": 11.634730538922156, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1406, "step": 7772 }, { "epoch": 11.636227544910179, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1454, "step": 7773 }, { "epoch": 11.637724550898204, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1435, "step": 7774 }, { "epoch": 11.639221556886227, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1459, "step": 7775 }, { "epoch": 11.640718562874252, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1453, "step": 7776 }, { "epoch": 11.642215568862275, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1421, "step": 7777 }, { "epoch": 11.6437125748503, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1412, "step": 7778 }, { "epoch": 11.645209580838323, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1361, "step": 7779 }, { "epoch": 11.646706586826348, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1408, "step": 7780 }, { "epoch": 11.64820359281437, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1402, "step": 7781 }, { "epoch": 11.649700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1456, "step": 7782 }, { "epoch": 11.651197604790418, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1497, "step": 7783 }, { "epoch": 11.652694610778443, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1496, "step": 7784 }, { "epoch": 11.654191616766466, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1371, "step": 7785 }, { "epoch": 11.655688622754491, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1473, "step": 7786 }, { "epoch": 11.657185628742514, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1414, "step": 7787 }, { "epoch": 11.658682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1491, "step": 7788 }, { "epoch": 11.660179640718562, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1421, "step": 7789 }, { "epoch": 11.661676646706587, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1465, "step": 7790 }, { "epoch": 11.66317365269461, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7791 }, { "epoch": 11.664670658682635, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1415, "step": 7792 }, { "epoch": 11.66616766467066, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1466, "step": 7793 }, { "epoch": 11.667664670658683, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1454, "step": 7794 }, { "epoch": 11.669161676646706, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1436, "step": 7795 }, { "epoch": 11.67065868263473, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1478, "step": 7796 }, { "epoch": 11.672155688622755, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1426, "step": 7797 }, { "epoch": 11.673652694610778, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 7798 }, { "epoch": 11.675149700598803, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1348, "step": 7799 }, { "epoch": 11.676646706586826, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1386, "step": 7800 }, { "epoch": 11.678143712574851, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1454, "step": 7801 }, { "epoch": 11.679640718562874, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7802 }, { "epoch": 11.681137724550899, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1399, "step": 7803 }, { "epoch": 11.682634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1473, "step": 7804 }, { "epoch": 11.684131736526947, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1409, "step": 7805 }, { "epoch": 11.68562874251497, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7806 }, { "epoch": 11.687125748502995, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1351, "step": 7807 }, { "epoch": 11.688622754491018, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.145, "step": 7808 }, { "epoch": 11.690119760479043, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 7809 }, { "epoch": 11.691616766467066, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1378, "step": 7810 }, { "epoch": 11.69311377245509, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1399, "step": 7811 }, { "epoch": 11.694610778443113, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1495, "step": 7812 }, { "epoch": 11.696107784431138, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.147, "step": 7813 }, { "epoch": 11.697604790419161, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1419, "step": 7814 }, { "epoch": 11.699101796407186, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1394, "step": 7815 }, { "epoch": 11.70059880239521, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1455, "step": 7816 }, { "epoch": 11.702095808383234, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1548, "step": 7817 }, { "epoch": 11.703592814371257, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1489, "step": 7818 }, { "epoch": 11.705089820359282, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1419, "step": 7819 }, { "epoch": 11.706586826347305, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1447, "step": 7820 }, { "epoch": 11.70808383233533, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 7821 }, { "epoch": 11.709580838323353, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1391, "step": 7822 }, { "epoch": 11.711077844311378, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1397, "step": 7823 }, { "epoch": 11.7125748502994, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1437, "step": 7824 }, { "epoch": 11.714071856287426, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1431, "step": 7825 }, { "epoch": 11.715568862275449, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1473, "step": 7826 }, { "epoch": 11.717065868263473, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1396, "step": 7827 }, { "epoch": 11.718562874251496, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1491, "step": 7828 }, { "epoch": 11.720059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1462, "step": 7829 }, { "epoch": 11.721556886227544, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1438, "step": 7830 }, { "epoch": 11.72305389221557, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.148, "step": 7831 }, { "epoch": 11.724550898203592, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1408, "step": 7832 }, { "epoch": 11.726047904191617, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1444, "step": 7833 }, { "epoch": 11.72754491017964, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1405, "step": 7834 }, { "epoch": 11.729041916167665, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1383, "step": 7835 }, { "epoch": 11.730538922155688, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1458, "step": 7836 }, { "epoch": 11.732035928143713, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.142, "step": 7837 }, { "epoch": 11.733532934131736, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1405, "step": 7838 }, { "epoch": 11.73502994011976, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1444, "step": 7839 }, { "epoch": 11.736526946107784, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1451, "step": 7840 }, { "epoch": 11.738023952095809, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1359, "step": 7841 }, { "epoch": 11.739520958083832, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1438, "step": 7842 }, { "epoch": 11.741017964071856, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1351, "step": 7843 }, { "epoch": 11.74251497005988, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.145, "step": 7844 }, { "epoch": 11.744011976047904, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1415, "step": 7845 }, { "epoch": 11.745508982035929, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1421, "step": 7846 }, { "epoch": 11.747005988023952, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1416, "step": 7847 }, { "epoch": 11.748502994011975, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1454, "step": 7848 }, { "epoch": 11.75, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1426, "step": 7849 }, { "epoch": 11.751497005988025, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.145, "step": 7850 }, { "epoch": 11.752994011976048, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.143, "step": 7851 }, { "epoch": 11.754491017964071, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1376, "step": 7852 }, { "epoch": 11.755988023952096, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1444, "step": 7853 }, { "epoch": 11.75748502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.14, "step": 7854 }, { "epoch": 11.758982035928144, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.142, "step": 7855 }, { "epoch": 11.760479041916168, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1427, "step": 7856 }, { "epoch": 11.761976047904191, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1449, "step": 7857 }, { "epoch": 11.763473053892216, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1457, "step": 7858 }, { "epoch": 11.76497005988024, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1445, "step": 7859 }, { "epoch": 11.766467065868264, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1401, "step": 7860 }, { "epoch": 11.767964071856287, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1457, "step": 7861 }, { "epoch": 11.769461077844312, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1434, "step": 7862 }, { "epoch": 11.770958083832335, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1444, "step": 7863 }, { "epoch": 11.77245508982036, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1407, "step": 7864 }, { "epoch": 11.773952095808383, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1377, "step": 7865 }, { "epoch": 11.775449101796408, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1487, "step": 7866 }, { "epoch": 11.77694610778443, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1471, "step": 7867 }, { "epoch": 11.778443113772456, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1397, "step": 7868 }, { "epoch": 11.779940119760479, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1483, "step": 7869 }, { "epoch": 11.781437125748504, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1394, "step": 7870 }, { "epoch": 11.782934131736527, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1378, "step": 7871 }, { "epoch": 11.784431137724551, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1486, "step": 7872 }, { "epoch": 11.785928143712574, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1469, "step": 7873 }, { "epoch": 11.7874251497006, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1385, "step": 7874 }, { "epoch": 11.788922155688622, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1456, "step": 7875 }, { "epoch": 11.790419161676647, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1471, "step": 7876 }, { "epoch": 11.79191616766467, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1407, "step": 7877 }, { "epoch": 11.793413173652695, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1419, "step": 7878 }, { "epoch": 11.794910179640718, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1421, "step": 7879 }, { "epoch": 11.796407185628743, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1454, "step": 7880 }, { "epoch": 11.797904191616766, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1467, "step": 7881 }, { "epoch": 11.79940119760479, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1409, "step": 7882 }, { "epoch": 11.800898203592814, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1465, "step": 7883 }, { "epoch": 11.802395209580839, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1518, "step": 7884 }, { "epoch": 11.803892215568862, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1376, "step": 7885 }, { "epoch": 11.805389221556887, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1458, "step": 7886 }, { "epoch": 11.80688622754491, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1376, "step": 7887 }, { "epoch": 11.808383233532934, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1508, "step": 7888 }, { "epoch": 11.809880239520957, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1447, "step": 7889 }, { "epoch": 11.811377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1427, "step": 7890 }, { "epoch": 11.812874251497005, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1437, "step": 7891 }, { "epoch": 11.81437125748503, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1421, "step": 7892 }, { "epoch": 11.815868263473053, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.147, "step": 7893 }, { "epoch": 11.817365269461078, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1366, "step": 7894 }, { "epoch": 11.818862275449101, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7895 }, { "epoch": 11.820359281437126, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1401, "step": 7896 }, { "epoch": 11.821856287425149, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1448, "step": 7897 }, { "epoch": 11.823353293413174, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1462, "step": 7898 }, { "epoch": 11.824850299401197, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1465, "step": 7899 }, { "epoch": 11.826347305389222, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1396, "step": 7900 }, { "epoch": 11.827844311377245, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1393, "step": 7901 }, { "epoch": 11.82934131736527, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1384, "step": 7902 }, { "epoch": 11.830838323353294, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1494, "step": 7903 }, { "epoch": 11.832335329341317, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1453, "step": 7904 }, { "epoch": 11.83383233532934, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1418, "step": 7905 }, { "epoch": 11.835329341317365, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1385, "step": 7906 }, { "epoch": 11.83682634730539, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1434, "step": 7907 }, { "epoch": 11.838323353293413, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.143, "step": 7908 }, { "epoch": 11.839820359281438, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1404, "step": 7909 }, { "epoch": 11.841317365269461, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.137, "step": 7910 }, { "epoch": 11.842814371257486, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1461, "step": 7911 }, { "epoch": 11.844311377245509, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1481, "step": 7912 }, { "epoch": 11.845808383233534, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.146, "step": 7913 }, { "epoch": 11.847305389221557, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1433, "step": 7914 }, { "epoch": 11.848802395209582, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7915 }, { "epoch": 11.850299401197605, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1415, "step": 7916 }, { "epoch": 11.85179640718563, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1414, "step": 7917 }, { "epoch": 11.853293413173652, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1399, "step": 7918 }, { "epoch": 11.854790419161677, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1448, "step": 7919 }, { "epoch": 11.8562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1382, "step": 7920 }, { "epoch": 11.857784431137725, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.138, "step": 7921 }, { "epoch": 11.859281437125748, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1455, "step": 7922 }, { "epoch": 11.860778443113773, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1459, "step": 7923 }, { "epoch": 11.862275449101796, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1446, "step": 7924 }, { "epoch": 11.863772455089821, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1503, "step": 7925 }, { "epoch": 11.865269461077844, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1455, "step": 7926 }, { "epoch": 11.866766467065869, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1505, "step": 7927 }, { "epoch": 11.868263473053892, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 7928 }, { "epoch": 11.869760479041917, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1405, "step": 7929 }, { "epoch": 11.87125748502994, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1465, "step": 7930 }, { "epoch": 11.872754491017965, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1379, "step": 7931 }, { "epoch": 11.874251497005988, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1452, "step": 7932 }, { "epoch": 11.875748502994012, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1434, "step": 7933 }, { "epoch": 11.877245508982035, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1441, "step": 7934 }, { "epoch": 11.87874251497006, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1401, "step": 7935 }, { "epoch": 11.880239520958083, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1434, "step": 7936 }, { "epoch": 11.881736526946108, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7937 }, { "epoch": 11.883233532934131, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7938 }, { "epoch": 11.884730538922156, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1427, "step": 7939 }, { "epoch": 11.886227544910179, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1443, "step": 7940 }, { "epoch": 11.887724550898204, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1426, "step": 7941 }, { "epoch": 11.889221556886227, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1456, "step": 7942 }, { "epoch": 11.890718562874252, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1337, "step": 7943 }, { "epoch": 11.892215568862275, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1394, "step": 7944 }, { "epoch": 11.8937125748503, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1403, "step": 7945 }, { "epoch": 11.895209580838323, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1413, "step": 7946 }, { "epoch": 11.896706586826348, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1434, "step": 7947 }, { "epoch": 11.89820359281437, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 7948 }, { "epoch": 11.899700598802395, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.144, "step": 7949 }, { "epoch": 11.901197604790418, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7950 }, { "epoch": 11.902694610778443, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7951 }, { "epoch": 11.904191616766466, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1401, "step": 7952 }, { "epoch": 11.905688622754491, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1384, "step": 7953 }, { "epoch": 11.907185628742514, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1434, "step": 7954 }, { "epoch": 11.908682634730539, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1425, "step": 7955 }, { "epoch": 11.910179640718562, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1442, "step": 7956 }, { "epoch": 11.911676646706587, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 7957 }, { "epoch": 11.91317365269461, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1446, "step": 7958 }, { "epoch": 11.914670658682635, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1412, "step": 7959 }, { "epoch": 11.91616766467066, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1388, "step": 7960 }, { "epoch": 11.917664670658683, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 7961 }, { "epoch": 11.919161676646706, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1448, "step": 7962 }, { "epoch": 11.92065868263473, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.142, "step": 7963 }, { "epoch": 11.922155688622755, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1364, "step": 7964 }, { "epoch": 11.923652694610778, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1348, "step": 7965 }, { "epoch": 11.925149700598803, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1425, "step": 7966 }, { "epoch": 11.926646706586826, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1372, "step": 7967 }, { "epoch": 11.928143712574851, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1397, "step": 7968 }, { "epoch": 11.929640718562874, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1386, "step": 7969 }, { "epoch": 11.931137724550899, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1407, "step": 7970 }, { "epoch": 11.932634730538922, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1425, "step": 7971 }, { "epoch": 11.934131736526947, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1487, "step": 7972 }, { "epoch": 11.93562874251497, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1433, "step": 7973 }, { "epoch": 11.937125748502995, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1472, "step": 7974 }, { "epoch": 11.938622754491018, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1441, "step": 7975 }, { "epoch": 11.940119760479043, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1424, "step": 7976 }, { "epoch": 11.941616766467066, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1441, "step": 7977 }, { "epoch": 11.94311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1387, "step": 7978 }, { "epoch": 11.944610778443113, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1465, "step": 7979 }, { "epoch": 11.946107784431138, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.144, "step": 7980 }, { "epoch": 11.947604790419161, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1489, "step": 7981 }, { "epoch": 11.949101796407186, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 7982 }, { "epoch": 11.95059880239521, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1425, "step": 7983 }, { "epoch": 11.952095808383234, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1383, "step": 7984 }, { "epoch": 11.953592814371257, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1467, "step": 7985 }, { "epoch": 11.955089820359282, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1425, "step": 7986 }, { "epoch": 11.956586826347305, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1384, "step": 7987 }, { "epoch": 11.95808383233533, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1446, "step": 7988 }, { "epoch": 11.959580838323353, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1458, "step": 7989 }, { "epoch": 11.961077844311378, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1357, "step": 7990 }, { "epoch": 11.9625748502994, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1409, "step": 7991 }, { "epoch": 11.964071856287426, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1429, "step": 7992 }, { "epoch": 11.965568862275449, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1496, "step": 7993 }, { "epoch": 11.967065868263473, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1397, "step": 7994 }, { "epoch": 11.968562874251496, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1388, "step": 7995 }, { "epoch": 11.970059880239521, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1428, "step": 7996 }, { "epoch": 11.971556886227544, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1493, "step": 7997 }, { "epoch": 11.97305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1395, "step": 7998 }, { "epoch": 11.974550898203592, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1389, "step": 7999 }, { "epoch": 11.976047904191617, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1449, "step": 8000 }, { "epoch": 11.97754491017964, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8001 }, { "epoch": 11.979041916167665, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 8002 }, { "epoch": 11.980538922155688, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1433, "step": 8003 }, { "epoch": 11.982035928143713, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1366, "step": 8004 }, { "epoch": 11.983532934131736, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1421, "step": 8005 }, { "epoch": 11.98502994011976, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1414, "step": 8006 }, { "epoch": 11.986526946107784, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1452, "step": 8007 }, { "epoch": 11.988023952095809, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1423, "step": 8008 }, { "epoch": 11.989520958083832, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1432, "step": 8009 }, { "epoch": 11.991017964071856, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1429, "step": 8010 }, { "epoch": 11.99251497005988, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1396, "step": 8011 }, { "epoch": 11.994011976047904, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1457, "step": 8012 }, { "epoch": 11.995508982035929, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1432, "step": 8013 }, { "epoch": 11.997005988023952, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1387, "step": 8014 }, { "epoch": 11.998502994011975, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1461, "step": 8015 }, { "epoch": 12.0, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1404, "step": 8016 }, { "epoch": 12.001497005988025, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8017 }, { "epoch": 12.002994011976048, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1431, "step": 8018 }, { "epoch": 12.004491017964073, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1443, "step": 8019 }, { "epoch": 12.005988023952096, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 8020 }, { "epoch": 12.00748502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.137, "step": 8021 }, { "epoch": 12.008982035928144, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1447, "step": 8022 }, { "epoch": 12.010479041916168, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1493, "step": 8023 }, { "epoch": 12.011976047904191, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1408, "step": 8024 }, { "epoch": 12.013473053892216, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 8025 }, { "epoch": 12.01497005988024, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1451, "step": 8026 }, { "epoch": 12.016467065868264, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1384, "step": 8027 }, { "epoch": 12.017964071856287, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1422, "step": 8028 }, { "epoch": 12.019461077844312, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8029 }, { "epoch": 12.020958083832335, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.143, "step": 8030 }, { "epoch": 12.02245508982036, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.145, "step": 8031 }, { "epoch": 12.023952095808383, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1446, "step": 8032 }, { "epoch": 12.025449101796408, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.142, "step": 8033 }, { "epoch": 12.02694610778443, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1441, "step": 8034 }, { "epoch": 12.028443113772456, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1374, "step": 8035 }, { "epoch": 12.029940119760479, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1424, "step": 8036 }, { "epoch": 12.031437125748504, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1437, "step": 8037 }, { "epoch": 12.032934131736527, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1403, "step": 8038 }, { "epoch": 12.034431137724551, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1352, "step": 8039 }, { "epoch": 12.035928143712574, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1424, "step": 8040 }, { "epoch": 12.0374251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1388, "step": 8041 }, { "epoch": 12.038922155688622, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1404, "step": 8042 }, { "epoch": 12.040419161676647, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1436, "step": 8043 }, { "epoch": 12.04191616766467, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1419, "step": 8044 }, { "epoch": 12.043413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1417, "step": 8045 }, { "epoch": 12.044910179640718, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.136, "step": 8046 }, { "epoch": 12.046407185628743, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1436, "step": 8047 }, { "epoch": 12.047904191616766, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8048 }, { "epoch": 12.04940119760479, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1424, "step": 8049 }, { "epoch": 12.050898203592814, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1452, "step": 8050 }, { "epoch": 12.052395209580839, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1498, "step": 8051 }, { "epoch": 12.053892215568862, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1422, "step": 8052 }, { "epoch": 12.055389221556887, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1463, "step": 8053 }, { "epoch": 12.05688622754491, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1412, "step": 8054 }, { "epoch": 12.058383233532934, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 8055 }, { "epoch": 12.059880239520957, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8056 }, { "epoch": 12.061377245508982, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8057 }, { "epoch": 12.062874251497005, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1396, "step": 8058 }, { "epoch": 12.06437125748503, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1382, "step": 8059 }, { "epoch": 12.065868263473053, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1435, "step": 8060 }, { "epoch": 12.067365269461078, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1423, "step": 8061 }, { "epoch": 12.068862275449101, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1507, "step": 8062 }, { "epoch": 12.070359281437126, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1411, "step": 8063 }, { "epoch": 12.071856287425149, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1398, "step": 8064 }, { "epoch": 12.073353293413174, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1448, "step": 8065 }, { "epoch": 12.074850299401197, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1347, "step": 8066 }, { "epoch": 12.076347305389222, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 8067 }, { "epoch": 12.077844311377245, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1427, "step": 8068 }, { "epoch": 12.07934131736527, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1383, "step": 8069 }, { "epoch": 12.080838323353293, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1462, "step": 8070 }, { "epoch": 12.082335329341317, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1376, "step": 8071 }, { "epoch": 12.08383233532934, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.142, "step": 8072 }, { "epoch": 12.085329341317365, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1444, "step": 8073 }, { "epoch": 12.08682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1405, "step": 8074 }, { "epoch": 12.088323353293413, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1438, "step": 8075 }, { "epoch": 12.089820359281438, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1387, "step": 8076 }, { "epoch": 12.091317365269461, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1384, "step": 8077 }, { "epoch": 12.092814371257486, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1414, "step": 8078 }, { "epoch": 12.094311377245509, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1387, "step": 8079 }, { "epoch": 12.095808383233534, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1456, "step": 8080 }, { "epoch": 12.097305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1428, "step": 8081 }, { "epoch": 12.098802395209582, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.138, "step": 8082 }, { "epoch": 12.100299401197605, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1368, "step": 8083 }, { "epoch": 12.10179640718563, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1466, "step": 8084 }, { "epoch": 12.103293413173652, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1426, "step": 8085 }, { "epoch": 12.104790419161677, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1483, "step": 8086 }, { "epoch": 12.1062874251497, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1458, "step": 8087 }, { "epoch": 12.107784431137725, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.141, "step": 8088 }, { "epoch": 12.109281437125748, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 8089 }, { "epoch": 12.110778443113773, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1417, "step": 8090 }, { "epoch": 12.112275449101796, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1407, "step": 8091 }, { "epoch": 12.113772455089821, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1425, "step": 8092 }, { "epoch": 12.115269461077844, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1392, "step": 8093 }, { "epoch": 12.116766467065869, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1403, "step": 8094 }, { "epoch": 12.118263473053892, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1391, "step": 8095 }, { "epoch": 12.119760479041917, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1407, "step": 8096 }, { "epoch": 12.12125748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1395, "step": 8097 }, { "epoch": 12.122754491017965, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8098 }, { "epoch": 12.124251497005988, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1437, "step": 8099 }, { "epoch": 12.125748502994012, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1386, "step": 8100 }, { "epoch": 12.127245508982035, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1443, "step": 8101 }, { "epoch": 12.12874251497006, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1497, "step": 8102 }, { "epoch": 12.130239520958083, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1449, "step": 8103 }, { "epoch": 12.131736526946108, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1475, "step": 8104 }, { "epoch": 12.133233532934131, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1443, "step": 8105 }, { "epoch": 12.134730538922156, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1465, "step": 8106 }, { "epoch": 12.136227544910179, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.146, "step": 8107 }, { "epoch": 12.137724550898204, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1372, "step": 8108 }, { "epoch": 12.139221556886227, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1381, "step": 8109 }, { "epoch": 12.140718562874252, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1415, "step": 8110 }, { "epoch": 12.142215568862275, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1439, "step": 8111 }, { "epoch": 12.1437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1427, "step": 8112 }, { "epoch": 12.145209580838323, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 8113 }, { "epoch": 12.146706586826348, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1437, "step": 8114 }, { "epoch": 12.14820359281437, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1397, "step": 8115 }, { "epoch": 12.149700598802395, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1361, "step": 8116 }, { "epoch": 12.151197604790418, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1431, "step": 8117 }, { "epoch": 12.152694610778443, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1408, "step": 8118 }, { "epoch": 12.154191616766466, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.144, "step": 8119 }, { "epoch": 12.155688622754491, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1462, "step": 8120 }, { "epoch": 12.157185628742514, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.145, "step": 8121 }, { "epoch": 12.158682634730539, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 8122 }, { "epoch": 12.160179640718562, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1378, "step": 8123 }, { "epoch": 12.161676646706587, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1426, "step": 8124 }, { "epoch": 12.16317365269461, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1405, "step": 8125 }, { "epoch": 12.164670658682635, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1449, "step": 8126 }, { "epoch": 12.16616766467066, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1395, "step": 8127 }, { "epoch": 12.167664670658683, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1486, "step": 8128 }, { "epoch": 12.169161676646707, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1418, "step": 8129 }, { "epoch": 12.17065868263473, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1426, "step": 8130 }, { "epoch": 12.172155688622755, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1384, "step": 8131 }, { "epoch": 12.173652694610778, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1481, "step": 8132 }, { "epoch": 12.175149700598803, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1451, "step": 8133 }, { "epoch": 12.176646706586826, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1474, "step": 8134 }, { "epoch": 12.178143712574851, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1461, "step": 8135 }, { "epoch": 12.179640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8136 }, { "epoch": 12.181137724550899, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1414, "step": 8137 }, { "epoch": 12.182634730538922, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1408, "step": 8138 }, { "epoch": 12.184131736526947, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1371, "step": 8139 }, { "epoch": 12.18562874251497, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1429, "step": 8140 }, { "epoch": 12.187125748502995, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1426, "step": 8141 }, { "epoch": 12.188622754491018, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.134, "step": 8142 }, { "epoch": 12.190119760479043, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1459, "step": 8143 }, { "epoch": 12.191616766467066, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1396, "step": 8144 }, { "epoch": 12.19311377245509, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1468, "step": 8145 }, { "epoch": 12.194610778443113, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1395, "step": 8146 }, { "epoch": 12.196107784431138, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1384, "step": 8147 }, { "epoch": 12.197604790419161, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.139, "step": 8148 }, { "epoch": 12.199101796407186, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1364, "step": 8149 }, { "epoch": 12.20059880239521, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1388, "step": 8150 }, { "epoch": 12.202095808383234, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1461, "step": 8151 }, { "epoch": 12.203592814371257, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 8152 }, { "epoch": 12.205089820359282, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.137, "step": 8153 }, { "epoch": 12.206586826347305, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1477, "step": 8154 }, { "epoch": 12.20808383233533, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1371, "step": 8155 }, { "epoch": 12.209580838323353, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.144, "step": 8156 }, { "epoch": 12.211077844311378, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1385, "step": 8157 }, { "epoch": 12.2125748502994, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1391, "step": 8158 }, { "epoch": 12.214071856287426, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.138, "step": 8159 }, { "epoch": 12.215568862275449, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1427, "step": 8160 }, { "epoch": 12.217065868263473, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1416, "step": 8161 }, { "epoch": 12.218562874251496, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1323, "step": 8162 }, { "epoch": 12.220059880239521, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.136, "step": 8163 }, { "epoch": 12.221556886227544, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1437, "step": 8164 }, { "epoch": 12.22305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1412, "step": 8165 }, { "epoch": 12.224550898203592, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1379, "step": 8166 }, { "epoch": 12.226047904191617, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1417, "step": 8167 }, { "epoch": 12.22754491017964, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1351, "step": 8168 }, { "epoch": 12.229041916167665, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.144, "step": 8169 }, { "epoch": 12.230538922155688, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1408, "step": 8170 }, { "epoch": 12.232035928143713, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8171 }, { "epoch": 12.233532934131736, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 8172 }, { "epoch": 12.23502994011976, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1433, "step": 8173 }, { "epoch": 12.236526946107784, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1459, "step": 8174 }, { "epoch": 12.238023952095809, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.143, "step": 8175 }, { "epoch": 12.239520958083832, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1383, "step": 8176 }, { "epoch": 12.241017964071856, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1396, "step": 8177 }, { "epoch": 12.24251497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1417, "step": 8178 }, { "epoch": 12.244011976047904, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1385, "step": 8179 }, { "epoch": 12.245508982035927, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.14, "step": 8180 }, { "epoch": 12.247005988023952, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1423, "step": 8181 }, { "epoch": 12.248502994011975, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1419, "step": 8182 }, { "epoch": 12.25, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 8183 }, { "epoch": 12.251497005988025, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.142, "step": 8184 }, { "epoch": 12.252994011976048, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.14, "step": 8185 }, { "epoch": 12.254491017964073, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1448, "step": 8186 }, { "epoch": 12.255988023952096, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.141, "step": 8187 }, { "epoch": 12.25748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1413, "step": 8188 }, { "epoch": 12.258982035928144, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1465, "step": 8189 }, { "epoch": 12.260479041916168, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1454, "step": 8190 }, { "epoch": 12.261976047904191, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1396, "step": 8191 }, { "epoch": 12.263473053892216, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1339, "step": 8192 }, { "epoch": 12.26497005988024, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1465, "step": 8193 }, { "epoch": 12.266467065868264, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1371, "step": 8194 }, { "epoch": 12.267964071856287, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1386, "step": 8195 }, { "epoch": 12.269461077844312, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1458, "step": 8196 }, { "epoch": 12.270958083832335, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.134, "step": 8197 }, { "epoch": 12.27245508982036, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.147, "step": 8198 }, { "epoch": 12.273952095808383, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1417, "step": 8199 }, { "epoch": 12.275449101796408, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.141, "step": 8200 }, { "epoch": 12.27694610778443, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1423, "step": 8201 }, { "epoch": 12.278443113772456, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1389, "step": 8202 }, { "epoch": 12.279940119760479, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1375, "step": 8203 }, { "epoch": 12.281437125748504, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1371, "step": 8204 }, { "epoch": 12.282934131736527, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1453, "step": 8205 }, { "epoch": 12.284431137724551, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1361, "step": 8206 }, { "epoch": 12.285928143712574, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1409, "step": 8207 }, { "epoch": 12.2874251497006, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1392, "step": 8208 }, { "epoch": 12.288922155688622, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1435, "step": 8209 }, { "epoch": 12.290419161676647, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.14, "step": 8210 }, { "epoch": 12.29191616766467, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1391, "step": 8211 }, { "epoch": 12.293413173652695, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1397, "step": 8212 }, { "epoch": 12.294910179640718, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1404, "step": 8213 }, { "epoch": 12.296407185628743, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1441, "step": 8214 }, { "epoch": 12.297904191616766, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1407, "step": 8215 }, { "epoch": 12.29940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1471, "step": 8216 }, { "epoch": 12.300898203592814, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1474, "step": 8217 }, { "epoch": 12.302395209580839, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1425, "step": 8218 }, { "epoch": 12.303892215568862, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1498, "step": 8219 }, { "epoch": 12.305389221556887, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 8220 }, { "epoch": 12.30688622754491, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1341, "step": 8221 }, { "epoch": 12.308383233532934, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1323, "step": 8222 }, { "epoch": 12.309880239520957, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1369, "step": 8223 }, { "epoch": 12.311377245508982, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.141, "step": 8224 }, { "epoch": 12.312874251497005, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1432, "step": 8225 }, { "epoch": 12.31437125748503, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8226 }, { "epoch": 12.315868263473053, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1408, "step": 8227 }, { "epoch": 12.317365269461078, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1437, "step": 8228 }, { "epoch": 12.318862275449101, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 8229 }, { "epoch": 12.320359281437126, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1364, "step": 8230 }, { "epoch": 12.321856287425149, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1458, "step": 8231 }, { "epoch": 12.323353293413174, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1365, "step": 8232 }, { "epoch": 12.324850299401197, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1436, "step": 8233 }, { "epoch": 12.326347305389222, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1397, "step": 8234 }, { "epoch": 12.327844311377245, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 8235 }, { "epoch": 12.32934131736527, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1392, "step": 8236 }, { "epoch": 12.330838323353294, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1486, "step": 8237 }, { "epoch": 12.332335329341317, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1474, "step": 8238 }, { "epoch": 12.33383233532934, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1375, "step": 8239 }, { "epoch": 12.335329341317365, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1441, "step": 8240 }, { "epoch": 12.33682634730539, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1388, "step": 8241 }, { "epoch": 12.338323353293413, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 8242 }, { "epoch": 12.339820359281438, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1423, "step": 8243 }, { "epoch": 12.341317365269461, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1449, "step": 8244 }, { "epoch": 12.342814371257486, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1374, "step": 8245 }, { "epoch": 12.344311377245509, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1383, "step": 8246 }, { "epoch": 12.345808383233534, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1396, "step": 8247 }, { "epoch": 12.347305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 8248 }, { "epoch": 12.348802395209582, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1414, "step": 8249 }, { "epoch": 12.350299401197605, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 8250 }, { "epoch": 12.35179640718563, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8251 }, { "epoch": 12.353293413173652, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 8252 }, { "epoch": 12.354790419161677, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1389, "step": 8253 }, { "epoch": 12.3562874251497, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8254 }, { "epoch": 12.357784431137725, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1453, "step": 8255 }, { "epoch": 12.359281437125748, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1398, "step": 8256 }, { "epoch": 12.360778443113773, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1445, "step": 8257 }, { "epoch": 12.362275449101796, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1496, "step": 8258 }, { "epoch": 12.363772455089821, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1383, "step": 8259 }, { "epoch": 12.365269461077844, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1399, "step": 8260 }, { "epoch": 12.366766467065869, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1447, "step": 8261 }, { "epoch": 12.368263473053892, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1375, "step": 8262 }, { "epoch": 12.369760479041917, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.136, "step": 8263 }, { "epoch": 12.37125748502994, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1403, "step": 8264 }, { "epoch": 12.372754491017965, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1358, "step": 8265 }, { "epoch": 12.374251497005988, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 8266 }, { "epoch": 12.375748502994012, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8267 }, { "epoch": 12.377245508982035, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1409, "step": 8268 }, { "epoch": 12.37874251497006, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1432, "step": 8269 }, { "epoch": 12.380239520958083, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.143, "step": 8270 }, { "epoch": 12.381736526946108, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1461, "step": 8271 }, { "epoch": 12.383233532934131, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1439, "step": 8272 }, { "epoch": 12.384730538922156, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1459, "step": 8273 }, { "epoch": 12.386227544910179, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1396, "step": 8274 }, { "epoch": 12.387724550898204, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.144, "step": 8275 }, { "epoch": 12.389221556886227, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1441, "step": 8276 }, { "epoch": 12.390718562874252, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1406, "step": 8277 }, { "epoch": 12.392215568862275, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1466, "step": 8278 }, { "epoch": 12.3937125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1457, "step": 8279 }, { "epoch": 12.395209580838323, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1391, "step": 8280 }, { "epoch": 12.396706586826348, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1491, "step": 8281 }, { "epoch": 12.39820359281437, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1406, "step": 8282 }, { "epoch": 12.399700598802395, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1492, "step": 8283 }, { "epoch": 12.401197604790418, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 8284 }, { "epoch": 12.402694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1387, "step": 8285 }, { "epoch": 12.404191616766466, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1444, "step": 8286 }, { "epoch": 12.405688622754491, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 8287 }, { "epoch": 12.407185628742514, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1447, "step": 8288 }, { "epoch": 12.408682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1449, "step": 8289 }, { "epoch": 12.410179640718562, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1455, "step": 8290 }, { "epoch": 12.411676646706587, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 8291 }, { "epoch": 12.41317365269461, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1369, "step": 8292 }, { "epoch": 12.414670658682635, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1372, "step": 8293 }, { "epoch": 12.41616766467066, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1466, "step": 8294 }, { "epoch": 12.417664670658683, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1427, "step": 8295 }, { "epoch": 12.419161676646706, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1499, "step": 8296 }, { "epoch": 12.42065868263473, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1431, "step": 8297 }, { "epoch": 12.422155688622755, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1434, "step": 8298 }, { "epoch": 12.423652694610778, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1402, "step": 8299 }, { "epoch": 12.425149700598803, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1438, "step": 8300 }, { "epoch": 12.426646706586826, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.138, "step": 8301 }, { "epoch": 12.428143712574851, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1423, "step": 8302 }, { "epoch": 12.429640718562874, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1403, "step": 8303 }, { "epoch": 12.431137724550899, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1427, "step": 8304 }, { "epoch": 12.432634730538922, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1422, "step": 8305 }, { "epoch": 12.434131736526947, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.142, "step": 8306 }, { "epoch": 12.43562874251497, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.139, "step": 8307 }, { "epoch": 12.437125748502995, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1449, "step": 8308 }, { "epoch": 12.438622754491018, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1351, "step": 8309 }, { "epoch": 12.440119760479043, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1507, "step": 8310 }, { "epoch": 12.441616766467066, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1382, "step": 8311 }, { "epoch": 12.44311377245509, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1448, "step": 8312 }, { "epoch": 12.444610778443113, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1435, "step": 8313 }, { "epoch": 12.446107784431138, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.139, "step": 8314 }, { "epoch": 12.447604790419161, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1368, "step": 8315 }, { "epoch": 12.449101796407186, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1424, "step": 8316 }, { "epoch": 12.45059880239521, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1432, "step": 8317 }, { "epoch": 12.452095808383234, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1417, "step": 8318 }, { "epoch": 12.453592814371257, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 8319 }, { "epoch": 12.455089820359282, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1402, "step": 8320 }, { "epoch": 12.456586826347305, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1431, "step": 8321 }, { "epoch": 12.45808383233533, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1352, "step": 8322 }, { "epoch": 12.459580838323353, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1392, "step": 8323 }, { "epoch": 12.461077844311378, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1405, "step": 8324 }, { "epoch": 12.4625748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1445, "step": 8325 }, { "epoch": 12.464071856287426, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1441, "step": 8326 }, { "epoch": 12.465568862275449, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1371, "step": 8327 }, { "epoch": 12.467065868263473, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1448, "step": 8328 }, { "epoch": 12.468562874251496, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1421, "step": 8329 }, { "epoch": 12.470059880239521, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1467, "step": 8330 }, { "epoch": 12.471556886227544, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8331 }, { "epoch": 12.47305389221557, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1396, "step": 8332 }, { "epoch": 12.474550898203592, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1396, "step": 8333 }, { "epoch": 12.476047904191617, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8334 }, { "epoch": 12.47754491017964, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1379, "step": 8335 }, { "epoch": 12.479041916167665, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1405, "step": 8336 }, { "epoch": 12.480538922155688, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1408, "step": 8337 }, { "epoch": 12.482035928143713, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1423, "step": 8338 }, { "epoch": 12.483532934131736, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1385, "step": 8339 }, { "epoch": 12.48502994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1435, "step": 8340 }, { "epoch": 12.486526946107784, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1441, "step": 8341 }, { "epoch": 12.488023952095809, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1377, "step": 8342 }, { "epoch": 12.489520958083832, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1337, "step": 8343 }, { "epoch": 12.491017964071856, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1413, "step": 8344 }, { "epoch": 12.49251497005988, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1463, "step": 8345 }, { "epoch": 12.494011976047904, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1472, "step": 8346 }, { "epoch": 12.495508982035927, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8347 }, { "epoch": 12.497005988023952, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 8348 }, { "epoch": 12.498502994011975, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1434, "step": 8349 }, { "epoch": 12.5, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1396, "step": 8350 }, { "epoch": 12.501497005988025, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.143, "step": 8351 }, { "epoch": 12.502994011976048, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1392, "step": 8352 }, { "epoch": 12.504491017964071, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 8353 }, { "epoch": 12.505988023952096, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1475, "step": 8354 }, { "epoch": 12.50748502994012, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1453, "step": 8355 }, { "epoch": 12.508982035928144, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1454, "step": 8356 }, { "epoch": 12.510479041916168, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.141, "step": 8357 }, { "epoch": 12.511976047904191, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1352, "step": 8358 }, { "epoch": 12.513473053892216, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.141, "step": 8359 }, { "epoch": 12.51497005988024, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.141, "step": 8360 }, { "epoch": 12.516467065868264, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1448, "step": 8361 }, { "epoch": 12.517964071856287, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1411, "step": 8362 }, { "epoch": 12.519461077844312, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1382, "step": 8363 }, { "epoch": 12.520958083832335, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1437, "step": 8364 }, { "epoch": 12.52245508982036, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.138, "step": 8365 }, { "epoch": 12.523952095808383, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1527, "step": 8366 }, { "epoch": 12.525449101796408, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1456, "step": 8367 }, { "epoch": 12.52694610778443, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1341, "step": 8368 }, { "epoch": 12.528443113772456, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1481, "step": 8369 }, { "epoch": 12.529940119760479, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1436, "step": 8370 }, { "epoch": 12.531437125748504, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1417, "step": 8371 }, { "epoch": 12.532934131736527, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1391, "step": 8372 }, { "epoch": 12.534431137724551, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1435, "step": 8373 }, { "epoch": 12.535928143712574, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1403, "step": 8374 }, { "epoch": 12.5374251497006, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1396, "step": 8375 }, { "epoch": 12.538922155688622, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1395, "step": 8376 }, { "epoch": 12.540419161676647, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1407, "step": 8377 }, { "epoch": 12.54191616766467, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1412, "step": 8378 }, { "epoch": 12.543413173652695, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1444, "step": 8379 }, { "epoch": 12.544910179640718, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.144, "step": 8380 }, { "epoch": 12.546407185628743, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 8381 }, { "epoch": 12.547904191616766, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.14, "step": 8382 }, { "epoch": 12.54940119760479, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8383 }, { "epoch": 12.550898203592814, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 8384 }, { "epoch": 12.552395209580839, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.135, "step": 8385 }, { "epoch": 12.553892215568862, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1402, "step": 8386 }, { "epoch": 12.555389221556887, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1476, "step": 8387 }, { "epoch": 12.55688622754491, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1391, "step": 8388 }, { "epoch": 12.558383233532934, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 8389 }, { "epoch": 12.559880239520957, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.145, "step": 8390 }, { "epoch": 12.561377245508982, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1455, "step": 8391 }, { "epoch": 12.562874251497005, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1433, "step": 8392 }, { "epoch": 12.56437125748503, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.143, "step": 8393 }, { "epoch": 12.565868263473053, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1375, "step": 8394 }, { "epoch": 12.567365269461078, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1389, "step": 8395 }, { "epoch": 12.568862275449101, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1468, "step": 8396 }, { "epoch": 12.570359281437126, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1454, "step": 8397 }, { "epoch": 12.571856287425149, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1437, "step": 8398 }, { "epoch": 12.573353293413174, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1445, "step": 8399 }, { "epoch": 12.574850299401197, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1412, "step": 8400 }, { "epoch": 12.576347305389222, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1451, "step": 8401 }, { "epoch": 12.577844311377245, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.141, "step": 8402 }, { "epoch": 12.57934131736527, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.146, "step": 8403 }, { "epoch": 12.580838323353294, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1471, "step": 8404 }, { "epoch": 12.582335329341317, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1366, "step": 8405 }, { "epoch": 12.58383233532934, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1405, "step": 8406 }, { "epoch": 12.585329341317365, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1442, "step": 8407 }, { "epoch": 12.58682634730539, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1509, "step": 8408 }, { "epoch": 12.588323353293413, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1415, "step": 8409 }, { "epoch": 12.589820359281438, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1466, "step": 8410 }, { "epoch": 12.591317365269461, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1434, "step": 8411 }, { "epoch": 12.592814371257486, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.138, "step": 8412 }, { "epoch": 12.594311377245509, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8413 }, { "epoch": 12.595808383233534, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1414, "step": 8414 }, { "epoch": 12.597305389221557, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1445, "step": 8415 }, { "epoch": 12.598802395209582, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.138, "step": 8416 }, { "epoch": 12.600299401197605, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1428, "step": 8417 }, { "epoch": 12.60179640718563, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 8418 }, { "epoch": 12.603293413173652, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1445, "step": 8419 }, { "epoch": 12.604790419161677, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 8420 }, { "epoch": 12.6062874251497, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8421 }, { "epoch": 12.607784431137725, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1454, "step": 8422 }, { "epoch": 12.609281437125748, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1411, "step": 8423 }, { "epoch": 12.610778443113773, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1463, "step": 8424 }, { "epoch": 12.612275449101796, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1443, "step": 8425 }, { "epoch": 12.613772455089821, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1424, "step": 8426 }, { "epoch": 12.615269461077844, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1458, "step": 8427 }, { "epoch": 12.616766467065869, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1453, "step": 8428 }, { "epoch": 12.618263473053892, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1413, "step": 8429 }, { "epoch": 12.619760479041917, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1446, "step": 8430 }, { "epoch": 12.62125748502994, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1389, "step": 8431 }, { "epoch": 12.622754491017965, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.144, "step": 8432 }, { "epoch": 12.624251497005988, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1363, "step": 8433 }, { "epoch": 12.625748502994012, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1362, "step": 8434 }, { "epoch": 12.627245508982035, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1474, "step": 8435 }, { "epoch": 12.62874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1376, "step": 8436 }, { "epoch": 12.630239520958083, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1344, "step": 8437 }, { "epoch": 12.631736526946108, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1365, "step": 8438 }, { "epoch": 12.633233532934131, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1435, "step": 8439 }, { "epoch": 12.634730538922156, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1356, "step": 8440 }, { "epoch": 12.636227544910179, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1417, "step": 8441 }, { "epoch": 12.637724550898204, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.142, "step": 8442 }, { "epoch": 12.639221556886227, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1395, "step": 8443 }, { "epoch": 12.640718562874252, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.134, "step": 8444 }, { "epoch": 12.642215568862275, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1432, "step": 8445 }, { "epoch": 12.6437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1367, "step": 8446 }, { "epoch": 12.645209580838323, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8447 }, { "epoch": 12.646706586826348, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1467, "step": 8448 }, { "epoch": 12.64820359281437, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.134, "step": 8449 }, { "epoch": 12.649700598802395, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1382, "step": 8450 }, { "epoch": 12.651197604790418, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1374, "step": 8451 }, { "epoch": 12.652694610778443, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1363, "step": 8452 }, { "epoch": 12.654191616766466, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1405, "step": 8453 }, { "epoch": 12.655688622754491, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1414, "step": 8454 }, { "epoch": 12.657185628742514, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1386, "step": 8455 }, { "epoch": 12.658682634730539, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1396, "step": 8456 }, { "epoch": 12.660179640718562, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1448, "step": 8457 }, { "epoch": 12.661676646706587, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1399, "step": 8458 }, { "epoch": 12.66317365269461, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1412, "step": 8459 }, { "epoch": 12.664670658682635, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1446, "step": 8460 }, { "epoch": 12.66616766467066, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1436, "step": 8461 }, { "epoch": 12.667664670658683, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1411, "step": 8462 }, { "epoch": 12.669161676646706, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.138, "step": 8463 }, { "epoch": 12.67065868263473, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1301, "step": 8464 }, { "epoch": 12.672155688622755, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1461, "step": 8465 }, { "epoch": 12.673652694610778, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1428, "step": 8466 }, { "epoch": 12.675149700598803, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1351, "step": 8467 }, { "epoch": 12.676646706586826, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1416, "step": 8468 }, { "epoch": 12.678143712574851, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1381, "step": 8469 }, { "epoch": 12.679640718562874, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1475, "step": 8470 }, { "epoch": 12.681137724550899, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1373, "step": 8471 }, { "epoch": 12.682634730538922, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1395, "step": 8472 }, { "epoch": 12.684131736526947, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1393, "step": 8473 }, { "epoch": 12.68562874251497, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1418, "step": 8474 }, { "epoch": 12.687125748502995, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1401, "step": 8475 }, { "epoch": 12.688622754491018, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8476 }, { "epoch": 12.690119760479043, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1452, "step": 8477 }, { "epoch": 12.691616766467066, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.14, "step": 8478 }, { "epoch": 12.69311377245509, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1327, "step": 8479 }, { "epoch": 12.694610778443113, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1388, "step": 8480 }, { "epoch": 12.696107784431138, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1448, "step": 8481 }, { "epoch": 12.697604790419161, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1366, "step": 8482 }, { "epoch": 12.699101796407186, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.139, "step": 8483 }, { "epoch": 12.70059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1427, "step": 8484 }, { "epoch": 12.702095808383234, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1476, "step": 8485 }, { "epoch": 12.703592814371257, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1465, "step": 8486 }, { "epoch": 12.705089820359282, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1419, "step": 8487 }, { "epoch": 12.706586826347305, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.141, "step": 8488 }, { "epoch": 12.70808383233533, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1466, "step": 8489 }, { "epoch": 12.709580838323353, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.145, "step": 8490 }, { "epoch": 12.711077844311378, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1448, "step": 8491 }, { "epoch": 12.7125748502994, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1408, "step": 8492 }, { "epoch": 12.714071856287426, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1419, "step": 8493 }, { "epoch": 12.715568862275449, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1346, "step": 8494 }, { "epoch": 12.717065868263473, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1464, "step": 8495 }, { "epoch": 12.718562874251496, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1357, "step": 8496 }, { "epoch": 12.720059880239521, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1502, "step": 8497 }, { "epoch": 12.721556886227544, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.142, "step": 8498 }, { "epoch": 12.72305389221557, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1436, "step": 8499 }, { "epoch": 12.724550898203592, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1461, "step": 8500 }, { "epoch": 12.726047904191617, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1442, "step": 8501 }, { "epoch": 12.72754491017964, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.143, "step": 8502 }, { "epoch": 12.729041916167665, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1386, "step": 8503 }, { "epoch": 12.730538922155688, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1308, "step": 8504 }, { "epoch": 12.732035928143713, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1455, "step": 8505 }, { "epoch": 12.733532934131736, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1411, "step": 8506 }, { "epoch": 12.73502994011976, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1439, "step": 8507 }, { "epoch": 12.736526946107784, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1347, "step": 8508 }, { "epoch": 12.738023952095809, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1437, "step": 8509 }, { "epoch": 12.739520958083832, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1415, "step": 8510 }, { "epoch": 12.741017964071856, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1406, "step": 8511 }, { "epoch": 12.74251497005988, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1391, "step": 8512 }, { "epoch": 12.744011976047904, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1401, "step": 8513 }, { "epoch": 12.745508982035929, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.138, "step": 8514 }, { "epoch": 12.747005988023952, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1447, "step": 8515 }, { "epoch": 12.748502994011975, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1455, "step": 8516 }, { "epoch": 12.75, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 8517 }, { "epoch": 12.751497005988025, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1437, "step": 8518 }, { "epoch": 12.752994011976048, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1361, "step": 8519 }, { "epoch": 12.754491017964071, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1411, "step": 8520 }, { "epoch": 12.755988023952096, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1402, "step": 8521 }, { "epoch": 12.75748502994012, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1385, "step": 8522 }, { "epoch": 12.758982035928144, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1383, "step": 8523 }, { "epoch": 12.760479041916168, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1355, "step": 8524 }, { "epoch": 12.761976047904191, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1316, "step": 8525 }, { "epoch": 12.763473053892216, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1385, "step": 8526 }, { "epoch": 12.76497005988024, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1481, "step": 8527 }, { "epoch": 12.766467065868264, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1398, "step": 8528 }, { "epoch": 12.767964071856287, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1421, "step": 8529 }, { "epoch": 12.769461077844312, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1417, "step": 8530 }, { "epoch": 12.770958083832335, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1402, "step": 8531 }, { "epoch": 12.77245508982036, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1391, "step": 8532 }, { "epoch": 12.773952095808383, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1433, "step": 8533 }, { "epoch": 12.775449101796408, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1412, "step": 8534 }, { "epoch": 12.77694610778443, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1463, "step": 8535 }, { "epoch": 12.778443113772456, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1458, "step": 8536 }, { "epoch": 12.779940119760479, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1471, "step": 8537 }, { "epoch": 12.781437125748504, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1428, "step": 8538 }, { "epoch": 12.782934131736527, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1376, "step": 8539 }, { "epoch": 12.784431137724551, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.145, "step": 8540 }, { "epoch": 12.785928143712574, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1417, "step": 8541 }, { "epoch": 12.7874251497006, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1463, "step": 8542 }, { "epoch": 12.788922155688622, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1441, "step": 8543 }, { "epoch": 12.790419161676647, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1438, "step": 8544 }, { "epoch": 12.79191616766467, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1455, "step": 8545 }, { "epoch": 12.793413173652695, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1404, "step": 8546 }, { "epoch": 12.794910179640718, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8547 }, { "epoch": 12.796407185628743, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1413, "step": 8548 }, { "epoch": 12.797904191616766, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1483, "step": 8549 }, { "epoch": 12.79940119760479, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1442, "step": 8550 }, { "epoch": 12.800898203592814, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8551 }, { "epoch": 12.802395209580839, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1399, "step": 8552 }, { "epoch": 12.803892215568862, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1382, "step": 8553 }, { "epoch": 12.805389221556887, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1365, "step": 8554 }, { "epoch": 12.80688622754491, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.137, "step": 8555 }, { "epoch": 12.808383233532934, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1467, "step": 8556 }, { "epoch": 12.809880239520957, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.144, "step": 8557 }, { "epoch": 12.811377245508982, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1455, "step": 8558 }, { "epoch": 12.812874251497005, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1424, "step": 8559 }, { "epoch": 12.81437125748503, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1362, "step": 8560 }, { "epoch": 12.815868263473053, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1438, "step": 8561 }, { "epoch": 12.817365269461078, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1424, "step": 8562 }, { "epoch": 12.818862275449101, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1433, "step": 8563 }, { "epoch": 12.820359281437126, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1336, "step": 8564 }, { "epoch": 12.821856287425149, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1424, "step": 8565 }, { "epoch": 12.823353293413174, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1461, "step": 8566 }, { "epoch": 12.824850299401197, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8567 }, { "epoch": 12.826347305389222, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1427, "step": 8568 }, { "epoch": 12.827844311377245, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1476, "step": 8569 }, { "epoch": 12.82934131736527, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1414, "step": 8570 }, { "epoch": 12.830838323353294, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 8571 }, { "epoch": 12.832335329341317, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1389, "step": 8572 }, { "epoch": 12.83383233532934, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1297, "step": 8573 }, { "epoch": 12.835329341317365, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1474, "step": 8574 }, { "epoch": 12.83682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1406, "step": 8575 }, { "epoch": 12.838323353293413, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1374, "step": 8576 }, { "epoch": 12.839820359281438, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1414, "step": 8577 }, { "epoch": 12.841317365269461, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1423, "step": 8578 }, { "epoch": 12.842814371257486, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1406, "step": 8579 }, { "epoch": 12.844311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1402, "step": 8580 }, { "epoch": 12.845808383233534, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1475, "step": 8581 }, { "epoch": 12.847305389221557, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1451, "step": 8582 }, { "epoch": 12.848802395209582, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1442, "step": 8583 }, { "epoch": 12.850299401197605, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1408, "step": 8584 }, { "epoch": 12.85179640718563, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1384, "step": 8585 }, { "epoch": 12.853293413173652, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8586 }, { "epoch": 12.854790419161677, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1399, "step": 8587 }, { "epoch": 12.8562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1486, "step": 8588 }, { "epoch": 12.857784431137725, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1376, "step": 8589 }, { "epoch": 12.859281437125748, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1443, "step": 8590 }, { "epoch": 12.860778443113773, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1394, "step": 8591 }, { "epoch": 12.862275449101796, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1389, "step": 8592 }, { "epoch": 12.863772455089821, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1445, "step": 8593 }, { "epoch": 12.865269461077844, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1446, "step": 8594 }, { "epoch": 12.866766467065869, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1384, "step": 8595 }, { "epoch": 12.868263473053892, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1416, "step": 8596 }, { "epoch": 12.869760479041917, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.14, "step": 8597 }, { "epoch": 12.87125748502994, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1387, "step": 8598 }, { "epoch": 12.872754491017965, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1439, "step": 8599 }, { "epoch": 12.874251497005988, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1384, "step": 8600 }, { "epoch": 12.875748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1417, "step": 8601 }, { "epoch": 12.877245508982035, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1374, "step": 8602 }, { "epoch": 12.87874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1381, "step": 8603 }, { "epoch": 12.880239520958083, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1453, "step": 8604 }, { "epoch": 12.881736526946108, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8605 }, { "epoch": 12.883233532934131, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1428, "step": 8606 }, { "epoch": 12.884730538922156, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1455, "step": 8607 }, { "epoch": 12.886227544910179, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1406, "step": 8608 }, { "epoch": 12.887724550898204, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1401, "step": 8609 }, { "epoch": 12.889221556886227, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1298, "step": 8610 }, { "epoch": 12.890718562874252, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1399, "step": 8611 }, { "epoch": 12.892215568862275, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.142, "step": 8612 }, { "epoch": 12.8937125748503, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 8613 }, { "epoch": 12.895209580838323, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1456, "step": 8614 }, { "epoch": 12.896706586826348, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 8615 }, { "epoch": 12.89820359281437, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8616 }, { "epoch": 12.899700598802395, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 8617 }, { "epoch": 12.901197604790418, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1396, "step": 8618 }, { "epoch": 12.902694610778443, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 8619 }, { "epoch": 12.904191616766466, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1366, "step": 8620 }, { "epoch": 12.905688622754491, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 8621 }, { "epoch": 12.907185628742514, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1394, "step": 8622 }, { "epoch": 12.908682634730539, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1425, "step": 8623 }, { "epoch": 12.910179640718562, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.143, "step": 8624 }, { "epoch": 12.911676646706587, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1388, "step": 8625 }, { "epoch": 12.91317365269461, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1402, "step": 8626 }, { "epoch": 12.914670658682635, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1302, "step": 8627 }, { "epoch": 12.91616766467066, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1358, "step": 8628 }, { "epoch": 12.917664670658683, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1443, "step": 8629 }, { "epoch": 12.919161676646706, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1429, "step": 8630 }, { "epoch": 12.92065868263473, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.145, "step": 8631 }, { "epoch": 12.922155688622755, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1471, "step": 8632 }, { "epoch": 12.923652694610778, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1492, "step": 8633 }, { "epoch": 12.925149700598803, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8634 }, { "epoch": 12.926646706586826, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1439, "step": 8635 }, { "epoch": 12.928143712574851, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1356, "step": 8636 }, { "epoch": 12.929640718562874, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1394, "step": 8637 }, { "epoch": 12.931137724550899, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1372, "step": 8638 }, { "epoch": 12.932634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.143, "step": 8639 }, { "epoch": 12.934131736526947, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 8640 }, { "epoch": 12.93562874251497, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1408, "step": 8641 }, { "epoch": 12.937125748502995, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1389, "step": 8642 }, { "epoch": 12.938622754491018, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1373, "step": 8643 }, { "epoch": 12.940119760479043, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 8644 }, { "epoch": 12.941616766467066, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1349, "step": 8645 }, { "epoch": 12.94311377245509, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1423, "step": 8646 }, { "epoch": 12.944610778443113, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1431, "step": 8647 }, { "epoch": 12.946107784431138, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1411, "step": 8648 }, { "epoch": 12.947604790419161, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.134, "step": 8649 }, { "epoch": 12.949101796407186, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1391, "step": 8650 }, { "epoch": 12.95059880239521, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1443, "step": 8651 }, { "epoch": 12.952095808383234, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1376, "step": 8652 }, { "epoch": 12.953592814371257, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1392, "step": 8653 }, { "epoch": 12.955089820359282, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1451, "step": 8654 }, { "epoch": 12.956586826347305, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1428, "step": 8655 }, { "epoch": 12.95808383233533, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1423, "step": 8656 }, { "epoch": 12.959580838323353, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1435, "step": 8657 }, { "epoch": 12.961077844311378, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1397, "step": 8658 }, { "epoch": 12.9625748502994, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.136, "step": 8659 }, { "epoch": 12.964071856287426, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1386, "step": 8660 }, { "epoch": 12.965568862275449, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1405, "step": 8661 }, { "epoch": 12.967065868263473, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1443, "step": 8662 }, { "epoch": 12.968562874251496, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1389, "step": 8663 }, { "epoch": 12.970059880239521, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1429, "step": 8664 }, { "epoch": 12.971556886227544, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8665 }, { "epoch": 12.97305389221557, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1421, "step": 8666 }, { "epoch": 12.974550898203592, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 8667 }, { "epoch": 12.976047904191617, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1372, "step": 8668 }, { "epoch": 12.97754491017964, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 8669 }, { "epoch": 12.979041916167665, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1406, "step": 8670 }, { "epoch": 12.980538922155688, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1436, "step": 8671 }, { "epoch": 12.982035928143713, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1382, "step": 8672 }, { "epoch": 12.983532934131736, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1442, "step": 8673 }, { "epoch": 12.98502994011976, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1443, "step": 8674 }, { "epoch": 12.986526946107784, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1416, "step": 8675 }, { "epoch": 12.988023952095809, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.143, "step": 8676 }, { "epoch": 12.989520958083832, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 8677 }, { "epoch": 12.991017964071856, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1388, "step": 8678 }, { "epoch": 12.99251497005988, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1426, "step": 8679 }, { "epoch": 12.994011976047904, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.146, "step": 8680 }, { "epoch": 12.995508982035929, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1354, "step": 8681 }, { "epoch": 12.997005988023952, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1413, "step": 8682 }, { "epoch": 12.998502994011975, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 8683 }, { "epoch": 13.0, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1391, "step": 8684 }, { "epoch": 13.001497005988025, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1427, "step": 8685 }, { "epoch": 13.002994011976048, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1372, "step": 8686 }, { "epoch": 13.004491017964073, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1372, "step": 8687 }, { "epoch": 13.005988023952096, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1409, "step": 8688 }, { "epoch": 13.00748502994012, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1393, "step": 8689 }, { "epoch": 13.008982035928144, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1348, "step": 8690 }, { "epoch": 13.010479041916168, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8691 }, { "epoch": 13.011976047904191, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1423, "step": 8692 }, { "epoch": 13.013473053892216, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1388, "step": 8693 }, { "epoch": 13.01497005988024, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1395, "step": 8694 }, { "epoch": 13.016467065868264, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.144, "step": 8695 }, { "epoch": 13.017964071856287, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1438, "step": 8696 }, { "epoch": 13.019461077844312, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.137, "step": 8697 }, { "epoch": 13.020958083832335, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1331, "step": 8698 }, { "epoch": 13.02245508982036, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1335, "step": 8699 }, { "epoch": 13.023952095808383, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1447, "step": 8700 }, { "epoch": 13.025449101796408, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 8701 }, { "epoch": 13.02694610778443, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.139, "step": 8702 }, { "epoch": 13.028443113772456, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1396, "step": 8703 }, { "epoch": 13.029940119760479, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1361, "step": 8704 }, { "epoch": 13.031437125748504, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1388, "step": 8705 }, { "epoch": 13.032934131736527, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1428, "step": 8706 }, { "epoch": 13.034431137724551, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1436, "step": 8707 }, { "epoch": 13.035928143712574, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8708 }, { "epoch": 13.0374251497006, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1415, "step": 8709 }, { "epoch": 13.038922155688622, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1344, "step": 8710 }, { "epoch": 13.040419161676647, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1349, "step": 8711 }, { "epoch": 13.04191616766467, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1392, "step": 8712 }, { "epoch": 13.043413173652695, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1412, "step": 8713 }, { "epoch": 13.044910179640718, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1436, "step": 8714 }, { "epoch": 13.046407185628743, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1358, "step": 8715 }, { "epoch": 13.047904191616766, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 8716 }, { "epoch": 13.04940119760479, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1382, "step": 8717 }, { "epoch": 13.050898203592814, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1408, "step": 8718 }, { "epoch": 13.052395209580839, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1407, "step": 8719 }, { "epoch": 13.053892215568862, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1442, "step": 8720 }, { "epoch": 13.055389221556887, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8721 }, { "epoch": 13.05688622754491, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1354, "step": 8722 }, { "epoch": 13.058383233532934, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8723 }, { "epoch": 13.059880239520957, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.14, "step": 8724 }, { "epoch": 13.061377245508982, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.145, "step": 8725 }, { "epoch": 13.062874251497005, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1421, "step": 8726 }, { "epoch": 13.06437125748503, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1463, "step": 8727 }, { "epoch": 13.065868263473053, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1319, "step": 8728 }, { "epoch": 13.067365269461078, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8729 }, { "epoch": 13.068862275449101, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.133, "step": 8730 }, { "epoch": 13.070359281437126, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 8731 }, { "epoch": 13.071856287425149, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 8732 }, { "epoch": 13.073353293413174, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1388, "step": 8733 }, { "epoch": 13.074850299401197, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1469, "step": 8734 }, { "epoch": 13.076347305389222, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1368, "step": 8735 }, { "epoch": 13.077844311377245, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1459, "step": 8736 }, { "epoch": 13.07934131736527, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.142, "step": 8737 }, { "epoch": 13.080838323353293, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1417, "step": 8738 }, { "epoch": 13.082335329341317, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 8739 }, { "epoch": 13.08383233532934, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1359, "step": 8740 }, { "epoch": 13.085329341317365, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1422, "step": 8741 }, { "epoch": 13.08682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1406, "step": 8742 }, { "epoch": 13.088323353293413, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1382, "step": 8743 }, { "epoch": 13.089820359281438, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1471, "step": 8744 }, { "epoch": 13.091317365269461, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1358, "step": 8745 }, { "epoch": 13.092814371257486, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1419, "step": 8746 }, { "epoch": 13.094311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1361, "step": 8747 }, { "epoch": 13.095808383233534, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1357, "step": 8748 }, { "epoch": 13.097305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1378, "step": 8749 }, { "epoch": 13.098802395209582, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 8750 }, { "epoch": 13.100299401197605, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1357, "step": 8751 }, { "epoch": 13.10179640718563, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1464, "step": 8752 }, { "epoch": 13.103293413173652, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1409, "step": 8753 }, { "epoch": 13.104790419161677, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1351, "step": 8754 }, { "epoch": 13.1062874251497, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1355, "step": 8755 }, { "epoch": 13.107784431137725, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1343, "step": 8756 }, { "epoch": 13.109281437125748, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 8757 }, { "epoch": 13.110778443113773, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1374, "step": 8758 }, { "epoch": 13.112275449101796, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1476, "step": 8759 }, { "epoch": 13.113772455089821, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1398, "step": 8760 }, { "epoch": 13.115269461077844, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1417, "step": 8761 }, { "epoch": 13.116766467065869, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1419, "step": 8762 }, { "epoch": 13.118263473053892, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1484, "step": 8763 }, { "epoch": 13.119760479041917, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1444, "step": 8764 }, { "epoch": 13.12125748502994, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1443, "step": 8765 }, { "epoch": 13.122754491017965, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1401, "step": 8766 }, { "epoch": 13.124251497005988, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1385, "step": 8767 }, { "epoch": 13.125748502994012, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8768 }, { "epoch": 13.127245508982035, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1378, "step": 8769 }, { "epoch": 13.12874251497006, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 8770 }, { "epoch": 13.130239520958083, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1411, "step": 8771 }, { "epoch": 13.131736526946108, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.135, "step": 8772 }, { "epoch": 13.133233532934131, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1408, "step": 8773 }, { "epoch": 13.134730538922156, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1457, "step": 8774 }, { "epoch": 13.136227544910179, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1448, "step": 8775 }, { "epoch": 13.137724550898204, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1364, "step": 8776 }, { "epoch": 13.139221556886227, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1342, "step": 8777 }, { "epoch": 13.140718562874252, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1362, "step": 8778 }, { "epoch": 13.142215568862275, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1402, "step": 8779 }, { "epoch": 13.1437125748503, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1357, "step": 8780 }, { "epoch": 13.145209580838323, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1392, "step": 8781 }, { "epoch": 13.146706586826348, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1464, "step": 8782 }, { "epoch": 13.14820359281437, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1457, "step": 8783 }, { "epoch": 13.149700598802395, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.136, "step": 8784 }, { "epoch": 13.151197604790418, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1413, "step": 8785 }, { "epoch": 13.152694610778443, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1405, "step": 8786 }, { "epoch": 13.154191616766466, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1366, "step": 8787 }, { "epoch": 13.155688622754491, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.141, "step": 8788 }, { "epoch": 13.157185628742514, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1407, "step": 8789 }, { "epoch": 13.158682634730539, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.144, "step": 8790 }, { "epoch": 13.160179640718562, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.135, "step": 8791 }, { "epoch": 13.161676646706587, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1398, "step": 8792 }, { "epoch": 13.16317365269461, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1454, "step": 8793 }, { "epoch": 13.164670658682635, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1347, "step": 8794 }, { "epoch": 13.16616766467066, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1439, "step": 8795 }, { "epoch": 13.167664670658683, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1399, "step": 8796 }, { "epoch": 13.169161676646707, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1397, "step": 8797 }, { "epoch": 13.17065868263473, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1344, "step": 8798 }, { "epoch": 13.172155688622755, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1415, "step": 8799 }, { "epoch": 13.173652694610778, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1421, "step": 8800 }, { "epoch": 13.175149700598803, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1411, "step": 8801 }, { "epoch": 13.176646706586826, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1387, "step": 8802 }, { "epoch": 13.178143712574851, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1412, "step": 8803 }, { "epoch": 13.179640718562874, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8804 }, { "epoch": 13.181137724550899, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1348, "step": 8805 }, { "epoch": 13.182634730538922, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1387, "step": 8806 }, { "epoch": 13.184131736526947, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1452, "step": 8807 }, { "epoch": 13.18562874251497, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.139, "step": 8808 }, { "epoch": 13.187125748502995, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1376, "step": 8809 }, { "epoch": 13.188622754491018, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1389, "step": 8810 }, { "epoch": 13.190119760479043, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1389, "step": 8811 }, { "epoch": 13.191616766467066, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 8812 }, { "epoch": 13.19311377245509, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1359, "step": 8813 }, { "epoch": 13.194610778443113, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 8814 }, { "epoch": 13.196107784431138, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1413, "step": 8815 }, { "epoch": 13.197604790419161, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1405, "step": 8816 }, { "epoch": 13.199101796407186, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.141, "step": 8817 }, { "epoch": 13.20059880239521, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1407, "step": 8818 }, { "epoch": 13.202095808383234, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1378, "step": 8819 }, { "epoch": 13.203592814371257, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 8820 }, { "epoch": 13.205089820359282, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.14, "step": 8821 }, { "epoch": 13.206586826347305, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1367, "step": 8822 }, { "epoch": 13.20808383233533, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1454, "step": 8823 }, { "epoch": 13.209580838323353, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1434, "step": 8824 }, { "epoch": 13.211077844311378, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 8825 }, { "epoch": 13.2125748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.142, "step": 8826 }, { "epoch": 13.214071856287426, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1437, "step": 8827 }, { "epoch": 13.215568862275449, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1345, "step": 8828 }, { "epoch": 13.217065868263473, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1483, "step": 8829 }, { "epoch": 13.218562874251496, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8830 }, { "epoch": 13.220059880239521, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1336, "step": 8831 }, { "epoch": 13.221556886227544, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.143, "step": 8832 }, { "epoch": 13.22305389221557, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1414, "step": 8833 }, { "epoch": 13.224550898203592, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1399, "step": 8834 }, { "epoch": 13.226047904191617, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1387, "step": 8835 }, { "epoch": 13.22754491017964, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1398, "step": 8836 }, { "epoch": 13.229041916167665, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1478, "step": 8837 }, { "epoch": 13.230538922155688, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.14, "step": 8838 }, { "epoch": 13.232035928143713, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.145, "step": 8839 }, { "epoch": 13.233532934131736, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1383, "step": 8840 }, { "epoch": 13.23502994011976, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 8841 }, { "epoch": 13.236526946107784, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8842 }, { "epoch": 13.238023952095809, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.139, "step": 8843 }, { "epoch": 13.239520958083832, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1393, "step": 8844 }, { "epoch": 13.241017964071856, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1376, "step": 8845 }, { "epoch": 13.24251497005988, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1386, "step": 8846 }, { "epoch": 13.244011976047904, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1391, "step": 8847 }, { "epoch": 13.245508982035927, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1481, "step": 8848 }, { "epoch": 13.247005988023952, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1423, "step": 8849 }, { "epoch": 13.248502994011975, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1423, "step": 8850 }, { "epoch": 13.25, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1403, "step": 8851 }, { "epoch": 13.251497005988025, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1409, "step": 8852 }, { "epoch": 13.252994011976048, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1323, "step": 8853 }, { "epoch": 13.254491017964073, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1398, "step": 8854 }, { "epoch": 13.255988023952096, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 8855 }, { "epoch": 13.25748502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1433, "step": 8856 }, { "epoch": 13.258982035928144, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1388, "step": 8857 }, { "epoch": 13.260479041916168, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1437, "step": 8858 }, { "epoch": 13.261976047904191, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 8859 }, { "epoch": 13.263473053892216, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1389, "step": 8860 }, { "epoch": 13.26497005988024, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1354, "step": 8861 }, { "epoch": 13.266467065868264, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1382, "step": 8862 }, { "epoch": 13.267964071856287, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1429, "step": 8863 }, { "epoch": 13.269461077844312, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.135, "step": 8864 }, { "epoch": 13.270958083832335, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1388, "step": 8865 }, { "epoch": 13.27245508982036, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1474, "step": 8866 }, { "epoch": 13.273952095808383, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.143, "step": 8867 }, { "epoch": 13.275449101796408, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1478, "step": 8868 }, { "epoch": 13.27694610778443, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.134, "step": 8869 }, { "epoch": 13.278443113772456, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1431, "step": 8870 }, { "epoch": 13.279940119760479, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.141, "step": 8871 }, { "epoch": 13.281437125748504, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1327, "step": 8872 }, { "epoch": 13.282934131736527, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.132, "step": 8873 }, { "epoch": 13.284431137724551, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1346, "step": 8874 }, { "epoch": 13.285928143712574, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.136, "step": 8875 }, { "epoch": 13.2874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1342, "step": 8876 }, { "epoch": 13.288922155688622, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.134, "step": 8877 }, { "epoch": 13.290419161676647, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1415, "step": 8878 }, { "epoch": 13.29191616766467, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1398, "step": 8879 }, { "epoch": 13.293413173652695, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.141, "step": 8880 }, { "epoch": 13.294910179640718, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1385, "step": 8881 }, { "epoch": 13.296407185628743, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1432, "step": 8882 }, { "epoch": 13.297904191616766, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1372, "step": 8883 }, { "epoch": 13.29940119760479, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1414, "step": 8884 }, { "epoch": 13.300898203592814, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1335, "step": 8885 }, { "epoch": 13.302395209580839, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1425, "step": 8886 }, { "epoch": 13.303892215568862, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.146, "step": 8887 }, { "epoch": 13.305389221556887, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8888 }, { "epoch": 13.30688622754491, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 8889 }, { "epoch": 13.308383233532934, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1421, "step": 8890 }, { "epoch": 13.309880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1487, "step": 8891 }, { "epoch": 13.311377245508982, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1403, "step": 8892 }, { "epoch": 13.312874251497005, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1414, "step": 8893 }, { "epoch": 13.31437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1319, "step": 8894 }, { "epoch": 13.315868263473053, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 8895 }, { "epoch": 13.317365269461078, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1409, "step": 8896 }, { "epoch": 13.318862275449101, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1371, "step": 8897 }, { "epoch": 13.320359281437126, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1472, "step": 8898 }, { "epoch": 13.321856287425149, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8899 }, { "epoch": 13.323353293413174, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.135, "step": 8900 }, { "epoch": 13.324850299401197, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1419, "step": 8901 }, { "epoch": 13.326347305389222, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1421, "step": 8902 }, { "epoch": 13.327844311377245, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1417, "step": 8903 }, { "epoch": 13.32934131736527, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1367, "step": 8904 }, { "epoch": 13.330838323353294, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1405, "step": 8905 }, { "epoch": 13.332335329341317, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.138, "step": 8906 }, { "epoch": 13.33383233532934, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1362, "step": 8907 }, { "epoch": 13.335329341317365, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1487, "step": 8908 }, { "epoch": 13.33682634730539, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.139, "step": 8909 }, { "epoch": 13.338323353293413, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1422, "step": 8910 }, { "epoch": 13.339820359281438, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8911 }, { "epoch": 13.341317365269461, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.143, "step": 8912 }, { "epoch": 13.342814371257486, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1378, "step": 8913 }, { "epoch": 13.344311377245509, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1399, "step": 8914 }, { "epoch": 13.345808383233534, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1362, "step": 8915 }, { "epoch": 13.347305389221557, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1395, "step": 8916 }, { "epoch": 13.348802395209582, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1458, "step": 8917 }, { "epoch": 13.350299401197605, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1423, "step": 8918 }, { "epoch": 13.35179640718563, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1411, "step": 8919 }, { "epoch": 13.353293413173652, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1411, "step": 8920 }, { "epoch": 13.354790419161677, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1462, "step": 8921 }, { "epoch": 13.3562874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.141, "step": 8922 }, { "epoch": 13.357784431137725, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1359, "step": 8923 }, { "epoch": 13.359281437125748, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1364, "step": 8924 }, { "epoch": 13.360778443113773, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1318, "step": 8925 }, { "epoch": 13.362275449101796, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1377, "step": 8926 }, { "epoch": 13.363772455089821, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1357, "step": 8927 }, { "epoch": 13.365269461077844, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1399, "step": 8928 }, { "epoch": 13.366766467065869, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1458, "step": 8929 }, { "epoch": 13.368263473053892, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1484, "step": 8930 }, { "epoch": 13.369760479041917, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1444, "step": 8931 }, { "epoch": 13.37125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1393, "step": 8932 }, { "epoch": 13.372754491017965, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1404, "step": 8933 }, { "epoch": 13.374251497005988, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1364, "step": 8934 }, { "epoch": 13.375748502994012, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1399, "step": 8935 }, { "epoch": 13.377245508982035, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1353, "step": 8936 }, { "epoch": 13.37874251497006, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1434, "step": 8937 }, { "epoch": 13.380239520958083, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1325, "step": 8938 }, { "epoch": 13.381736526946108, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1466, "step": 8939 }, { "epoch": 13.383233532934131, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1382, "step": 8940 }, { "epoch": 13.384730538922156, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1383, "step": 8941 }, { "epoch": 13.386227544910179, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1404, "step": 8942 }, { "epoch": 13.387724550898204, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1375, "step": 8943 }, { "epoch": 13.389221556886227, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1373, "step": 8944 }, { "epoch": 13.390718562874252, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1419, "step": 8945 }, { "epoch": 13.392215568862275, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1397, "step": 8946 }, { "epoch": 13.3937125748503, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 8947 }, { "epoch": 13.395209580838323, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1376, "step": 8948 }, { "epoch": 13.396706586826348, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1434, "step": 8949 }, { "epoch": 13.39820359281437, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1439, "step": 8950 }, { "epoch": 13.399700598802395, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1467, "step": 8951 }, { "epoch": 13.401197604790418, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8952 }, { "epoch": 13.402694610778443, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.147, "step": 8953 }, { "epoch": 13.404191616766466, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1434, "step": 8954 }, { "epoch": 13.405688622754491, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.139, "step": 8955 }, { "epoch": 13.407185628742514, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 8956 }, { "epoch": 13.408682634730539, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1311, "step": 8957 }, { "epoch": 13.410179640718562, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1445, "step": 8958 }, { "epoch": 13.411676646706587, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 8959 }, { "epoch": 13.41317365269461, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1447, "step": 8960 }, { "epoch": 13.414670658682635, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1388, "step": 8961 }, { "epoch": 13.41616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1424, "step": 8962 }, { "epoch": 13.417664670658683, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1366, "step": 8963 }, { "epoch": 13.419161676646706, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.131, "step": 8964 }, { "epoch": 13.42065868263473, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.138, "step": 8965 }, { "epoch": 13.422155688622755, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1377, "step": 8966 }, { "epoch": 13.423652694610778, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1436, "step": 8967 }, { "epoch": 13.425149700598803, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1352, "step": 8968 }, { "epoch": 13.426646706586826, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1355, "step": 8969 }, { "epoch": 13.428143712574851, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1357, "step": 8970 }, { "epoch": 13.429640718562874, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.135, "step": 8971 }, { "epoch": 13.431137724550899, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1388, "step": 8972 }, { "epoch": 13.432634730538922, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1444, "step": 8973 }, { "epoch": 13.434131736526947, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1372, "step": 8974 }, { "epoch": 13.43562874251497, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1467, "step": 8975 }, { "epoch": 13.437125748502995, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1367, "step": 8976 }, { "epoch": 13.438622754491018, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1357, "step": 8977 }, { "epoch": 13.440119760479043, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1374, "step": 8978 }, { "epoch": 13.441616766467066, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 8979 }, { "epoch": 13.44311377245509, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1401, "step": 8980 }, { "epoch": 13.444610778443113, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1377, "step": 8981 }, { "epoch": 13.446107784431138, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1372, "step": 8982 }, { "epoch": 13.447604790419161, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1404, "step": 8983 }, { "epoch": 13.449101796407186, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1363, "step": 8984 }, { "epoch": 13.45059880239521, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1459, "step": 8985 }, { "epoch": 13.452095808383234, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1397, "step": 8986 }, { "epoch": 13.453592814371257, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1467, "step": 8987 }, { "epoch": 13.455089820359282, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1404, "step": 8988 }, { "epoch": 13.456586826347305, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1438, "step": 8989 }, { "epoch": 13.45808383233533, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.144, "step": 8990 }, { "epoch": 13.459580838323353, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1395, "step": 8991 }, { "epoch": 13.461077844311378, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1374, "step": 8992 }, { "epoch": 13.4625748502994, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1432, "step": 8993 }, { "epoch": 13.464071856287426, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1448, "step": 8994 }, { "epoch": 13.465568862275449, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1426, "step": 8995 }, { "epoch": 13.467065868263473, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1377, "step": 8996 }, { "epoch": 13.468562874251496, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1382, "step": 8997 }, { "epoch": 13.470059880239521, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1382, "step": 8998 }, { "epoch": 13.471556886227544, "grad_norm": 0.048583984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 8999 }, { "epoch": 13.47305389221557, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1345, "step": 9000 }, { "epoch": 13.474550898203592, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.142, "step": 9001 }, { "epoch": 13.476047904191617, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1375, "step": 9002 }, { "epoch": 13.47754491017964, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 9003 }, { "epoch": 13.479041916167665, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1427, "step": 9004 }, { "epoch": 13.480538922155688, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1382, "step": 9005 }, { "epoch": 13.482035928143713, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1405, "step": 9006 }, { "epoch": 13.483532934131736, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 9007 }, { "epoch": 13.48502994011976, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1382, "step": 9008 }, { "epoch": 13.486526946107784, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 9009 }, { "epoch": 13.488023952095809, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1412, "step": 9010 }, { "epoch": 13.489520958083832, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1413, "step": 9011 }, { "epoch": 13.491017964071856, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1409, "step": 9012 }, { "epoch": 13.49251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1424, "step": 9013 }, { "epoch": 13.494011976047904, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1367, "step": 9014 }, { "epoch": 13.495508982035927, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1448, "step": 9015 }, { "epoch": 13.497005988023952, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1413, "step": 9016 }, { "epoch": 13.498502994011975, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1353, "step": 9017 }, { "epoch": 13.5, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1412, "step": 9018 }, { "epoch": 13.501497005988025, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1346, "step": 9019 }, { "epoch": 13.502994011976048, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1366, "step": 9020 }, { "epoch": 13.504491017964071, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1426, "step": 9021 }, { "epoch": 13.505988023952096, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1357, "step": 9022 }, { "epoch": 13.50748502994012, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1359, "step": 9023 }, { "epoch": 13.508982035928144, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1344, "step": 9024 }, { "epoch": 13.510479041916168, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1389, "step": 9025 }, { "epoch": 13.511976047904191, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 9026 }, { "epoch": 13.513473053892216, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1412, "step": 9027 }, { "epoch": 13.51497005988024, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.145, "step": 9028 }, { "epoch": 13.516467065868264, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1372, "step": 9029 }, { "epoch": 13.517964071856287, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1379, "step": 9030 }, { "epoch": 13.519461077844312, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 9031 }, { "epoch": 13.520958083832335, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1395, "step": 9032 }, { "epoch": 13.52245508982036, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1416, "step": 9033 }, { "epoch": 13.523952095808383, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1431, "step": 9034 }, { "epoch": 13.525449101796408, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1433, "step": 9035 }, { "epoch": 13.52694610778443, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1373, "step": 9036 }, { "epoch": 13.528443113772456, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 9037 }, { "epoch": 13.529940119760479, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1388, "step": 9038 }, { "epoch": 13.531437125748504, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1408, "step": 9039 }, { "epoch": 13.532934131736527, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1405, "step": 9040 }, { "epoch": 13.534431137724551, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1342, "step": 9041 }, { "epoch": 13.535928143712574, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1409, "step": 9042 }, { "epoch": 13.5374251497006, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.147, "step": 9043 }, { "epoch": 13.538922155688622, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1399, "step": 9044 }, { "epoch": 13.540419161676647, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1449, "step": 9045 }, { "epoch": 13.54191616766467, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1365, "step": 9046 }, { "epoch": 13.543413173652695, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1342, "step": 9047 }, { "epoch": 13.544910179640718, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1391, "step": 9048 }, { "epoch": 13.546407185628743, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1349, "step": 9049 }, { "epoch": 13.547904191616766, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1386, "step": 9050 }, { "epoch": 13.54940119760479, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1356, "step": 9051 }, { "epoch": 13.550898203592814, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1397, "step": 9052 }, { "epoch": 13.552395209580839, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1293, "step": 9053 }, { "epoch": 13.553892215568862, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1358, "step": 9054 }, { "epoch": 13.555389221556887, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1461, "step": 9055 }, { "epoch": 13.55688622754491, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9056 }, { "epoch": 13.558383233532934, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.141, "step": 9057 }, { "epoch": 13.559880239520957, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1443, "step": 9058 }, { "epoch": 13.561377245508982, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1374, "step": 9059 }, { "epoch": 13.562874251497005, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 9060 }, { "epoch": 13.56437125748503, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1366, "step": 9061 }, { "epoch": 13.565868263473053, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1419, "step": 9062 }, { "epoch": 13.567365269461078, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1416, "step": 9063 }, { "epoch": 13.568862275449101, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1385, "step": 9064 }, { "epoch": 13.570359281437126, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1344, "step": 9065 }, { "epoch": 13.571856287425149, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1365, "step": 9066 }, { "epoch": 13.573353293413174, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 9067 }, { "epoch": 13.574850299401197, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1364, "step": 9068 }, { "epoch": 13.576347305389222, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1373, "step": 9069 }, { "epoch": 13.577844311377245, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1343, "step": 9070 }, { "epoch": 13.57934131736527, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1374, "step": 9071 }, { "epoch": 13.580838323353294, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1323, "step": 9072 }, { "epoch": 13.582335329341317, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 9073 }, { "epoch": 13.58383233532934, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1356, "step": 9074 }, { "epoch": 13.585329341317365, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1368, "step": 9075 }, { "epoch": 13.58682634730539, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1382, "step": 9076 }, { "epoch": 13.588323353293413, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1335, "step": 9077 }, { "epoch": 13.589820359281438, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1403, "step": 9078 }, { "epoch": 13.591317365269461, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1423, "step": 9079 }, { "epoch": 13.592814371257486, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1428, "step": 9080 }, { "epoch": 13.594311377245509, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1397, "step": 9081 }, { "epoch": 13.595808383233534, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1374, "step": 9082 }, { "epoch": 13.597305389221557, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9083 }, { "epoch": 13.598802395209582, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1389, "step": 9084 }, { "epoch": 13.600299401197605, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1442, "step": 9085 }, { "epoch": 13.60179640718563, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1447, "step": 9086 }, { "epoch": 13.603293413173652, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.13, "step": 9087 }, { "epoch": 13.604790419161677, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1404, "step": 9088 }, { "epoch": 13.6062874251497, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1377, "step": 9089 }, { "epoch": 13.607784431137725, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1353, "step": 9090 }, { "epoch": 13.609281437125748, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9091 }, { "epoch": 13.610778443113773, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1405, "step": 9092 }, { "epoch": 13.612275449101796, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1389, "step": 9093 }, { "epoch": 13.613772455089821, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1392, "step": 9094 }, { "epoch": 13.615269461077844, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1427, "step": 9095 }, { "epoch": 13.616766467065869, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1312, "step": 9096 }, { "epoch": 13.618263473053892, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1372, "step": 9097 }, { "epoch": 13.619760479041917, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1424, "step": 9098 }, { "epoch": 13.62125748502994, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1381, "step": 9099 }, { "epoch": 13.622754491017965, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1408, "step": 9100 }, { "epoch": 13.624251497005988, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1409, "step": 9101 }, { "epoch": 13.625748502994012, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1415, "step": 9102 }, { "epoch": 13.627245508982035, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1378, "step": 9103 }, { "epoch": 13.62874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 9104 }, { "epoch": 13.630239520958083, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1357, "step": 9105 }, { "epoch": 13.631736526946108, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1424, "step": 9106 }, { "epoch": 13.633233532934131, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9107 }, { "epoch": 13.634730538922156, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1405, "step": 9108 }, { "epoch": 13.636227544910179, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1417, "step": 9109 }, { "epoch": 13.637724550898204, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 9110 }, { "epoch": 13.639221556886227, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1328, "step": 9111 }, { "epoch": 13.640718562874252, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1367, "step": 9112 }, { "epoch": 13.642215568862275, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.137, "step": 9113 }, { "epoch": 13.6437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1384, "step": 9114 }, { "epoch": 13.645209580838323, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1402, "step": 9115 }, { "epoch": 13.646706586826348, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1428, "step": 9116 }, { "epoch": 13.64820359281437, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1361, "step": 9117 }, { "epoch": 13.649700598802395, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1413, "step": 9118 }, { "epoch": 13.651197604790418, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.137, "step": 9119 }, { "epoch": 13.652694610778443, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1358, "step": 9120 }, { "epoch": 13.654191616766466, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.143, "step": 9121 }, { "epoch": 13.655688622754491, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 9122 }, { "epoch": 13.657185628742514, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 9123 }, { "epoch": 13.658682634730539, "grad_norm": 0.1923828125, "learning_rate": 0.0008, "loss": 1.1396, "step": 9124 }, { "epoch": 13.660179640718562, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1354, "step": 9125 }, { "epoch": 13.661676646706587, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.138, "step": 9126 }, { "epoch": 13.66317365269461, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1403, "step": 9127 }, { "epoch": 13.664670658682635, "grad_norm": 0.2177734375, "learning_rate": 0.0008, "loss": 1.1355, "step": 9128 }, { "epoch": 13.66616766467066, "grad_norm": 0.310546875, "learning_rate": 0.0008, "loss": 1.1381, "step": 9129 }, { "epoch": 13.667664670658683, "grad_norm": 0.451171875, "learning_rate": 0.0008, "loss": 1.1491, "step": 9130 }, { "epoch": 13.669161676646706, "grad_norm": 0.73828125, "learning_rate": 0.0008, "loss": 1.152, "step": 9131 }, { "epoch": 13.67065868263473, "grad_norm": 1.3984375, "learning_rate": 0.0008, "loss": 1.1754, "step": 9132 }, { "epoch": 13.672155688622755, "grad_norm": 0.56640625, "learning_rate": 0.0008, "loss": 1.155, "step": 9133 }, { "epoch": 13.673652694610778, "grad_norm": 0.37109375, "learning_rate": 0.0008, "loss": 1.1507, "step": 9134 }, { "epoch": 13.675149700598803, "grad_norm": 0.5859375, "learning_rate": 0.0008, "loss": 1.1575, "step": 9135 }, { "epoch": 13.676646706586826, "grad_norm": 0.478515625, "learning_rate": 0.0008, "loss": 1.1395, "step": 9136 }, { "epoch": 13.678143712574851, "grad_norm": 0.447265625, "learning_rate": 0.0008, "loss": 1.1566, "step": 9137 }, { "epoch": 13.679640718562874, "grad_norm": 0.1787109375, "learning_rate": 0.0008, "loss": 1.1481, "step": 9138 }, { "epoch": 13.681137724550899, "grad_norm": 0.2890625, "learning_rate": 0.0008, "loss": 1.1555, "step": 9139 }, { "epoch": 13.682634730538922, "grad_norm": 0.25, "learning_rate": 0.0008, "loss": 1.155, "step": 9140 }, { "epoch": 13.684131736526947, "grad_norm": 0.1650390625, "learning_rate": 0.0008, "loss": 1.1502, "step": 9141 }, { "epoch": 13.68562874251497, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.1448, "step": 9142 }, { "epoch": 13.687125748502995, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1403, "step": 9143 }, { "epoch": 13.688622754491018, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1551, "step": 9144 }, { "epoch": 13.690119760479043, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1438, "step": 9145 }, { "epoch": 13.691616766467066, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1416, "step": 9146 }, { "epoch": 13.69311377245509, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1453, "step": 9147 }, { "epoch": 13.694610778443113, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.148, "step": 9148 }, { "epoch": 13.696107784431138, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1403, "step": 9149 }, { "epoch": 13.697604790419161, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1429, "step": 9150 }, { "epoch": 13.699101796407186, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1487, "step": 9151 }, { "epoch": 13.70059880239521, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1416, "step": 9152 }, { "epoch": 13.702095808383234, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1483, "step": 9153 }, { "epoch": 13.703592814371257, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1328, "step": 9154 }, { "epoch": 13.705089820359282, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 9155 }, { "epoch": 13.706586826347305, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1447, "step": 9156 }, { "epoch": 13.70808383233533, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1471, "step": 9157 }, { "epoch": 13.709580838323353, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1443, "step": 9158 }, { "epoch": 13.711077844311378, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1417, "step": 9159 }, { "epoch": 13.7125748502994, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1388, "step": 9160 }, { "epoch": 13.714071856287426, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9161 }, { "epoch": 13.715568862275449, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1391, "step": 9162 }, { "epoch": 13.717065868263473, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1406, "step": 9163 }, { "epoch": 13.718562874251496, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1399, "step": 9164 }, { "epoch": 13.720059880239521, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1519, "step": 9165 }, { "epoch": 13.721556886227544, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.143, "step": 9166 }, { "epoch": 13.72305389221557, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1452, "step": 9167 }, { "epoch": 13.724550898203592, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.14, "step": 9168 }, { "epoch": 13.726047904191617, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1366, "step": 9169 }, { "epoch": 13.72754491017964, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1405, "step": 9170 }, { "epoch": 13.729041916167665, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.145, "step": 9171 }, { "epoch": 13.730538922155688, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9172 }, { "epoch": 13.732035928143713, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.138, "step": 9173 }, { "epoch": 13.733532934131736, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1455, "step": 9174 }, { "epoch": 13.73502994011976, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 9175 }, { "epoch": 13.736526946107784, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1455, "step": 9176 }, { "epoch": 13.738023952095809, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1393, "step": 9177 }, { "epoch": 13.739520958083832, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1437, "step": 9178 }, { "epoch": 13.741017964071856, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1413, "step": 9179 }, { "epoch": 13.74251497005988, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1384, "step": 9180 }, { "epoch": 13.744011976047904, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1458, "step": 9181 }, { "epoch": 13.745508982035929, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.138, "step": 9182 }, { "epoch": 13.747005988023952, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9183 }, { "epoch": 13.748502994011975, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1455, "step": 9184 }, { "epoch": 13.75, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.134, "step": 9185 }, { "epoch": 13.751497005988025, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1397, "step": 9186 }, { "epoch": 13.752994011976048, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.139, "step": 9187 }, { "epoch": 13.754491017964071, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1442, "step": 9188 }, { "epoch": 13.755988023952096, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1453, "step": 9189 }, { "epoch": 13.75748502994012, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1342, "step": 9190 }, { "epoch": 13.758982035928144, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1403, "step": 9191 }, { "epoch": 13.760479041916168, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1397, "step": 9192 }, { "epoch": 13.761976047904191, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1321, "step": 9193 }, { "epoch": 13.763473053892216, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1395, "step": 9194 }, { "epoch": 13.76497005988024, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1366, "step": 9195 }, { "epoch": 13.766467065868264, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1384, "step": 9196 }, { "epoch": 13.767964071856287, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1414, "step": 9197 }, { "epoch": 13.769461077844312, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1388, "step": 9198 }, { "epoch": 13.770958083832335, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1394, "step": 9199 }, { "epoch": 13.77245508982036, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1433, "step": 9200 }, { "epoch": 13.773952095808383, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1395, "step": 9201 }, { "epoch": 13.775449101796408, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1447, "step": 9202 }, { "epoch": 13.77694610778443, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1398, "step": 9203 }, { "epoch": 13.778443113772456, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1436, "step": 9204 }, { "epoch": 13.779940119760479, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.141, "step": 9205 }, { "epoch": 13.781437125748504, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1476, "step": 9206 }, { "epoch": 13.782934131736527, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1382, "step": 9207 }, { "epoch": 13.784431137724551, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1356, "step": 9208 }, { "epoch": 13.785928143712574, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1387, "step": 9209 }, { "epoch": 13.7874251497006, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1426, "step": 9210 }, { "epoch": 13.788922155688622, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1394, "step": 9211 }, { "epoch": 13.790419161676647, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 9212 }, { "epoch": 13.79191616766467, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.141, "step": 9213 }, { "epoch": 13.793413173652695, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1373, "step": 9214 }, { "epoch": 13.794910179640718, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1326, "step": 9215 }, { "epoch": 13.796407185628743, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1374, "step": 9216 }, { "epoch": 13.797904191616766, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.138, "step": 9217 }, { "epoch": 13.79940119760479, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 9218 }, { "epoch": 13.800898203592814, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1354, "step": 9219 }, { "epoch": 13.802395209580839, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1359, "step": 9220 }, { "epoch": 13.803892215568862, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1429, "step": 9221 }, { "epoch": 13.805389221556887, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1326, "step": 9222 }, { "epoch": 13.80688622754491, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1406, "step": 9223 }, { "epoch": 13.808383233532934, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 9224 }, { "epoch": 13.809880239520957, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.141, "step": 9225 }, { "epoch": 13.811377245508982, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1444, "step": 9226 }, { "epoch": 13.812874251497005, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1319, "step": 9227 }, { "epoch": 13.81437125748503, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1411, "step": 9228 }, { "epoch": 13.815868263473053, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1374, "step": 9229 }, { "epoch": 13.817365269461078, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1447, "step": 9230 }, { "epoch": 13.818862275449101, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1404, "step": 9231 }, { "epoch": 13.820359281437126, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1344, "step": 9232 }, { "epoch": 13.821856287425149, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1397, "step": 9233 }, { "epoch": 13.823353293413174, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1416, "step": 9234 }, { "epoch": 13.824850299401197, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.146, "step": 9235 }, { "epoch": 13.826347305389222, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1356, "step": 9236 }, { "epoch": 13.827844311377245, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1454, "step": 9237 }, { "epoch": 13.82934131736527, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.142, "step": 9238 }, { "epoch": 13.830838323353294, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1419, "step": 9239 }, { "epoch": 13.832335329341317, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1395, "step": 9240 }, { "epoch": 13.83383233532934, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1455, "step": 9241 }, { "epoch": 13.835329341317365, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1389, "step": 9242 }, { "epoch": 13.83682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 9243 }, { "epoch": 13.838323353293413, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1466, "step": 9244 }, { "epoch": 13.839820359281438, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1441, "step": 9245 }, { "epoch": 13.841317365269461, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1396, "step": 9246 }, { "epoch": 13.842814371257486, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 9247 }, { "epoch": 13.844311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.143, "step": 9248 }, { "epoch": 13.845808383233534, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1346, "step": 9249 }, { "epoch": 13.847305389221557, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.142, "step": 9250 }, { "epoch": 13.848802395209582, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1402, "step": 9251 }, { "epoch": 13.850299401197605, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 9252 }, { "epoch": 13.85179640718563, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1411, "step": 9253 }, { "epoch": 13.853293413173652, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1324, "step": 9254 }, { "epoch": 13.854790419161677, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1349, "step": 9255 }, { "epoch": 13.8562874251497, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9256 }, { "epoch": 13.857784431137725, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1396, "step": 9257 }, { "epoch": 13.859281437125748, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1379, "step": 9258 }, { "epoch": 13.860778443113773, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1471, "step": 9259 }, { "epoch": 13.862275449101796, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1367, "step": 9260 }, { "epoch": 13.863772455089821, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1434, "step": 9261 }, { "epoch": 13.865269461077844, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.138, "step": 9262 }, { "epoch": 13.866766467065869, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.137, "step": 9263 }, { "epoch": 13.868263473053892, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1411, "step": 9264 }, { "epoch": 13.869760479041917, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1379, "step": 9265 }, { "epoch": 13.87125748502994, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1408, "step": 9266 }, { "epoch": 13.872754491017965, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 9267 }, { "epoch": 13.874251497005988, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1338, "step": 9268 }, { "epoch": 13.875748502994012, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9269 }, { "epoch": 13.877245508982035, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1426, "step": 9270 }, { "epoch": 13.87874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1376, "step": 9271 }, { "epoch": 13.880239520958083, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.137, "step": 9272 }, { "epoch": 13.881736526946108, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1384, "step": 9273 }, { "epoch": 13.883233532934131, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1363, "step": 9274 }, { "epoch": 13.884730538922156, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1368, "step": 9275 }, { "epoch": 13.886227544910179, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1437, "step": 9276 }, { "epoch": 13.887724550898204, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9277 }, { "epoch": 13.889221556886227, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1403, "step": 9278 }, { "epoch": 13.890718562874252, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1379, "step": 9279 }, { "epoch": 13.892215568862275, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1351, "step": 9280 }, { "epoch": 13.8937125748503, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 9281 }, { "epoch": 13.895209580838323, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 9282 }, { "epoch": 13.896706586826348, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1371, "step": 9283 }, { "epoch": 13.89820359281437, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1412, "step": 9284 }, { "epoch": 13.899700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1407, "step": 9285 }, { "epoch": 13.901197604790418, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1354, "step": 9286 }, { "epoch": 13.902694610778443, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 9287 }, { "epoch": 13.904191616766466, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.143, "step": 9288 }, { "epoch": 13.905688622754491, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1396, "step": 9289 }, { "epoch": 13.907185628742514, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 9290 }, { "epoch": 13.908682634730539, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1381, "step": 9291 }, { "epoch": 13.910179640718562, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1398, "step": 9292 }, { "epoch": 13.911676646706587, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1412, "step": 9293 }, { "epoch": 13.91317365269461, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1409, "step": 9294 }, { "epoch": 13.914670658682635, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1405, "step": 9295 }, { "epoch": 13.91616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1404, "step": 9296 }, { "epoch": 13.917664670658683, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 9297 }, { "epoch": 13.919161676646706, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1421, "step": 9298 }, { "epoch": 13.92065868263473, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1439, "step": 9299 }, { "epoch": 13.922155688622755, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1393, "step": 9300 }, { "epoch": 13.923652694610778, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1402, "step": 9301 }, { "epoch": 13.925149700598803, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1355, "step": 9302 }, { "epoch": 13.926646706586826, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1317, "step": 9303 }, { "epoch": 13.928143712574851, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1378, "step": 9304 }, { "epoch": 13.929640718562874, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1366, "step": 9305 }, { "epoch": 13.931137724550899, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1416, "step": 9306 }, { "epoch": 13.932634730538922, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1417, "step": 9307 }, { "epoch": 13.934131736526947, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1453, "step": 9308 }, { "epoch": 13.93562874251497, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1476, "step": 9309 }, { "epoch": 13.937125748502995, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1403, "step": 9310 }, { "epoch": 13.938622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1452, "step": 9311 }, { "epoch": 13.940119760479043, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1422, "step": 9312 }, { "epoch": 13.941616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 9313 }, { "epoch": 13.94311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1416, "step": 9314 }, { "epoch": 13.944610778443113, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1411, "step": 9315 }, { "epoch": 13.946107784431138, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9316 }, { "epoch": 13.947604790419161, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1432, "step": 9317 }, { "epoch": 13.949101796407186, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.143, "step": 9318 }, { "epoch": 13.95059880239521, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1363, "step": 9319 }, { "epoch": 13.952095808383234, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 9320 }, { "epoch": 13.953592814371257, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1354, "step": 9321 }, { "epoch": 13.955089820359282, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1367, "step": 9322 }, { "epoch": 13.956586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 9323 }, { "epoch": 13.95808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1404, "step": 9324 }, { "epoch": 13.959580838323353, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1421, "step": 9325 }, { "epoch": 13.961077844311378, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1381, "step": 9326 }, { "epoch": 13.9625748502994, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 9327 }, { "epoch": 13.964071856287426, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1374, "step": 9328 }, { "epoch": 13.965568862275449, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1303, "step": 9329 }, { "epoch": 13.967065868263473, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.131, "step": 9330 }, { "epoch": 13.968562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1366, "step": 9331 }, { "epoch": 13.970059880239521, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1426, "step": 9332 }, { "epoch": 13.971556886227544, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1364, "step": 9333 }, { "epoch": 13.97305389221557, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1458, "step": 9334 }, { "epoch": 13.974550898203592, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.139, "step": 9335 }, { "epoch": 13.976047904191617, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1389, "step": 9336 }, { "epoch": 13.97754491017964, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1401, "step": 9337 }, { "epoch": 13.979041916167665, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 9338 }, { "epoch": 13.980538922155688, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1392, "step": 9339 }, { "epoch": 13.982035928143713, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1346, "step": 9340 }, { "epoch": 13.983532934131736, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1364, "step": 9341 }, { "epoch": 13.98502994011976, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.139, "step": 9342 }, { "epoch": 13.986526946107784, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.143, "step": 9343 }, { "epoch": 13.988023952095809, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.142, "step": 9344 }, { "epoch": 13.989520958083832, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1386, "step": 9345 }, { "epoch": 13.991017964071856, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1374, "step": 9346 }, { "epoch": 13.99251497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1476, "step": 9347 }, { "epoch": 13.994011976047904, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1424, "step": 9348 }, { "epoch": 13.995508982035929, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1418, "step": 9349 }, { "epoch": 13.997005988023952, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1399, "step": 9350 }, { "epoch": 13.998502994011975, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1425, "step": 9351 }, { "epoch": 14.0, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1443, "step": 9352 }, { "epoch": 14.001497005988025, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1446, "step": 9353 }, { "epoch": 14.002994011976048, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1345, "step": 9354 }, { "epoch": 14.004491017964073, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1434, "step": 9355 }, { "epoch": 14.005988023952096, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1307, "step": 9356 }, { "epoch": 14.00748502994012, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1419, "step": 9357 }, { "epoch": 14.008982035928144, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1299, "step": 9358 }, { "epoch": 14.010479041916168, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1316, "step": 9359 }, { "epoch": 14.011976047904191, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1351, "step": 9360 }, { "epoch": 14.013473053892216, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.146, "step": 9361 }, { "epoch": 14.01497005988024, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1422, "step": 9362 }, { "epoch": 14.016467065868264, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1433, "step": 9363 }, { "epoch": 14.017964071856287, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1401, "step": 9364 }, { "epoch": 14.019461077844312, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1408, "step": 9365 }, { "epoch": 14.020958083832335, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1395, "step": 9366 }, { "epoch": 14.02245508982036, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 9367 }, { "epoch": 14.023952095808383, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1393, "step": 9368 }, { "epoch": 14.025449101796408, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1397, "step": 9369 }, { "epoch": 14.02694610778443, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1353, "step": 9370 }, { "epoch": 14.028443113772456, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1415, "step": 9371 }, { "epoch": 14.029940119760479, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.136, "step": 9372 }, { "epoch": 14.031437125748504, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1351, "step": 9373 }, { "epoch": 14.032934131736527, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1334, "step": 9374 }, { "epoch": 14.034431137724551, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1345, "step": 9375 }, { "epoch": 14.035928143712574, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1366, "step": 9376 }, { "epoch": 14.0374251497006, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1414, "step": 9377 }, { "epoch": 14.038922155688622, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1481, "step": 9378 }, { "epoch": 14.040419161676647, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9379 }, { "epoch": 14.04191616766467, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.137, "step": 9380 }, { "epoch": 14.043413173652695, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1364, "step": 9381 }, { "epoch": 14.044910179640718, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 9382 }, { "epoch": 14.046407185628743, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 9383 }, { "epoch": 14.047904191616766, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1443, "step": 9384 }, { "epoch": 14.04940119760479, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1371, "step": 9385 }, { "epoch": 14.050898203592814, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9386 }, { "epoch": 14.052395209580839, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1404, "step": 9387 }, { "epoch": 14.053892215568862, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1349, "step": 9388 }, { "epoch": 14.055389221556887, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1346, "step": 9389 }, { "epoch": 14.05688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 9390 }, { "epoch": 14.058383233532934, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1417, "step": 9391 }, { "epoch": 14.059880239520957, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1346, "step": 9392 }, { "epoch": 14.061377245508982, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1408, "step": 9393 }, { "epoch": 14.062874251497005, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1393, "step": 9394 }, { "epoch": 14.06437125748503, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.132, "step": 9395 }, { "epoch": 14.065868263473053, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1382, "step": 9396 }, { "epoch": 14.067365269461078, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1372, "step": 9397 }, { "epoch": 14.068862275449101, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1393, "step": 9398 }, { "epoch": 14.070359281437126, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1343, "step": 9399 }, { "epoch": 14.071856287425149, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1429, "step": 9400 }, { "epoch": 14.073353293413174, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1356, "step": 9401 }, { "epoch": 14.074850299401197, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1418, "step": 9402 }, { "epoch": 14.076347305389222, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1427, "step": 9403 }, { "epoch": 14.077844311377245, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1388, "step": 9404 }, { "epoch": 14.07934131736527, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1461, "step": 9405 }, { "epoch": 14.080838323353293, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1329, "step": 9406 }, { "epoch": 14.082335329341317, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1444, "step": 9407 }, { "epoch": 14.08383233532934, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1353, "step": 9408 }, { "epoch": 14.085329341317365, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1465, "step": 9409 }, { "epoch": 14.08682634730539, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1302, "step": 9410 }, { "epoch": 14.088323353293413, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1397, "step": 9411 }, { "epoch": 14.089820359281438, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.138, "step": 9412 }, { "epoch": 14.091317365269461, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1413, "step": 9413 }, { "epoch": 14.092814371257486, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1374, "step": 9414 }, { "epoch": 14.094311377245509, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1391, "step": 9415 }, { "epoch": 14.095808383233534, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1394, "step": 9416 }, { "epoch": 14.097305389221557, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1397, "step": 9417 }, { "epoch": 14.098802395209582, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1444, "step": 9418 }, { "epoch": 14.100299401197605, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1459, "step": 9419 }, { "epoch": 14.10179640718563, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1376, "step": 9420 }, { "epoch": 14.103293413173652, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1393, "step": 9421 }, { "epoch": 14.104790419161677, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1404, "step": 9422 }, { "epoch": 14.1062874251497, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1363, "step": 9423 }, { "epoch": 14.107784431137725, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.139, "step": 9424 }, { "epoch": 14.109281437125748, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1444, "step": 9425 }, { "epoch": 14.110778443113773, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1368, "step": 9426 }, { "epoch": 14.112275449101796, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1358, "step": 9427 }, { "epoch": 14.113772455089821, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1424, "step": 9428 }, { "epoch": 14.115269461077844, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1377, "step": 9429 }, { "epoch": 14.116766467065869, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1391, "step": 9430 }, { "epoch": 14.118263473053892, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1316, "step": 9431 }, { "epoch": 14.119760479041917, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1423, "step": 9432 }, { "epoch": 14.12125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1423, "step": 9433 }, { "epoch": 14.122754491017965, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.134, "step": 9434 }, { "epoch": 14.124251497005988, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1397, "step": 9435 }, { "epoch": 14.125748502994012, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1376, "step": 9436 }, { "epoch": 14.127245508982035, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.139, "step": 9437 }, { "epoch": 14.12874251497006, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1388, "step": 9438 }, { "epoch": 14.130239520958083, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1434, "step": 9439 }, { "epoch": 14.131736526946108, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1336, "step": 9440 }, { "epoch": 14.133233532934131, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1409, "step": 9441 }, { "epoch": 14.134730538922156, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1394, "step": 9442 }, { "epoch": 14.136227544910179, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1426, "step": 9443 }, { "epoch": 14.137724550898204, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1355, "step": 9444 }, { "epoch": 14.139221556886227, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 9445 }, { "epoch": 14.140718562874252, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.134, "step": 9446 }, { "epoch": 14.142215568862275, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1389, "step": 9447 }, { "epoch": 14.1437125748503, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1432, "step": 9448 }, { "epoch": 14.145209580838323, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1377, "step": 9449 }, { "epoch": 14.146706586826348, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1394, "step": 9450 }, { "epoch": 14.14820359281437, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1324, "step": 9451 }, { "epoch": 14.149700598802395, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1476, "step": 9452 }, { "epoch": 14.151197604790418, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 9453 }, { "epoch": 14.152694610778443, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9454 }, { "epoch": 14.154191616766466, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1378, "step": 9455 }, { "epoch": 14.155688622754491, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 9456 }, { "epoch": 14.157185628742514, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9457 }, { "epoch": 14.158682634730539, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1433, "step": 9458 }, { "epoch": 14.160179640718562, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1384, "step": 9459 }, { "epoch": 14.161676646706587, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1363, "step": 9460 }, { "epoch": 14.16317365269461, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.135, "step": 9461 }, { "epoch": 14.164670658682635, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1397, "step": 9462 }, { "epoch": 14.16616766467066, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1367, "step": 9463 }, { "epoch": 14.167664670658683, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1416, "step": 9464 }, { "epoch": 14.169161676646707, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1359, "step": 9465 }, { "epoch": 14.17065868263473, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.136, "step": 9466 }, { "epoch": 14.172155688622755, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1359, "step": 9467 }, { "epoch": 14.173652694610778, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 9468 }, { "epoch": 14.175149700598803, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 9469 }, { "epoch": 14.176646706586826, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1369, "step": 9470 }, { "epoch": 14.178143712574851, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1362, "step": 9471 }, { "epoch": 14.179640718562874, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1377, "step": 9472 }, { "epoch": 14.181137724550899, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1425, "step": 9473 }, { "epoch": 14.182634730538922, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1458, "step": 9474 }, { "epoch": 14.184131736526947, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1402, "step": 9475 }, { "epoch": 14.18562874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1438, "step": 9476 }, { "epoch": 14.187125748502995, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1366, "step": 9477 }, { "epoch": 14.188622754491018, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 9478 }, { "epoch": 14.190119760479043, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1355, "step": 9479 }, { "epoch": 14.191616766467066, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1355, "step": 9480 }, { "epoch": 14.19311377245509, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1355, "step": 9481 }, { "epoch": 14.194610778443113, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1369, "step": 9482 }, { "epoch": 14.196107784431138, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 9483 }, { "epoch": 14.197604790419161, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1334, "step": 9484 }, { "epoch": 14.199101796407186, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1369, "step": 9485 }, { "epoch": 14.20059880239521, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1366, "step": 9486 }, { "epoch": 14.202095808383234, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1398, "step": 9487 }, { "epoch": 14.203592814371257, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1392, "step": 9488 }, { "epoch": 14.205089820359282, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1368, "step": 9489 }, { "epoch": 14.206586826347305, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1445, "step": 9490 }, { "epoch": 14.20808383233533, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1434, "step": 9491 }, { "epoch": 14.209580838323353, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1362, "step": 9492 }, { "epoch": 14.211077844311378, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1405, "step": 9493 }, { "epoch": 14.2125748502994, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9494 }, { "epoch": 14.214071856287426, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.136, "step": 9495 }, { "epoch": 14.215568862275449, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 9496 }, { "epoch": 14.217065868263473, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1337, "step": 9497 }, { "epoch": 14.218562874251496, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1379, "step": 9498 }, { "epoch": 14.220059880239521, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1423, "step": 9499 }, { "epoch": 14.221556886227544, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1276, "step": 9500 }, { "epoch": 14.22305389221557, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 9501 }, { "epoch": 14.224550898203592, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1443, "step": 9502 }, { "epoch": 14.226047904191617, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.137, "step": 9503 }, { "epoch": 14.22754491017964, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.1429, "step": 9504 }, { "epoch": 14.229041916167665, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1399, "step": 9505 }, { "epoch": 14.230538922155688, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1388, "step": 9506 }, { "epoch": 14.232035928143713, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1337, "step": 9507 }, { "epoch": 14.233532934131736, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1423, "step": 9508 }, { "epoch": 14.23502994011976, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1341, "step": 9509 }, { "epoch": 14.236526946107784, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1339, "step": 9510 }, { "epoch": 14.238023952095809, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1353, "step": 9511 }, { "epoch": 14.239520958083832, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1404, "step": 9512 }, { "epoch": 14.241017964071856, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1336, "step": 9513 }, { "epoch": 14.24251497005988, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1348, "step": 9514 }, { "epoch": 14.244011976047904, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1326, "step": 9515 }, { "epoch": 14.245508982035927, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1401, "step": 9516 }, { "epoch": 14.247005988023952, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1406, "step": 9517 }, { "epoch": 14.248502994011975, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1319, "step": 9518 }, { "epoch": 14.25, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1396, "step": 9519 }, { "epoch": 14.251497005988025, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1352, "step": 9520 }, { "epoch": 14.252994011976048, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1383, "step": 9521 }, { "epoch": 14.254491017964073, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1357, "step": 9522 }, { "epoch": 14.255988023952096, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.139, "step": 9523 }, { "epoch": 14.25748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1404, "step": 9524 }, { "epoch": 14.258982035928144, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1417, "step": 9525 }, { "epoch": 14.260479041916168, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 9526 }, { "epoch": 14.261976047904191, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1399, "step": 9527 }, { "epoch": 14.263473053892216, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1392, "step": 9528 }, { "epoch": 14.26497005988024, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1426, "step": 9529 }, { "epoch": 14.266467065868264, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.138, "step": 9530 }, { "epoch": 14.267964071856287, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 9531 }, { "epoch": 14.269461077844312, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1383, "step": 9532 }, { "epoch": 14.270958083832335, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1409, "step": 9533 }, { "epoch": 14.27245508982036, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1347, "step": 9534 }, { "epoch": 14.273952095808383, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1364, "step": 9535 }, { "epoch": 14.275449101796408, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1376, "step": 9536 }, { "epoch": 14.27694610778443, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1407, "step": 9537 }, { "epoch": 14.278443113772456, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.137, "step": 9538 }, { "epoch": 14.279940119760479, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1341, "step": 9539 }, { "epoch": 14.281437125748504, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1419, "step": 9540 }, { "epoch": 14.282934131736527, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9541 }, { "epoch": 14.284431137724551, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1417, "step": 9542 }, { "epoch": 14.285928143712574, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1296, "step": 9543 }, { "epoch": 14.2874251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1411, "step": 9544 }, { "epoch": 14.288922155688622, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.143, "step": 9545 }, { "epoch": 14.290419161676647, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9546 }, { "epoch": 14.29191616766467, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9547 }, { "epoch": 14.293413173652695, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1362, "step": 9548 }, { "epoch": 14.294910179640718, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1367, "step": 9549 }, { "epoch": 14.296407185628743, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1399, "step": 9550 }, { "epoch": 14.297904191616766, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1333, "step": 9551 }, { "epoch": 14.29940119760479, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1347, "step": 9552 }, { "epoch": 14.300898203592814, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1379, "step": 9553 }, { "epoch": 14.302395209580839, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.138, "step": 9554 }, { "epoch": 14.303892215568862, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 9555 }, { "epoch": 14.305389221556887, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1425, "step": 9556 }, { "epoch": 14.30688622754491, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.14, "step": 9557 }, { "epoch": 14.308383233532934, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1351, "step": 9558 }, { "epoch": 14.309880239520957, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1394, "step": 9559 }, { "epoch": 14.311377245508982, "grad_norm": 0.193359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 9560 }, { "epoch": 14.312874251497005, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1423, "step": 9561 }, { "epoch": 14.31437125748503, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1486, "step": 9562 }, { "epoch": 14.315868263473053, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1305, "step": 9563 }, { "epoch": 14.317365269461078, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1371, "step": 9564 }, { "epoch": 14.318862275449101, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1348, "step": 9565 }, { "epoch": 14.320359281437126, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1356, "step": 9566 }, { "epoch": 14.321856287425149, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1429, "step": 9567 }, { "epoch": 14.323353293413174, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1344, "step": 9568 }, { "epoch": 14.324850299401197, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1307, "step": 9569 }, { "epoch": 14.326347305389222, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.137, "step": 9570 }, { "epoch": 14.327844311377245, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1341, "step": 9571 }, { "epoch": 14.32934131736527, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1399, "step": 9572 }, { "epoch": 14.330838323353294, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1395, "step": 9573 }, { "epoch": 14.332335329341317, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1412, "step": 9574 }, { "epoch": 14.33383233532934, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1346, "step": 9575 }, { "epoch": 14.335329341317365, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 9576 }, { "epoch": 14.33682634730539, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1405, "step": 9577 }, { "epoch": 14.338323353293413, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1419, "step": 9578 }, { "epoch": 14.339820359281438, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1396, "step": 9579 }, { "epoch": 14.341317365269461, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.134, "step": 9580 }, { "epoch": 14.342814371257486, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.141, "step": 9581 }, { "epoch": 14.344311377245509, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1429, "step": 9582 }, { "epoch": 14.345808383233534, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1394, "step": 9583 }, { "epoch": 14.347305389221557, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1411, "step": 9584 }, { "epoch": 14.348802395209582, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1325, "step": 9585 }, { "epoch": 14.350299401197605, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1371, "step": 9586 }, { "epoch": 14.35179640718563, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1365, "step": 9587 }, { "epoch": 14.353293413173652, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9588 }, { "epoch": 14.354790419161677, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1344, "step": 9589 }, { "epoch": 14.3562874251497, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1335, "step": 9590 }, { "epoch": 14.357784431137725, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1407, "step": 9591 }, { "epoch": 14.359281437125748, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1369, "step": 9592 }, { "epoch": 14.360778443113773, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1411, "step": 9593 }, { "epoch": 14.362275449101796, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1318, "step": 9594 }, { "epoch": 14.363772455089821, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1364, "step": 9595 }, { "epoch": 14.365269461077844, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1387, "step": 9596 }, { "epoch": 14.366766467065869, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1385, "step": 9597 }, { "epoch": 14.368263473053892, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1355, "step": 9598 }, { "epoch": 14.369760479041917, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9599 }, { "epoch": 14.37125748502994, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1352, "step": 9600 }, { "epoch": 14.372754491017965, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1418, "step": 9601 }, { "epoch": 14.374251497005988, "grad_norm": 0.04931640625, "learning_rate": 0.0008, "loss": 1.1313, "step": 9602 }, { "epoch": 14.375748502994012, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1401, "step": 9603 }, { "epoch": 14.377245508982035, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.135, "step": 9604 }, { "epoch": 14.37874251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1435, "step": 9605 }, { "epoch": 14.380239520958083, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1391, "step": 9606 }, { "epoch": 14.381736526946108, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1368, "step": 9607 }, { "epoch": 14.383233532934131, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.141, "step": 9608 }, { "epoch": 14.384730538922156, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1325, "step": 9609 }, { "epoch": 14.386227544910179, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1373, "step": 9610 }, { "epoch": 14.387724550898204, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1372, "step": 9611 }, { "epoch": 14.389221556886227, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1356, "step": 9612 }, { "epoch": 14.390718562874252, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 9613 }, { "epoch": 14.392215568862275, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1368, "step": 9614 }, { "epoch": 14.3937125748503, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1439, "step": 9615 }, { "epoch": 14.395209580838323, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1388, "step": 9616 }, { "epoch": 14.396706586826348, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.131, "step": 9617 }, { "epoch": 14.39820359281437, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1369, "step": 9618 }, { "epoch": 14.399700598802395, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.129, "step": 9619 }, { "epoch": 14.401197604790418, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1419, "step": 9620 }, { "epoch": 14.402694610778443, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 9621 }, { "epoch": 14.404191616766466, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1419, "step": 9622 }, { "epoch": 14.405688622754491, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1344, "step": 9623 }, { "epoch": 14.407185628742514, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1391, "step": 9624 }, { "epoch": 14.408682634730539, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 9625 }, { "epoch": 14.410179640718562, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1409, "step": 9626 }, { "epoch": 14.411676646706587, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1379, "step": 9627 }, { "epoch": 14.41317365269461, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1376, "step": 9628 }, { "epoch": 14.414670658682635, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1367, "step": 9629 }, { "epoch": 14.41616766467066, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1344, "step": 9630 }, { "epoch": 14.417664670658683, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1412, "step": 9631 }, { "epoch": 14.419161676646706, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9632 }, { "epoch": 14.42065868263473, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1367, "step": 9633 }, { "epoch": 14.422155688622755, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1408, "step": 9634 }, { "epoch": 14.423652694610778, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1324, "step": 9635 }, { "epoch": 14.425149700598803, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1359, "step": 9636 }, { "epoch": 14.426646706586826, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1383, "step": 9637 }, { "epoch": 14.428143712574851, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 9638 }, { "epoch": 14.429640718562874, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1387, "step": 9639 }, { "epoch": 14.431137724550899, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1371, "step": 9640 }, { "epoch": 14.432634730538922, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1343, "step": 9641 }, { "epoch": 14.434131736526947, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1361, "step": 9642 }, { "epoch": 14.43562874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1369, "step": 9643 }, { "epoch": 14.437125748502995, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1481, "step": 9644 }, { "epoch": 14.438622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1441, "step": 9645 }, { "epoch": 14.440119760479043, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1365, "step": 9646 }, { "epoch": 14.441616766467066, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1355, "step": 9647 }, { "epoch": 14.44311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1336, "step": 9648 }, { "epoch": 14.444610778443113, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1476, "step": 9649 }, { "epoch": 14.446107784431138, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1347, "step": 9650 }, { "epoch": 14.447604790419161, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1278, "step": 9651 }, { "epoch": 14.449101796407186, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9652 }, { "epoch": 14.45059880239521, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1304, "step": 9653 }, { "epoch": 14.452095808383234, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1384, "step": 9654 }, { "epoch": 14.453592814371257, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1442, "step": 9655 }, { "epoch": 14.455089820359282, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 9656 }, { "epoch": 14.456586826347305, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1435, "step": 9657 }, { "epoch": 14.45808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 9658 }, { "epoch": 14.459580838323353, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1374, "step": 9659 }, { "epoch": 14.461077844311378, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1463, "step": 9660 }, { "epoch": 14.4625748502994, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1366, "step": 9661 }, { "epoch": 14.464071856287426, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1338, "step": 9662 }, { "epoch": 14.465568862275449, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 9663 }, { "epoch": 14.467065868263473, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1323, "step": 9664 }, { "epoch": 14.468562874251496, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1367, "step": 9665 }, { "epoch": 14.470059880239521, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1404, "step": 9666 }, { "epoch": 14.471556886227544, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1444, "step": 9667 }, { "epoch": 14.47305389221557, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1425, "step": 9668 }, { "epoch": 14.474550898203592, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.139, "step": 9669 }, { "epoch": 14.476047904191617, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1376, "step": 9670 }, { "epoch": 14.47754491017964, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1337, "step": 9671 }, { "epoch": 14.479041916167665, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1434, "step": 9672 }, { "epoch": 14.480538922155688, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9673 }, { "epoch": 14.482035928143713, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1369, "step": 9674 }, { "epoch": 14.483532934131736, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1345, "step": 9675 }, { "epoch": 14.48502994011976, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1413, "step": 9676 }, { "epoch": 14.486526946107784, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1346, "step": 9677 }, { "epoch": 14.488023952095809, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1328, "step": 9678 }, { "epoch": 14.489520958083832, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1409, "step": 9679 }, { "epoch": 14.491017964071856, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1432, "step": 9680 }, { "epoch": 14.49251497005988, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1356, "step": 9681 }, { "epoch": 14.494011976047904, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1398, "step": 9682 }, { "epoch": 14.495508982035927, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1382, "step": 9683 }, { "epoch": 14.497005988023952, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.134, "step": 9684 }, { "epoch": 14.498502994011975, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1372, "step": 9685 }, { "epoch": 14.5, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1395, "step": 9686 }, { "epoch": 14.501497005988025, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.1391, "step": 9687 }, { "epoch": 14.502994011976048, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1307, "step": 9688 }, { "epoch": 14.504491017964071, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1376, "step": 9689 }, { "epoch": 14.505988023952096, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9690 }, { "epoch": 14.50748502994012, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1407, "step": 9691 }, { "epoch": 14.508982035928144, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1352, "step": 9692 }, { "epoch": 14.510479041916168, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1319, "step": 9693 }, { "epoch": 14.511976047904191, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1311, "step": 9694 }, { "epoch": 14.513473053892216, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1314, "step": 9695 }, { "epoch": 14.51497005988024, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1384, "step": 9696 }, { "epoch": 14.516467065868264, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1357, "step": 9697 }, { "epoch": 14.517964071856287, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1379, "step": 9698 }, { "epoch": 14.519461077844312, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1339, "step": 9699 }, { "epoch": 14.520958083832335, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1419, "step": 9700 }, { "epoch": 14.52245508982036, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1328, "step": 9701 }, { "epoch": 14.523952095808383, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1415, "step": 9702 }, { "epoch": 14.525449101796408, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1371, "step": 9703 }, { "epoch": 14.52694610778443, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1366, "step": 9704 }, { "epoch": 14.528443113772456, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.133, "step": 9705 }, { "epoch": 14.529940119760479, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.134, "step": 9706 }, { "epoch": 14.531437125748504, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1328, "step": 9707 }, { "epoch": 14.532934131736527, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 9708 }, { "epoch": 14.534431137724551, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1428, "step": 9709 }, { "epoch": 14.535928143712574, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1375, "step": 9710 }, { "epoch": 14.5374251497006, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9711 }, { "epoch": 14.538922155688622, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.14, "step": 9712 }, { "epoch": 14.540419161676647, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.138, "step": 9713 }, { "epoch": 14.54191616766467, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1396, "step": 9714 }, { "epoch": 14.543413173652695, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.132, "step": 9715 }, { "epoch": 14.544910179640718, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1378, "step": 9716 }, { "epoch": 14.546407185628743, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1349, "step": 9717 }, { "epoch": 14.547904191616766, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1378, "step": 9718 }, { "epoch": 14.54940119760479, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.142, "step": 9719 }, { "epoch": 14.550898203592814, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1448, "step": 9720 }, { "epoch": 14.552395209580839, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1376, "step": 9721 }, { "epoch": 14.553892215568862, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1421, "step": 9722 }, { "epoch": 14.555389221556887, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1371, "step": 9723 }, { "epoch": 14.55688622754491, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1344, "step": 9724 }, { "epoch": 14.558383233532934, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1443, "step": 9725 }, { "epoch": 14.559880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9726 }, { "epoch": 14.561377245508982, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 9727 }, { "epoch": 14.562874251497005, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1369, "step": 9728 }, { "epoch": 14.56437125748503, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 9729 }, { "epoch": 14.565868263473053, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1305, "step": 9730 }, { "epoch": 14.567365269461078, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.14, "step": 9731 }, { "epoch": 14.568862275449101, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1326, "step": 9732 }, { "epoch": 14.570359281437126, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 9733 }, { "epoch": 14.571856287425149, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1412, "step": 9734 }, { "epoch": 14.573353293413174, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1333, "step": 9735 }, { "epoch": 14.574850299401197, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1396, "step": 9736 }, { "epoch": 14.576347305389222, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1364, "step": 9737 }, { "epoch": 14.577844311377245, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.136, "step": 9738 }, { "epoch": 14.57934131736527, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1442, "step": 9739 }, { "epoch": 14.580838323353294, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1401, "step": 9740 }, { "epoch": 14.582335329341317, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1449, "step": 9741 }, { "epoch": 14.58383233532934, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 9742 }, { "epoch": 14.585329341317365, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.144, "step": 9743 }, { "epoch": 14.58682634730539, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1329, "step": 9744 }, { "epoch": 14.588323353293413, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 9745 }, { "epoch": 14.589820359281438, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1416, "step": 9746 }, { "epoch": 14.591317365269461, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1427, "step": 9747 }, { "epoch": 14.592814371257486, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9748 }, { "epoch": 14.594311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1378, "step": 9749 }, { "epoch": 14.595808383233534, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1382, "step": 9750 }, { "epoch": 14.597305389221557, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.135, "step": 9751 }, { "epoch": 14.598802395209582, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1386, "step": 9752 }, { "epoch": 14.600299401197605, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 9753 }, { "epoch": 14.60179640718563, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1416, "step": 9754 }, { "epoch": 14.603293413173652, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 9755 }, { "epoch": 14.604790419161677, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 9756 }, { "epoch": 14.6062874251497, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1301, "step": 9757 }, { "epoch": 14.607784431137725, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1421, "step": 9758 }, { "epoch": 14.609281437125748, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1352, "step": 9759 }, { "epoch": 14.610778443113773, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1331, "step": 9760 }, { "epoch": 14.612275449101796, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1367, "step": 9761 }, { "epoch": 14.613772455089821, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1329, "step": 9762 }, { "epoch": 14.615269461077844, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1332, "step": 9763 }, { "epoch": 14.616766467065869, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.133, "step": 9764 }, { "epoch": 14.618263473053892, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1363, "step": 9765 }, { "epoch": 14.619760479041917, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1396, "step": 9766 }, { "epoch": 14.62125748502994, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1327, "step": 9767 }, { "epoch": 14.622754491017965, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1349, "step": 9768 }, { "epoch": 14.624251497005988, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1345, "step": 9769 }, { "epoch": 14.625748502994012, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1415, "step": 9770 }, { "epoch": 14.627245508982035, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1351, "step": 9771 }, { "epoch": 14.62874251497006, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1425, "step": 9772 }, { "epoch": 14.630239520958083, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1402, "step": 9773 }, { "epoch": 14.631736526946108, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1457, "step": 9774 }, { "epoch": 14.633233532934131, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 9775 }, { "epoch": 14.634730538922156, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1374, "step": 9776 }, { "epoch": 14.636227544910179, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1376, "step": 9777 }, { "epoch": 14.637724550898204, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1354, "step": 9778 }, { "epoch": 14.639221556886227, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1425, "step": 9779 }, { "epoch": 14.640718562874252, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1353, "step": 9780 }, { "epoch": 14.642215568862275, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1327, "step": 9781 }, { "epoch": 14.6437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 9782 }, { "epoch": 14.645209580838323, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 9783 }, { "epoch": 14.646706586826348, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1397, "step": 9784 }, { "epoch": 14.64820359281437, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1393, "step": 9785 }, { "epoch": 14.649700598802395, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1387, "step": 9786 }, { "epoch": 14.651197604790418, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1391, "step": 9787 }, { "epoch": 14.652694610778443, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1415, "step": 9788 }, { "epoch": 14.654191616766466, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.142, "step": 9789 }, { "epoch": 14.655688622754491, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.139, "step": 9790 }, { "epoch": 14.657185628742514, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1357, "step": 9791 }, { "epoch": 14.658682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1379, "step": 9792 }, { "epoch": 14.660179640718562, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 9793 }, { "epoch": 14.661676646706587, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1324, "step": 9794 }, { "epoch": 14.66317365269461, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 9795 }, { "epoch": 14.664670658682635, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 9796 }, { "epoch": 14.66616766467066, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1344, "step": 9797 }, { "epoch": 14.667664670658683, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1329, "step": 9798 }, { "epoch": 14.669161676646706, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1419, "step": 9799 }, { "epoch": 14.67065868263473, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1382, "step": 9800 }, { "epoch": 14.672155688622755, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.141, "step": 9801 }, { "epoch": 14.673652694610778, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.132, "step": 9802 }, { "epoch": 14.675149700598803, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.134, "step": 9803 }, { "epoch": 14.676646706586826, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1417, "step": 9804 }, { "epoch": 14.678143712574851, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.134, "step": 9805 }, { "epoch": 14.679640718562874, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1377, "step": 9806 }, { "epoch": 14.681137724550899, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1472, "step": 9807 }, { "epoch": 14.682634730538922, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1417, "step": 9808 }, { "epoch": 14.684131736526947, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1365, "step": 9809 }, { "epoch": 14.68562874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1357, "step": 9810 }, { "epoch": 14.687125748502995, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1298, "step": 9811 }, { "epoch": 14.688622754491018, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1348, "step": 9812 }, { "epoch": 14.690119760479043, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 9813 }, { "epoch": 14.691616766467066, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1315, "step": 9814 }, { "epoch": 14.69311377245509, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1408, "step": 9815 }, { "epoch": 14.694610778443113, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1383, "step": 9816 }, { "epoch": 14.696107784431138, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1444, "step": 9817 }, { "epoch": 14.697604790419161, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9818 }, { "epoch": 14.699101796407186, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1377, "step": 9819 }, { "epoch": 14.70059880239521, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1384, "step": 9820 }, { "epoch": 14.702095808383234, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 9821 }, { "epoch": 14.703592814371257, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1443, "step": 9822 }, { "epoch": 14.705089820359282, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1397, "step": 9823 }, { "epoch": 14.706586826347305, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1364, "step": 9824 }, { "epoch": 14.70808383233533, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1434, "step": 9825 }, { "epoch": 14.709580838323353, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1411, "step": 9826 }, { "epoch": 14.711077844311378, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9827 }, { "epoch": 14.7125748502994, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1339, "step": 9828 }, { "epoch": 14.714071856287426, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.138, "step": 9829 }, { "epoch": 14.715568862275449, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 9830 }, { "epoch": 14.717065868263473, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1412, "step": 9831 }, { "epoch": 14.718562874251496, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.146, "step": 9832 }, { "epoch": 14.720059880239521, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1425, "step": 9833 }, { "epoch": 14.721556886227544, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1374, "step": 9834 }, { "epoch": 14.72305389221557, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1437, "step": 9835 }, { "epoch": 14.724550898203592, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1417, "step": 9836 }, { "epoch": 14.726047904191617, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 9837 }, { "epoch": 14.72754491017964, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 9838 }, { "epoch": 14.729041916167665, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1313, "step": 9839 }, { "epoch": 14.730538922155688, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1403, "step": 9840 }, { "epoch": 14.732035928143713, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 9841 }, { "epoch": 14.733532934131736, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.131, "step": 9842 }, { "epoch": 14.73502994011976, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1366, "step": 9843 }, { "epoch": 14.736526946107784, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1409, "step": 9844 }, { "epoch": 14.738023952095809, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.142, "step": 9845 }, { "epoch": 14.739520958083832, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.135, "step": 9846 }, { "epoch": 14.741017964071856, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1405, "step": 9847 }, { "epoch": 14.74251497005988, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1279, "step": 9848 }, { "epoch": 14.744011976047904, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1379, "step": 9849 }, { "epoch": 14.745508982035929, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 9850 }, { "epoch": 14.747005988023952, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1342, "step": 9851 }, { "epoch": 14.748502994011975, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1441, "step": 9852 }, { "epoch": 14.75, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1375, "step": 9853 }, { "epoch": 14.751497005988025, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1305, "step": 9854 }, { "epoch": 14.752994011976048, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1414, "step": 9855 }, { "epoch": 14.754491017964071, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1388, "step": 9856 }, { "epoch": 14.755988023952096, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1398, "step": 9857 }, { "epoch": 14.75748502994012, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1392, "step": 9858 }, { "epoch": 14.758982035928144, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1314, "step": 9859 }, { "epoch": 14.760479041916168, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1369, "step": 9860 }, { "epoch": 14.761976047904191, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9861 }, { "epoch": 14.763473053892216, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9862 }, { "epoch": 14.76497005988024, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1404, "step": 9863 }, { "epoch": 14.766467065868264, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9864 }, { "epoch": 14.767964071856287, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.143, "step": 9865 }, { "epoch": 14.769461077844312, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9866 }, { "epoch": 14.770958083832335, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1415, "step": 9867 }, { "epoch": 14.77245508982036, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.138, "step": 9868 }, { "epoch": 14.773952095808383, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1321, "step": 9869 }, { "epoch": 14.775449101796408, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 9870 }, { "epoch": 14.77694610778443, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1401, "step": 9871 }, { "epoch": 14.778443113772456, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1429, "step": 9872 }, { "epoch": 14.779940119760479, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1346, "step": 9873 }, { "epoch": 14.781437125748504, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1366, "step": 9874 }, { "epoch": 14.782934131736527, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1408, "step": 9875 }, { "epoch": 14.784431137724551, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 9876 }, { "epoch": 14.785928143712574, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.1406, "step": 9877 }, { "epoch": 14.7874251497006, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1459, "step": 9878 }, { "epoch": 14.788922155688622, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1348, "step": 9879 }, { "epoch": 14.790419161676647, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1329, "step": 9880 }, { "epoch": 14.79191616766467, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1374, "step": 9881 }, { "epoch": 14.793413173652695, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1398, "step": 9882 }, { "epoch": 14.794910179640718, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1356, "step": 9883 }, { "epoch": 14.796407185628743, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1429, "step": 9884 }, { "epoch": 14.797904191616766, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1395, "step": 9885 }, { "epoch": 14.79940119760479, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1382, "step": 9886 }, { "epoch": 14.800898203592814, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 9887 }, { "epoch": 14.802395209580839, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1433, "step": 9888 }, { "epoch": 14.803892215568862, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1326, "step": 9889 }, { "epoch": 14.805389221556887, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1314, "step": 9890 }, { "epoch": 14.80688622754491, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1359, "step": 9891 }, { "epoch": 14.808383233532934, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1428, "step": 9892 }, { "epoch": 14.809880239520957, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.132, "step": 9893 }, { "epoch": 14.811377245508982, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1419, "step": 9894 }, { "epoch": 14.812874251497005, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1373, "step": 9895 }, { "epoch": 14.81437125748503, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1366, "step": 9896 }, { "epoch": 14.815868263473053, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1409, "step": 9897 }, { "epoch": 14.817365269461078, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9898 }, { "epoch": 14.818862275449101, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 9899 }, { "epoch": 14.820359281437126, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1361, "step": 9900 }, { "epoch": 14.821856287425149, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1398, "step": 9901 }, { "epoch": 14.823353293413174, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.138, "step": 9902 }, { "epoch": 14.824850299401197, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1411, "step": 9903 }, { "epoch": 14.826347305389222, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 9904 }, { "epoch": 14.827844311377245, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1406, "step": 9905 }, { "epoch": 14.82934131736527, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1392, "step": 9906 }, { "epoch": 14.830838323353294, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1391, "step": 9907 }, { "epoch": 14.832335329341317, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1321, "step": 9908 }, { "epoch": 14.83383233532934, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1394, "step": 9909 }, { "epoch": 14.835329341317365, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 9910 }, { "epoch": 14.83682634730539, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1463, "step": 9911 }, { "epoch": 14.838323353293413, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1368, "step": 9912 }, { "epoch": 14.839820359281438, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1373, "step": 9913 }, { "epoch": 14.841317365269461, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9914 }, { "epoch": 14.842814371257486, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1325, "step": 9915 }, { "epoch": 14.844311377245509, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.146, "step": 9916 }, { "epoch": 14.845808383233534, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1378, "step": 9917 }, { "epoch": 14.847305389221557, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1317, "step": 9918 }, { "epoch": 14.848802395209582, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1421, "step": 9919 }, { "epoch": 14.850299401197605, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1357, "step": 9920 }, { "epoch": 14.85179640718563, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9921 }, { "epoch": 14.853293413173652, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1386, "step": 9922 }, { "epoch": 14.854790419161677, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.139, "step": 9923 }, { "epoch": 14.8562874251497, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1376, "step": 9924 }, { "epoch": 14.857784431137725, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1383, "step": 9925 }, { "epoch": 14.859281437125748, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1427, "step": 9926 }, { "epoch": 14.860778443113773, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.144, "step": 9927 }, { "epoch": 14.862275449101796, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1363, "step": 9928 }, { "epoch": 14.863772455089821, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1399, "step": 9929 }, { "epoch": 14.865269461077844, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1452, "step": 9930 }, { "epoch": 14.866766467065869, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 9931 }, { "epoch": 14.868263473053892, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1459, "step": 9932 }, { "epoch": 14.869760479041917, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1328, "step": 9933 }, { "epoch": 14.87125748502994, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1413, "step": 9934 }, { "epoch": 14.872754491017965, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1349, "step": 9935 }, { "epoch": 14.874251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1413, "step": 9936 }, { "epoch": 14.875748502994012, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1395, "step": 9937 }, { "epoch": 14.877245508982035, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1441, "step": 9938 }, { "epoch": 14.87874251497006, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 9939 }, { "epoch": 14.880239520958083, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1382, "step": 9940 }, { "epoch": 14.881736526946108, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1407, "step": 9941 }, { "epoch": 14.883233532934131, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1375, "step": 9942 }, { "epoch": 14.884730538922156, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1416, "step": 9943 }, { "epoch": 14.886227544910179, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1396, "step": 9944 }, { "epoch": 14.887724550898204, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 9945 }, { "epoch": 14.889221556886227, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1376, "step": 9946 }, { "epoch": 14.890718562874252, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.136, "step": 9947 }, { "epoch": 14.892215568862275, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1352, "step": 9948 }, { "epoch": 14.8937125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1379, "step": 9949 }, { "epoch": 14.895209580838323, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1401, "step": 9950 }, { "epoch": 14.896706586826348, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1396, "step": 9951 }, { "epoch": 14.89820359281437, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1388, "step": 9952 }, { "epoch": 14.899700598802395, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1397, "step": 9953 }, { "epoch": 14.901197604790418, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1354, "step": 9954 }, { "epoch": 14.902694610778443, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1372, "step": 9955 }, { "epoch": 14.904191616766466, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1423, "step": 9956 }, { "epoch": 14.905688622754491, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1389, "step": 9957 }, { "epoch": 14.907185628742514, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9958 }, { "epoch": 14.908682634730539, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1388, "step": 9959 }, { "epoch": 14.910179640718562, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1325, "step": 9960 }, { "epoch": 14.911676646706587, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1393, "step": 9961 }, { "epoch": 14.91317365269461, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 9962 }, { "epoch": 14.914670658682635, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1406, "step": 9963 }, { "epoch": 14.91616766467066, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1381, "step": 9964 }, { "epoch": 14.917664670658683, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 9965 }, { "epoch": 14.919161676646706, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1353, "step": 9966 }, { "epoch": 14.92065868263473, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1407, "step": 9967 }, { "epoch": 14.922155688622755, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1313, "step": 9968 }, { "epoch": 14.923652694610778, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.139, "step": 9969 }, { "epoch": 14.925149700598803, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1323, "step": 9970 }, { "epoch": 14.926646706586826, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.13, "step": 9971 }, { "epoch": 14.928143712574851, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1388, "step": 9972 }, { "epoch": 14.929640718562874, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1424, "step": 9973 }, { "epoch": 14.931137724550899, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1328, "step": 9974 }, { "epoch": 14.932634730538922, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1447, "step": 9975 }, { "epoch": 14.934131736526947, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9976 }, { "epoch": 14.93562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.137, "step": 9977 }, { "epoch": 14.937125748502995, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1317, "step": 9978 }, { "epoch": 14.938622754491018, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1367, "step": 9979 }, { "epoch": 14.940119760479043, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1448, "step": 9980 }, { "epoch": 14.941616766467066, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.134, "step": 9981 }, { "epoch": 14.94311377245509, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1383, "step": 9982 }, { "epoch": 14.944610778443113, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1364, "step": 9983 }, { "epoch": 14.946107784431138, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1376, "step": 9984 }, { "epoch": 14.947604790419161, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1423, "step": 9985 }, { "epoch": 14.949101796407186, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1327, "step": 9986 }, { "epoch": 14.95059880239521, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1366, "step": 9987 }, { "epoch": 14.952095808383234, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 9988 }, { "epoch": 14.953592814371257, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1348, "step": 9989 }, { "epoch": 14.955089820359282, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1382, "step": 9990 }, { "epoch": 14.956586826347305, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9991 }, { "epoch": 14.95808383233533, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1445, "step": 9992 }, { "epoch": 14.959580838323353, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9993 }, { "epoch": 14.961077844311378, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.138, "step": 9994 }, { "epoch": 14.9625748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1307, "step": 9995 }, { "epoch": 14.964071856287426, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.134, "step": 9996 }, { "epoch": 14.965568862275449, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1401, "step": 9997 }, { "epoch": 14.967065868263473, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1395, "step": 9998 }, { "epoch": 14.968562874251496, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1343, "step": 9999 }, { "epoch": 14.970059880239521, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1363, "step": 10000 }, { "epoch": 14.971556886227544, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1435, "step": 10001 }, { "epoch": 14.97305389221557, "grad_norm": 0.7265625, "learning_rate": 0.0008, "loss": 1.1953, "step": 10002 }, { "epoch": 14.974550898203592, "grad_norm": 0.5703125, "learning_rate": 0.0008, "loss": 1.197, "step": 10003 }, { "epoch": 14.976047904191617, "grad_norm": 1.0859375, "learning_rate": 0.0008, "loss": 1.2462, "step": 10004 }, { "epoch": 14.97754491017964, "grad_norm": 0.8125, "learning_rate": 0.0008, "loss": 1.322, "step": 10005 }, { "epoch": 14.979041916167665, "grad_norm": 0.9453125, "learning_rate": 0.0008, "loss": 1.3121, "step": 10006 }, { "epoch": 14.980538922155688, "grad_norm": 1.90625, "learning_rate": 0.0008, "loss": 1.4386, "step": 10007 }, { "epoch": 14.982035928143713, "grad_norm": 0.91796875, "learning_rate": 0.0008, "loss": 1.415, "step": 10008 }, { "epoch": 14.983532934131736, "grad_norm": 1.421875, "learning_rate": 0.0008, "loss": 1.4384, "step": 10009 }, { "epoch": 14.98502994011976, "grad_norm": 1.09375, "learning_rate": 0.0008, "loss": 1.4024, "step": 10010 }, { "epoch": 14.986526946107784, "grad_norm": 1.4453125, "learning_rate": 0.0008, "loss": 1.4154, "step": 10011 }, { "epoch": 14.988023952095809, "grad_norm": 0.63671875, "learning_rate": 0.0008, "loss": 1.3349, "step": 10012 }, { "epoch": 14.989520958083832, "grad_norm": 0.71484375, "learning_rate": 0.0008, "loss": 1.3392, "step": 10013 }, { "epoch": 14.991017964071856, "grad_norm": 0.71875, "learning_rate": 0.0008, "loss": 1.3292, "step": 10014 }, { "epoch": 14.99251497005988, "grad_norm": 0.77734375, "learning_rate": 0.0008, "loss": 1.33, "step": 10015 }, { "epoch": 14.994011976047904, "grad_norm": 0.78125, "learning_rate": 0.0008, "loss": 1.3175, "step": 10016 }, { "epoch": 14.995508982035929, "grad_norm": 0.69921875, "learning_rate": 0.0008, "loss": 1.3063, "step": 10017 }, { "epoch": 14.997005988023952, "grad_norm": 0.73828125, "learning_rate": 0.0008, "loss": 1.2948, "step": 10018 }, { "epoch": 14.998502994011975, "grad_norm": 0.71484375, "learning_rate": 0.0008, "loss": 1.2935, "step": 10019 }, { "epoch": 15.0, "grad_norm": 0.6171875, "learning_rate": 0.0008, "loss": 1.3024, "step": 10020 }, { "epoch": 15.001497005988025, "grad_norm": 0.51953125, "learning_rate": 0.0008, "loss": 1.2768, "step": 10021 }, { "epoch": 15.002994011976048, "grad_norm": 0.5, "learning_rate": 0.0008, "loss": 1.2798, "step": 10022 }, { "epoch": 15.004491017964073, "grad_norm": 0.41796875, "learning_rate": 0.0008, "loss": 1.2622, "step": 10023 }, { "epoch": 15.005988023952096, "grad_norm": 0.4296875, "learning_rate": 0.0008, "loss": 1.2656, "step": 10024 }, { "epoch": 15.00748502994012, "grad_norm": 0.353515625, "learning_rate": 0.0008, "loss": 1.2513, "step": 10025 }, { "epoch": 15.008982035928144, "grad_norm": 0.330078125, "learning_rate": 0.0008, "loss": 1.2537, "step": 10026 }, { "epoch": 15.010479041916168, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.2473, "step": 10027 }, { "epoch": 15.011976047904191, "grad_norm": 0.279296875, "learning_rate": 0.0008, "loss": 1.2435, "step": 10028 }, { "epoch": 15.013473053892216, "grad_norm": 0.26171875, "learning_rate": 0.0008, "loss": 1.2405, "step": 10029 }, { "epoch": 15.01497005988024, "grad_norm": 0.2412109375, "learning_rate": 0.0008, "loss": 1.2356, "step": 10030 }, { "epoch": 15.016467065868264, "grad_norm": 0.2392578125, "learning_rate": 0.0008, "loss": 1.2365, "step": 10031 }, { "epoch": 15.017964071856287, "grad_norm": 0.2138671875, "learning_rate": 0.0008, "loss": 1.2272, "step": 10032 }, { "epoch": 15.019461077844312, "grad_norm": 0.20703125, "learning_rate": 0.0008, "loss": 1.2254, "step": 10033 }, { "epoch": 15.020958083832335, "grad_norm": 0.2177734375, "learning_rate": 0.0008, "loss": 1.2223, "step": 10034 }, { "epoch": 15.02245508982036, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.2255, "step": 10035 }, { "epoch": 15.023952095808383, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.2276, "step": 10036 }, { "epoch": 15.025449101796408, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.2096, "step": 10037 }, { "epoch": 15.02694610778443, "grad_norm": 0.220703125, "learning_rate": 0.0008, "loss": 1.2236, "step": 10038 }, { "epoch": 15.028443113772456, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.2168, "step": 10039 }, { "epoch": 15.029940119760479, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.2155, "step": 10040 }, { "epoch": 15.031437125748504, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.2094, "step": 10041 }, { "epoch": 15.032934131736527, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2182, "step": 10042 }, { "epoch": 15.034431137724551, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.2101, "step": 10043 }, { "epoch": 15.035928143712574, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.2025, "step": 10044 }, { "epoch": 15.0374251497006, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.2118, "step": 10045 }, { "epoch": 15.038922155688622, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.2121, "step": 10046 }, { "epoch": 15.040419161676647, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.2078, "step": 10047 }, { "epoch": 15.04191616766467, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1975, "step": 10048 }, { "epoch": 15.043413173652695, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.2134, "step": 10049 }, { "epoch": 15.044910179640718, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.2077, "step": 10050 }, { "epoch": 15.046407185628743, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.2121, "step": 10051 }, { "epoch": 15.047904191616766, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.2069, "step": 10052 }, { "epoch": 15.04940119760479, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.2011, "step": 10053 }, { "epoch": 15.050898203592814, "grad_norm": 0.1650390625, "learning_rate": 0.0008, "loss": 1.2021, "step": 10054 }, { "epoch": 15.052395209580839, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2066, "step": 10055 }, { "epoch": 15.053892215568862, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2001, "step": 10056 }, { "epoch": 15.055389221556887, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.2066, "step": 10057 }, { "epoch": 15.05688622754491, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1997, "step": 10058 }, { "epoch": 15.058383233532934, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.2026, "step": 10059 }, { "epoch": 15.059880239520957, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1989, "step": 10060 }, { "epoch": 15.061377245508982, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1982, "step": 10061 }, { "epoch": 15.062874251497005, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.195, "step": 10062 }, { "epoch": 15.06437125748503, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1939, "step": 10063 }, { "epoch": 15.065868263473053, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1967, "step": 10064 }, { "epoch": 15.067365269461078, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1991, "step": 10065 }, { "epoch": 15.068862275449101, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1929, "step": 10066 }, { "epoch": 15.070359281437126, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1991, "step": 10067 }, { "epoch": 15.071856287425149, "grad_norm": 0.2080078125, "learning_rate": 0.0008, "loss": 1.1991, "step": 10068 }, { "epoch": 15.073353293413174, "grad_norm": 0.18359375, "learning_rate": 0.0008, "loss": 1.2034, "step": 10069 }, { "epoch": 15.074850299401197, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1983, "step": 10070 }, { "epoch": 15.076347305389222, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.2024, "step": 10071 }, { "epoch": 15.077844311377245, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.1969, "step": 10072 }, { "epoch": 15.07934131736527, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1962, "step": 10073 }, { "epoch": 15.080838323353293, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.2029, "step": 10074 }, { "epoch": 15.082335329341317, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.2002, "step": 10075 }, { "epoch": 15.08383233532934, "grad_norm": 0.271484375, "learning_rate": 0.0008, "loss": 1.2045, "step": 10076 }, { "epoch": 15.085329341317365, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1967, "step": 10077 }, { "epoch": 15.08682634730539, "grad_norm": 0.2080078125, "learning_rate": 0.0008, "loss": 1.1967, "step": 10078 }, { "epoch": 15.088323353293413, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.1964, "step": 10079 }, { "epoch": 15.089820359281438, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1886, "step": 10080 }, { "epoch": 15.091317365269461, "grad_norm": 0.2080078125, "learning_rate": 0.0008, "loss": 1.2011, "step": 10081 }, { "epoch": 15.092814371257486, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1915, "step": 10082 }, { "epoch": 15.094311377245509, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.1937, "step": 10083 }, { "epoch": 15.095808383233534, "grad_norm": 0.1943359375, "learning_rate": 0.0008, "loss": 1.1951, "step": 10084 }, { "epoch": 15.097305389221557, "grad_norm": 0.2001953125, "learning_rate": 0.0008, "loss": 1.1879, "step": 10085 }, { "epoch": 15.098802395209582, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.1895, "step": 10086 }, { "epoch": 15.100299401197605, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1964, "step": 10087 }, { "epoch": 15.10179640718563, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1939, "step": 10088 }, { "epoch": 15.103293413173652, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.2019, "step": 10089 }, { "epoch": 15.104790419161677, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1935, "step": 10090 }, { "epoch": 15.1062874251497, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1927, "step": 10091 }, { "epoch": 15.107784431137725, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1872, "step": 10092 }, { "epoch": 15.109281437125748, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1908, "step": 10093 }, { "epoch": 15.110778443113773, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.185, "step": 10094 }, { "epoch": 15.112275449101796, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1925, "step": 10095 }, { "epoch": 15.113772455089821, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.187, "step": 10096 }, { "epoch": 15.115269461077844, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1898, "step": 10097 }, { "epoch": 15.116766467065869, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1908, "step": 10098 }, { "epoch": 15.118263473053892, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1927, "step": 10099 }, { "epoch": 15.119760479041917, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1933, "step": 10100 }, { "epoch": 15.12125748502994, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1867, "step": 10101 }, { "epoch": 15.122754491017965, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1823, "step": 10102 }, { "epoch": 15.124251497005988, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1814, "step": 10103 }, { "epoch": 15.125748502994012, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1911, "step": 10104 }, { "epoch": 15.127245508982035, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1813, "step": 10105 }, { "epoch": 15.12874251497006, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1893, "step": 10106 }, { "epoch": 15.130239520958083, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1838, "step": 10107 }, { "epoch": 15.131736526946108, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1872, "step": 10108 }, { "epoch": 15.133233532934131, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1911, "step": 10109 }, { "epoch": 15.134730538922156, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1896, "step": 10110 }, { "epoch": 15.136227544910179, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.1874, "step": 10111 }, { "epoch": 15.137724550898204, "grad_norm": 0.185546875, "learning_rate": 0.0008, "loss": 1.1867, "step": 10112 }, { "epoch": 15.139221556886227, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.1852, "step": 10113 }, { "epoch": 15.140718562874252, "grad_norm": 0.232421875, "learning_rate": 0.0008, "loss": 1.1812, "step": 10114 }, { "epoch": 15.142215568862275, "grad_norm": 0.2451171875, "learning_rate": 0.0008, "loss": 1.1879, "step": 10115 }, { "epoch": 15.1437125748503, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1872, "step": 10116 }, { "epoch": 15.145209580838323, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.1864, "step": 10117 }, { "epoch": 15.146706586826348, "grad_norm": 0.23828125, "learning_rate": 0.0008, "loss": 1.1896, "step": 10118 }, { "epoch": 15.14820359281437, "grad_norm": 0.357421875, "learning_rate": 0.0008, "loss": 1.1956, "step": 10119 }, { "epoch": 15.149700598802395, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.1972, "step": 10120 }, { "epoch": 15.151197604790418, "grad_norm": 0.171875, "learning_rate": 0.0008, "loss": 1.1886, "step": 10121 }, { "epoch": 15.152694610778443, "grad_norm": 0.2392578125, "learning_rate": 0.0008, "loss": 1.1964, "step": 10122 }, { "epoch": 15.154191616766466, "grad_norm": 0.275390625, "learning_rate": 0.0008, "loss": 1.1911, "step": 10123 }, { "epoch": 15.155688622754491, "grad_norm": 0.275390625, "learning_rate": 0.0008, "loss": 1.1974, "step": 10124 }, { "epoch": 15.157185628742514, "grad_norm": 0.263671875, "learning_rate": 0.0008, "loss": 1.1901, "step": 10125 }, { "epoch": 15.158682634730539, "grad_norm": 0.1630859375, "learning_rate": 0.0008, "loss": 1.1833, "step": 10126 }, { "epoch": 15.160179640718562, "grad_norm": 0.2314453125, "learning_rate": 0.0008, "loss": 1.184, "step": 10127 }, { "epoch": 15.161676646706587, "grad_norm": 0.23046875, "learning_rate": 0.0008, "loss": 1.1825, "step": 10128 }, { "epoch": 15.16317365269461, "grad_norm": 0.1884765625, "learning_rate": 0.0008, "loss": 1.1908, "step": 10129 }, { "epoch": 15.164670658682635, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.186, "step": 10130 }, { "epoch": 15.16616766467066, "grad_norm": 0.25, "learning_rate": 0.0008, "loss": 1.1848, "step": 10131 }, { "epoch": 15.167664670658683, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.1879, "step": 10132 }, { "epoch": 15.169161676646707, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1837, "step": 10133 }, { "epoch": 15.17065868263473, "grad_norm": 0.251953125, "learning_rate": 0.0008, "loss": 1.192, "step": 10134 }, { "epoch": 15.172155688622755, "grad_norm": 0.255859375, "learning_rate": 0.0008, "loss": 1.1921, "step": 10135 }, { "epoch": 15.173652694610778, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1854, "step": 10136 }, { "epoch": 15.175149700598803, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1867, "step": 10137 }, { "epoch": 15.176646706586826, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.187, "step": 10138 }, { "epoch": 15.178143712574851, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1886, "step": 10139 }, { "epoch": 15.179640718562874, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.1865, "step": 10140 }, { "epoch": 15.181137724550899, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1871, "step": 10141 }, { "epoch": 15.182634730538922, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1883, "step": 10142 }, { "epoch": 15.184131736526947, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1778, "step": 10143 }, { "epoch": 15.18562874251497, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.182, "step": 10144 }, { "epoch": 15.187125748502995, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1829, "step": 10145 }, { "epoch": 15.188622754491018, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1779, "step": 10146 }, { "epoch": 15.190119760479043, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1845, "step": 10147 }, { "epoch": 15.191616766467066, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1856, "step": 10148 }, { "epoch": 15.19311377245509, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1764, "step": 10149 }, { "epoch": 15.194610778443113, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1791, "step": 10150 }, { "epoch": 15.196107784431138, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1829, "step": 10151 }, { "epoch": 15.197604790419161, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1779, "step": 10152 }, { "epoch": 15.199101796407186, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1816, "step": 10153 }, { "epoch": 15.20059880239521, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1851, "step": 10154 }, { "epoch": 15.202095808383234, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1788, "step": 10155 }, { "epoch": 15.203592814371257, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1736, "step": 10156 }, { "epoch": 15.205089820359282, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1862, "step": 10157 }, { "epoch": 15.206586826347305, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.179, "step": 10158 }, { "epoch": 15.20808383233533, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1776, "step": 10159 }, { "epoch": 15.209580838323353, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.172, "step": 10160 }, { "epoch": 15.211077844311378, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.176, "step": 10161 }, { "epoch": 15.2125748502994, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.1837, "step": 10162 }, { "epoch": 15.214071856287426, "grad_norm": 0.21484375, "learning_rate": 0.0008, "loss": 1.1855, "step": 10163 }, { "epoch": 15.215568862275449, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.1963, "step": 10164 }, { "epoch": 15.217065868263473, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1825, "step": 10165 }, { "epoch": 15.218562874251496, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1807, "step": 10166 }, { "epoch": 15.220059880239521, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.1775, "step": 10167 }, { "epoch": 15.221556886227544, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1853, "step": 10168 }, { "epoch": 15.22305389221557, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.1803, "step": 10169 }, { "epoch": 15.224550898203592, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.188, "step": 10170 }, { "epoch": 15.226047904191617, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1799, "step": 10171 }, { "epoch": 15.22754491017964, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.175, "step": 10172 }, { "epoch": 15.229041916167665, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1806, "step": 10173 }, { "epoch": 15.230538922155688, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1758, "step": 10174 }, { "epoch": 15.232035928143713, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1775, "step": 10175 }, { "epoch": 15.233532934131736, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1863, "step": 10176 }, { "epoch": 15.23502994011976, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1771, "step": 10177 }, { "epoch": 15.236526946107784, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1763, "step": 10178 }, { "epoch": 15.238023952095809, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1832, "step": 10179 }, { "epoch": 15.239520958083832, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.183, "step": 10180 }, { "epoch": 15.241017964071856, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1787, "step": 10181 }, { "epoch": 15.24251497005988, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1767, "step": 10182 }, { "epoch": 15.244011976047904, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1749, "step": 10183 }, { "epoch": 15.245508982035927, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1788, "step": 10184 }, { "epoch": 15.247005988023952, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1799, "step": 10185 }, { "epoch": 15.248502994011975, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1837, "step": 10186 }, { "epoch": 15.25, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1717, "step": 10187 }, { "epoch": 15.251497005988025, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1763, "step": 10188 }, { "epoch": 15.252994011976048, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1751, "step": 10189 }, { "epoch": 15.254491017964073, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1748, "step": 10190 }, { "epoch": 15.255988023952096, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1754, "step": 10191 }, { "epoch": 15.25748502994012, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1802, "step": 10192 }, { "epoch": 15.258982035928144, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1813, "step": 10193 }, { "epoch": 15.260479041916168, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1771, "step": 10194 }, { "epoch": 15.261976047904191, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1715, "step": 10195 }, { "epoch": 15.263473053892216, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1703, "step": 10196 }, { "epoch": 15.26497005988024, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.166, "step": 10197 }, { "epoch": 15.266467065868264, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1747, "step": 10198 }, { "epoch": 15.267964071856287, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1742, "step": 10199 }, { "epoch": 15.269461077844312, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1746, "step": 10200 }, { "epoch": 15.270958083832335, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1698, "step": 10201 }, { "epoch": 15.27245508982036, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1697, "step": 10202 }, { "epoch": 15.273952095808383, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1774, "step": 10203 }, { "epoch": 15.275449101796408, "grad_norm": 0.212890625, "learning_rate": 0.0008, "loss": 1.1756, "step": 10204 }, { "epoch": 15.27694610778443, "grad_norm": 0.28515625, "learning_rate": 0.0008, "loss": 1.1756, "step": 10205 }, { "epoch": 15.278443113772456, "grad_norm": 0.4375, "learning_rate": 0.0008, "loss": 1.1811, "step": 10206 }, { "epoch": 15.279940119760479, "grad_norm": 0.66796875, "learning_rate": 0.0008, "loss": 1.1842, "step": 10207 }, { "epoch": 15.281437125748504, "grad_norm": 1.0625, "learning_rate": 0.0008, "loss": 1.2017, "step": 10208 }, { "epoch": 15.282934131736527, "grad_norm": 1.3203125, "learning_rate": 0.0008, "loss": 1.211, "step": 10209 }, { "epoch": 15.284431137724551, "grad_norm": 0.400390625, "learning_rate": 0.0008, "loss": 1.1962, "step": 10210 }, { "epoch": 15.285928143712574, "grad_norm": 0.7109375, "learning_rate": 0.0008, "loss": 1.2019, "step": 10211 }, { "epoch": 15.2874251497006, "grad_norm": 0.6171875, "learning_rate": 0.0008, "loss": 1.19, "step": 10212 }, { "epoch": 15.288922155688622, "grad_norm": 0.37890625, "learning_rate": 0.0008, "loss": 1.1935, "step": 10213 }, { "epoch": 15.290419161676647, "grad_norm": 0.58984375, "learning_rate": 0.0008, "loss": 1.1886, "step": 10214 }, { "epoch": 15.29191616766467, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.1925, "step": 10215 }, { "epoch": 15.293413173652695, "grad_norm": 0.361328125, "learning_rate": 0.0008, "loss": 1.1823, "step": 10216 }, { "epoch": 15.294910179640718, "grad_norm": 0.314453125, "learning_rate": 0.0008, "loss": 1.1803, "step": 10217 }, { "epoch": 15.296407185628743, "grad_norm": 0.2138671875, "learning_rate": 0.0008, "loss": 1.1814, "step": 10218 }, { "epoch": 15.297904191616766, "grad_norm": 0.216796875, "learning_rate": 0.0008, "loss": 1.1821, "step": 10219 }, { "epoch": 15.29940119760479, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.1714, "step": 10220 }, { "epoch": 15.300898203592814, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.1753, "step": 10221 }, { "epoch": 15.302395209580839, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.1812, "step": 10222 }, { "epoch": 15.303892215568862, "grad_norm": 0.1572265625, "learning_rate": 0.0008, "loss": 1.1746, "step": 10223 }, { "epoch": 15.305389221556887, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.1741, "step": 10224 }, { "epoch": 15.30688622754491, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1798, "step": 10225 }, { "epoch": 15.308383233532934, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1718, "step": 10226 }, { "epoch": 15.309880239520957, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1733, "step": 10227 }, { "epoch": 15.311377245508982, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.17, "step": 10228 }, { "epoch": 15.312874251497005, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1695, "step": 10229 }, { "epoch": 15.31437125748503, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1696, "step": 10230 }, { "epoch": 15.315868263473053, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1637, "step": 10231 }, { "epoch": 15.317365269461078, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1708, "step": 10232 }, { "epoch": 15.318862275449101, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.172, "step": 10233 }, { "epoch": 15.320359281437126, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1734, "step": 10234 }, { "epoch": 15.321856287425149, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1715, "step": 10235 }, { "epoch": 15.323353293413174, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1648, "step": 10236 }, { "epoch": 15.324850299401197, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1736, "step": 10237 }, { "epoch": 15.326347305389222, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1684, "step": 10238 }, { "epoch": 15.327844311377245, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1694, "step": 10239 }, { "epoch": 15.32934131736527, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1621, "step": 10240 }, { "epoch": 15.330838323353294, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1726, "step": 10241 }, { "epoch": 15.332335329341317, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1757, "step": 10242 }, { "epoch": 15.33383233532934, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1782, "step": 10243 }, { "epoch": 15.335329341317365, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1713, "step": 10244 }, { "epoch": 15.33682634730539, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1693, "step": 10245 }, { "epoch": 15.338323353293413, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1737, "step": 10246 }, { "epoch": 15.339820359281438, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1747, "step": 10247 }, { "epoch": 15.341317365269461, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1707, "step": 10248 }, { "epoch": 15.342814371257486, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.164, "step": 10249 }, { "epoch": 15.344311377245509, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1663, "step": 10250 }, { "epoch": 15.345808383233534, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1603, "step": 10251 }, { "epoch": 15.347305389221557, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1664, "step": 10252 }, { "epoch": 15.348802395209582, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1656, "step": 10253 }, { "epoch": 15.350299401197605, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1663, "step": 10254 }, { "epoch": 15.35179640718563, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1658, "step": 10255 }, { "epoch": 15.353293413173652, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1655, "step": 10256 }, { "epoch": 15.354790419161677, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.1631, "step": 10257 }, { "epoch": 15.3562874251497, "grad_norm": 0.181640625, "learning_rate": 0.0008, "loss": 1.1724, "step": 10258 }, { "epoch": 15.357784431137725, "grad_norm": 0.18359375, "learning_rate": 0.0008, "loss": 1.1674, "step": 10259 }, { "epoch": 15.359281437125748, "grad_norm": 0.1787109375, "learning_rate": 0.0008, "loss": 1.1656, "step": 10260 }, { "epoch": 15.360778443113773, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1653, "step": 10261 }, { "epoch": 15.362275449101796, "grad_norm": 0.208984375, "learning_rate": 0.0008, "loss": 1.1645, "step": 10262 }, { "epoch": 15.363772455089821, "grad_norm": 0.291015625, "learning_rate": 0.0008, "loss": 1.1746, "step": 10263 }, { "epoch": 15.365269461077844, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.1789, "step": 10264 }, { "epoch": 15.366766467065869, "grad_norm": 0.1748046875, "learning_rate": 0.0008, "loss": 1.1737, "step": 10265 }, { "epoch": 15.368263473053892, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.1677, "step": 10266 }, { "epoch": 15.369760479041917, "grad_norm": 0.2255859375, "learning_rate": 0.0008, "loss": 1.1637, "step": 10267 }, { "epoch": 15.37125748502994, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1624, "step": 10268 }, { "epoch": 15.372754491017965, "grad_norm": 0.21875, "learning_rate": 0.0008, "loss": 1.1657, "step": 10269 }, { "epoch": 15.374251497005988, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.164, "step": 10270 }, { "epoch": 15.375748502994012, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.1686, "step": 10271 }, { "epoch": 15.377245508982035, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1666, "step": 10272 }, { "epoch": 15.37874251497006, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.1666, "step": 10273 }, { "epoch": 15.380239520958083, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.159, "step": 10274 }, { "epoch": 15.381736526946108, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.163, "step": 10275 }, { "epoch": 15.383233532934131, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1669, "step": 10276 }, { "epoch": 15.384730538922156, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1725, "step": 10277 }, { "epoch": 15.386227544910179, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1594, "step": 10278 }, { "epoch": 15.387724550898204, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1586, "step": 10279 }, { "epoch": 15.389221556886227, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1613, "step": 10280 }, { "epoch": 15.390718562874252, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1625, "step": 10281 }, { "epoch": 15.392215568862275, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1702, "step": 10282 }, { "epoch": 15.3937125748503, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1576, "step": 10283 }, { "epoch": 15.395209580838323, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1593, "step": 10284 }, { "epoch": 15.396706586826348, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1568, "step": 10285 }, { "epoch": 15.39820359281437, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1603, "step": 10286 }, { "epoch": 15.399700598802395, "grad_norm": 0.203125, "learning_rate": 0.0008, "loss": 1.1577, "step": 10287 }, { "epoch": 15.401197604790418, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1656, "step": 10288 }, { "epoch": 15.402694610778443, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1624, "step": 10289 }, { "epoch": 15.404191616766466, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1657, "step": 10290 }, { "epoch": 15.405688622754491, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1584, "step": 10291 }, { "epoch": 15.407185628742514, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1592, "step": 10292 }, { "epoch": 15.408682634730539, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1533, "step": 10293 }, { "epoch": 15.410179640718562, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1605, "step": 10294 }, { "epoch": 15.411676646706587, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1563, "step": 10295 }, { "epoch": 15.41317365269461, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1627, "step": 10296 }, { "epoch": 15.414670658682635, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1555, "step": 10297 }, { "epoch": 15.41616766467066, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.157, "step": 10298 }, { "epoch": 15.417664670658683, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1598, "step": 10299 }, { "epoch": 15.419161676646706, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1579, "step": 10300 }, { "epoch": 15.42065868263473, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1524, "step": 10301 }, { "epoch": 15.422155688622755, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1597, "step": 10302 }, { "epoch": 15.423652694610778, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1601, "step": 10303 }, { "epoch": 15.425149700598803, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1565, "step": 10304 }, { "epoch": 15.426646706586826, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1585, "step": 10305 }, { "epoch": 15.428143712574851, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1529, "step": 10306 }, { "epoch": 15.429640718562874, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1569, "step": 10307 }, { "epoch": 15.431137724550899, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1521, "step": 10308 }, { "epoch": 15.432634730538922, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1514, "step": 10309 }, { "epoch": 15.434131736526947, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1517, "step": 10310 }, { "epoch": 15.43562874251497, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1582, "step": 10311 }, { "epoch": 15.437125748502995, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1616, "step": 10312 }, { "epoch": 15.438622754491018, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1554, "step": 10313 }, { "epoch": 15.440119760479043, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.152, "step": 10314 }, { "epoch": 15.441616766467066, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1562, "step": 10315 }, { "epoch": 15.44311377245509, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1528, "step": 10316 }, { "epoch": 15.444610778443113, "grad_norm": 0.208984375, "learning_rate": 0.0008, "loss": 1.1558, "step": 10317 }, { "epoch": 15.446107784431138, "grad_norm": 0.24609375, "learning_rate": 0.0008, "loss": 1.1567, "step": 10318 }, { "epoch": 15.447604790419161, "grad_norm": 0.27734375, "learning_rate": 0.0008, "loss": 1.1573, "step": 10319 }, { "epoch": 15.449101796407186, "grad_norm": 0.291015625, "learning_rate": 0.0008, "loss": 1.1663, "step": 10320 }, { "epoch": 15.45059880239521, "grad_norm": 0.279296875, "learning_rate": 0.0008, "loss": 1.1592, "step": 10321 }, { "epoch": 15.452095808383234, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1534, "step": 10322 }, { "epoch": 15.453592814371257, "grad_norm": 0.1884765625, "learning_rate": 0.0008, "loss": 1.1607, "step": 10323 }, { "epoch": 15.455089820359282, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1545, "step": 10324 }, { "epoch": 15.456586826347305, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1568, "step": 10325 }, { "epoch": 15.45808383233533, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1555, "step": 10326 }, { "epoch": 15.459580838323353, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1584, "step": 10327 }, { "epoch": 15.461077844311378, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.16, "step": 10328 }, { "epoch": 15.4625748502994, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1597, "step": 10329 }, { "epoch": 15.464071856287426, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1532, "step": 10330 }, { "epoch": 15.465568862275449, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.1508, "step": 10331 }, { "epoch": 15.467065868263473, "grad_norm": 0.2197265625, "learning_rate": 0.0008, "loss": 1.1585, "step": 10332 }, { "epoch": 15.468562874251496, "grad_norm": 0.271484375, "learning_rate": 0.0008, "loss": 1.1617, "step": 10333 }, { "epoch": 15.470059880239521, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.1609, "step": 10334 }, { "epoch": 15.471556886227544, "grad_norm": 0.30078125, "learning_rate": 0.0008, "loss": 1.1608, "step": 10335 }, { "epoch": 15.47305389221557, "grad_norm": 0.32421875, "learning_rate": 0.0008, "loss": 1.1592, "step": 10336 }, { "epoch": 15.474550898203592, "grad_norm": 0.34375, "learning_rate": 0.0008, "loss": 1.1598, "step": 10337 }, { "epoch": 15.476047904191617, "grad_norm": 0.2177734375, "learning_rate": 0.0008, "loss": 1.156, "step": 10338 }, { "epoch": 15.47754491017964, "grad_norm": 0.265625, "learning_rate": 0.0008, "loss": 1.16, "step": 10339 }, { "epoch": 15.479041916167665, "grad_norm": 0.2890625, "learning_rate": 0.0008, "loss": 1.161, "step": 10340 }, { "epoch": 15.480538922155688, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1555, "step": 10341 }, { "epoch": 15.482035928143713, "grad_norm": 0.212890625, "learning_rate": 0.0008, "loss": 1.1576, "step": 10342 }, { "epoch": 15.483532934131736, "grad_norm": 0.2353515625, "learning_rate": 0.0008, "loss": 1.1636, "step": 10343 }, { "epoch": 15.48502994011976, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.1559, "step": 10344 }, { "epoch": 15.486526946107784, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.1585, "step": 10345 }, { "epoch": 15.488023952095809, "grad_norm": 0.20703125, "learning_rate": 0.0008, "loss": 1.1547, "step": 10346 }, { "epoch": 15.489520958083832, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1529, "step": 10347 }, { "epoch": 15.491017964071856, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1527, "step": 10348 }, { "epoch": 15.49251497005988, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1539, "step": 10349 }, { "epoch": 15.494011976047904, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1589, "step": 10350 }, { "epoch": 15.495508982035927, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.161, "step": 10351 }, { "epoch": 15.497005988023952, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.151, "step": 10352 }, { "epoch": 15.498502994011975, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1538, "step": 10353 }, { "epoch": 15.5, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1566, "step": 10354 }, { "epoch": 15.501497005988025, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.154, "step": 10355 }, { "epoch": 15.502994011976048, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.155, "step": 10356 }, { "epoch": 15.504491017964071, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1457, "step": 10357 }, { "epoch": 15.505988023952096, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1486, "step": 10358 }, { "epoch": 15.50748502994012, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1554, "step": 10359 }, { "epoch": 15.508982035928144, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1525, "step": 10360 }, { "epoch": 15.510479041916168, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1507, "step": 10361 }, { "epoch": 15.511976047904191, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1575, "step": 10362 }, { "epoch": 15.513473053892216, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.155, "step": 10363 }, { "epoch": 15.51497005988024, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1564, "step": 10364 }, { "epoch": 15.516467065868264, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.152, "step": 10365 }, { "epoch": 15.517964071856287, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1519, "step": 10366 }, { "epoch": 15.519461077844312, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1526, "step": 10367 }, { "epoch": 15.520958083832335, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1488, "step": 10368 }, { "epoch": 15.52245508982036, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1573, "step": 10369 }, { "epoch": 15.523952095808383, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1499, "step": 10370 }, { "epoch": 15.525449101796408, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1482, "step": 10371 }, { "epoch": 15.52694610778443, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1481, "step": 10372 }, { "epoch": 15.528443113772456, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1562, "step": 10373 }, { "epoch": 15.529940119760479, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1533, "step": 10374 }, { "epoch": 15.531437125748504, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.143, "step": 10375 }, { "epoch": 15.532934131736527, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1512, "step": 10376 }, { "epoch": 15.534431137724551, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1476, "step": 10377 }, { "epoch": 15.535928143712574, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1492, "step": 10378 }, { "epoch": 15.5374251497006, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1475, "step": 10379 }, { "epoch": 15.538922155688622, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1484, "step": 10380 }, { "epoch": 15.540419161676647, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1562, "step": 10381 }, { "epoch": 15.54191616766467, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1464, "step": 10382 }, { "epoch": 15.543413173652695, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1474, "step": 10383 }, { "epoch": 15.544910179640718, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1546, "step": 10384 }, { "epoch": 15.546407185628743, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1411, "step": 10385 }, { "epoch": 15.547904191616766, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1474, "step": 10386 }, { "epoch": 15.54940119760479, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1559, "step": 10387 }, { "epoch": 15.550898203592814, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1433, "step": 10388 }, { "epoch": 15.552395209580839, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1505, "step": 10389 }, { "epoch": 15.553892215568862, "grad_norm": 0.1630859375, "learning_rate": 0.0008, "loss": 1.1525, "step": 10390 }, { "epoch": 15.555389221556887, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1492, "step": 10391 }, { "epoch": 15.55688622754491, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1506, "step": 10392 }, { "epoch": 15.558383233532934, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1434, "step": 10393 }, { "epoch": 15.559880239520957, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.1532, "step": 10394 }, { "epoch": 15.561377245508982, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1472, "step": 10395 }, { "epoch": 15.562874251497005, "grad_norm": 0.1884765625, "learning_rate": 0.0008, "loss": 1.1519, "step": 10396 }, { "epoch": 15.56437125748503, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1481, "step": 10397 }, { "epoch": 15.565868263473053, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.153, "step": 10398 }, { "epoch": 15.567365269461078, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1453, "step": 10399 }, { "epoch": 15.568862275449101, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.1464, "step": 10400 }, { "epoch": 15.570359281437126, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1493, "step": 10401 }, { "epoch": 15.571856287425149, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1427, "step": 10402 }, { "epoch": 15.573353293413174, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1519, "step": 10403 }, { "epoch": 15.574850299401197, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1465, "step": 10404 }, { "epoch": 15.576347305389222, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.148, "step": 10405 }, { "epoch": 15.577844311377245, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1472, "step": 10406 }, { "epoch": 15.57934131736527, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1467, "step": 10407 }, { "epoch": 15.580838323353294, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1441, "step": 10408 }, { "epoch": 15.582335329341317, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1511, "step": 10409 }, { "epoch": 15.58383233532934, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.1476, "step": 10410 }, { "epoch": 15.585329341317365, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1513, "step": 10411 }, { "epoch": 15.58682634730539, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1509, "step": 10412 }, { "epoch": 15.588323353293413, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1504, "step": 10413 }, { "epoch": 15.589820359281438, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1544, "step": 10414 }, { "epoch": 15.591317365269461, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.151, "step": 10415 }, { "epoch": 15.592814371257486, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1552, "step": 10416 }, { "epoch": 15.594311377245509, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1474, "step": 10417 }, { "epoch": 15.595808383233534, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1487, "step": 10418 }, { "epoch": 15.597305389221557, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1431, "step": 10419 }, { "epoch": 15.598802395209582, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1476, "step": 10420 }, { "epoch": 15.600299401197605, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1438, "step": 10421 }, { "epoch": 15.60179640718563, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1488, "step": 10422 }, { "epoch": 15.603293413173652, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.153, "step": 10423 }, { "epoch": 15.604790419161677, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1501, "step": 10424 }, { "epoch": 15.6062874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1453, "step": 10425 }, { "epoch": 15.607784431137725, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1452, "step": 10426 }, { "epoch": 15.609281437125748, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1471, "step": 10427 }, { "epoch": 15.610778443113773, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1495, "step": 10428 }, { "epoch": 15.612275449101796, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1411, "step": 10429 }, { "epoch": 15.613772455089821, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.157, "step": 10430 }, { "epoch": 15.615269461077844, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1513, "step": 10431 }, { "epoch": 15.616766467065869, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.151, "step": 10432 }, { "epoch": 15.618263473053892, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1431, "step": 10433 }, { "epoch": 15.619760479041917, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1483, "step": 10434 }, { "epoch": 15.62125748502994, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.1463, "step": 10435 }, { "epoch": 15.622754491017965, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1538, "step": 10436 }, { "epoch": 15.624251497005988, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1529, "step": 10437 }, { "epoch": 15.625748502994012, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1431, "step": 10438 }, { "epoch": 15.627245508982035, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.144, "step": 10439 }, { "epoch": 15.62874251497006, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1475, "step": 10440 }, { "epoch": 15.630239520958083, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1427, "step": 10441 }, { "epoch": 15.631736526946108, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.148, "step": 10442 }, { "epoch": 15.633233532934131, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1465, "step": 10443 }, { "epoch": 15.634730538922156, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1437, "step": 10444 }, { "epoch": 15.636227544910179, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1385, "step": 10445 }, { "epoch": 15.637724550898204, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1467, "step": 10446 }, { "epoch": 15.639221556886227, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1459, "step": 10447 }, { "epoch": 15.640718562874252, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1487, "step": 10448 }, { "epoch": 15.642215568862275, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1483, "step": 10449 }, { "epoch": 15.6437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1457, "step": 10450 }, { "epoch": 15.645209580838323, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1418, "step": 10451 }, { "epoch": 15.646706586826348, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1379, "step": 10452 }, { "epoch": 15.64820359281437, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1488, "step": 10453 }, { "epoch": 15.649700598802395, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1457, "step": 10454 }, { "epoch": 15.651197604790418, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1458, "step": 10455 }, { "epoch": 15.652694610778443, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1448, "step": 10456 }, { "epoch": 15.654191616766466, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1507, "step": 10457 }, { "epoch": 15.655688622754491, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1427, "step": 10458 }, { "epoch": 15.657185628742514, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1534, "step": 10459 }, { "epoch": 15.658682634730539, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1493, "step": 10460 }, { "epoch": 15.660179640718562, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1466, "step": 10461 }, { "epoch": 15.661676646706587, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1468, "step": 10462 }, { "epoch": 15.66317365269461, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1452, "step": 10463 }, { "epoch": 15.664670658682635, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1491, "step": 10464 }, { "epoch": 15.66616766467066, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1479, "step": 10465 }, { "epoch": 15.667664670658683, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1441, "step": 10466 }, { "epoch": 15.669161676646706, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1496, "step": 10467 }, { "epoch": 15.67065868263473, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1452, "step": 10468 }, { "epoch": 15.672155688622755, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1437, "step": 10469 }, { "epoch": 15.673652694610778, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1394, "step": 10470 }, { "epoch": 15.675149700598803, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1458, "step": 10471 }, { "epoch": 15.676646706586826, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.151, "step": 10472 }, { "epoch": 15.678143712574851, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1455, "step": 10473 }, { "epoch": 15.679640718562874, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1443, "step": 10474 }, { "epoch": 15.681137724550899, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1401, "step": 10475 }, { "epoch": 15.682634730538922, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1457, "step": 10476 }, { "epoch": 15.684131736526947, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1479, "step": 10477 }, { "epoch": 15.68562874251497, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1421, "step": 10478 }, { "epoch": 15.687125748502995, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1446, "step": 10479 }, { "epoch": 15.688622754491018, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1518, "step": 10480 }, { "epoch": 15.690119760479043, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1494, "step": 10481 }, { "epoch": 15.691616766467066, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1452, "step": 10482 }, { "epoch": 15.69311377245509, "grad_norm": 0.193359375, "learning_rate": 0.0008, "loss": 1.1471, "step": 10483 }, { "epoch": 15.694610778443113, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1422, "step": 10484 }, { "epoch": 15.696107784431138, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.1413, "step": 10485 }, { "epoch": 15.697604790419161, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1405, "step": 10486 }, { "epoch": 15.699101796407186, "grad_norm": 0.1572265625, "learning_rate": 0.0008, "loss": 1.1456, "step": 10487 }, { "epoch": 15.70059880239521, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1536, "step": 10488 }, { "epoch": 15.702095808383234, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.144, "step": 10489 }, { "epoch": 15.703592814371257, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.15, "step": 10490 }, { "epoch": 15.705089820359282, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1397, "step": 10491 }, { "epoch": 15.706586826347305, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1479, "step": 10492 }, { "epoch": 15.70808383233533, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1458, "step": 10493 }, { "epoch": 15.709580838323353, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1487, "step": 10494 }, { "epoch": 15.711077844311378, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1413, "step": 10495 }, { "epoch": 15.7125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.14, "step": 10496 }, { "epoch": 15.714071856287426, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1449, "step": 10497 }, { "epoch": 15.715568862275449, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1458, "step": 10498 }, { "epoch": 15.717065868263473, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1502, "step": 10499 }, { "epoch": 15.718562874251496, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1477, "step": 10500 }, { "epoch": 15.720059880239521, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1439, "step": 10501 }, { "epoch": 15.721556886227544, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1439, "step": 10502 }, { "epoch": 15.72305389221557, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1473, "step": 10503 }, { "epoch": 15.724550898203592, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1467, "step": 10504 }, { "epoch": 15.726047904191617, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1422, "step": 10505 }, { "epoch": 15.72754491017964, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1403, "step": 10506 }, { "epoch": 15.729041916167665, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1448, "step": 10507 }, { "epoch": 15.730538922155688, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1419, "step": 10508 }, { "epoch": 15.732035928143713, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1428, "step": 10509 }, { "epoch": 15.733532934131736, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1477, "step": 10510 }, { "epoch": 15.73502994011976, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1471, "step": 10511 }, { "epoch": 15.736526946107784, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1534, "step": 10512 }, { "epoch": 15.738023952095809, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1514, "step": 10513 }, { "epoch": 15.739520958083832, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1449, "step": 10514 }, { "epoch": 15.741017964071856, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1422, "step": 10515 }, { "epoch": 15.74251497005988, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1485, "step": 10516 }, { "epoch": 15.744011976047904, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1467, "step": 10517 }, { "epoch": 15.745508982035929, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1424, "step": 10518 }, { "epoch": 15.747005988023952, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.138, "step": 10519 }, { "epoch": 15.748502994011975, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1437, "step": 10520 }, { "epoch": 15.75, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1447, "step": 10521 }, { "epoch": 15.751497005988025, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1414, "step": 10522 }, { "epoch": 15.752994011976048, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1486, "step": 10523 }, { "epoch": 15.754491017964071, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1414, "step": 10524 }, { "epoch": 15.755988023952096, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1365, "step": 10525 }, { "epoch": 15.75748502994012, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1507, "step": 10526 }, { "epoch": 15.758982035928144, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1415, "step": 10527 }, { "epoch": 15.760479041916168, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1464, "step": 10528 }, { "epoch": 15.761976047904191, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1475, "step": 10529 }, { "epoch": 15.763473053892216, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1485, "step": 10530 }, { "epoch": 15.76497005988024, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1409, "step": 10531 }, { "epoch": 15.766467065868264, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1464, "step": 10532 }, { "epoch": 15.767964071856287, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 10533 }, { "epoch": 15.769461077844312, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1496, "step": 10534 }, { "epoch": 15.770958083832335, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1402, "step": 10535 }, { "epoch": 15.77245508982036, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1477, "step": 10536 }, { "epoch": 15.773952095808383, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.144, "step": 10537 }, { "epoch": 15.775449101796408, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1485, "step": 10538 }, { "epoch": 15.77694610778443, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.142, "step": 10539 }, { "epoch": 15.778443113772456, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1416, "step": 10540 }, { "epoch": 15.779940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1424, "step": 10541 }, { "epoch": 15.781437125748504, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 10542 }, { "epoch": 15.782934131736527, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1454, "step": 10543 }, { "epoch": 15.784431137724551, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1494, "step": 10544 }, { "epoch": 15.785928143712574, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1421, "step": 10545 }, { "epoch": 15.7874251497006, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1412, "step": 10546 }, { "epoch": 15.788922155688622, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1499, "step": 10547 }, { "epoch": 15.790419161676647, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1399, "step": 10548 }, { "epoch": 15.79191616766467, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.146, "step": 10549 }, { "epoch": 15.793413173652695, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1474, "step": 10550 }, { "epoch": 15.794910179640718, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1454, "step": 10551 }, { "epoch": 15.796407185628743, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1452, "step": 10552 }, { "epoch": 15.797904191616766, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1505, "step": 10553 }, { "epoch": 15.79940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1438, "step": 10554 }, { "epoch": 15.800898203592814, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1446, "step": 10555 }, { "epoch": 15.802395209580839, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1507, "step": 10556 }, { "epoch": 15.803892215568862, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1467, "step": 10557 }, { "epoch": 15.805389221556887, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1439, "step": 10558 }, { "epoch": 15.80688622754491, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1466, "step": 10559 }, { "epoch": 15.808383233532934, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.148, "step": 10560 }, { "epoch": 15.809880239520957, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1465, "step": 10561 }, { "epoch": 15.811377245508982, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.141, "step": 10562 }, { "epoch": 15.812874251497005, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 10563 }, { "epoch": 15.81437125748503, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1439, "step": 10564 }, { "epoch": 15.815868263473053, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1533, "step": 10565 }, { "epoch": 15.817365269461078, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1393, "step": 10566 }, { "epoch": 15.818862275449101, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1441, "step": 10567 }, { "epoch": 15.820359281437126, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1385, "step": 10568 }, { "epoch": 15.821856287425149, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1419, "step": 10569 }, { "epoch": 15.823353293413174, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1478, "step": 10570 }, { "epoch": 15.824850299401197, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1389, "step": 10571 }, { "epoch": 15.826347305389222, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1359, "step": 10572 }, { "epoch": 15.827844311377245, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1461, "step": 10573 }, { "epoch": 15.82934131736527, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 10574 }, { "epoch": 15.830838323353294, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.144, "step": 10575 }, { "epoch": 15.832335329341317, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1383, "step": 10576 }, { "epoch": 15.83383233532934, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1435, "step": 10577 }, { "epoch": 15.835329341317365, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1417, "step": 10578 }, { "epoch": 15.83682634730539, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1489, "step": 10579 }, { "epoch": 15.838323353293413, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1457, "step": 10580 }, { "epoch": 15.839820359281438, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1444, "step": 10581 }, { "epoch": 15.841317365269461, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.14, "step": 10582 }, { "epoch": 15.842814371257486, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1501, "step": 10583 }, { "epoch": 15.844311377245509, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.145, "step": 10584 }, { "epoch": 15.845808383233534, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1432, "step": 10585 }, { "epoch": 15.847305389221557, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1445, "step": 10586 }, { "epoch": 15.848802395209582, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1474, "step": 10587 }, { "epoch": 15.850299401197605, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1506, "step": 10588 }, { "epoch": 15.85179640718563, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1476, "step": 10589 }, { "epoch": 15.853293413173652, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1415, "step": 10590 }, { "epoch": 15.854790419161677, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1436, "step": 10591 }, { "epoch": 15.8562874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1389, "step": 10592 }, { "epoch": 15.857784431137725, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.149, "step": 10593 }, { "epoch": 15.859281437125748, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1458, "step": 10594 }, { "epoch": 15.860778443113773, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1417, "step": 10595 }, { "epoch": 15.862275449101796, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1472, "step": 10596 }, { "epoch": 15.863772455089821, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1382, "step": 10597 }, { "epoch": 15.865269461077844, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1476, "step": 10598 }, { "epoch": 15.866766467065869, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1405, "step": 10599 }, { "epoch": 15.868263473053892, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.145, "step": 10600 }, { "epoch": 15.869760479041917, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1487, "step": 10601 }, { "epoch": 15.87125748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1436, "step": 10602 }, { "epoch": 15.872754491017965, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1502, "step": 10603 }, { "epoch": 15.874251497005988, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1456, "step": 10604 }, { "epoch": 15.875748502994012, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1485, "step": 10605 }, { "epoch": 15.877245508982035, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1384, "step": 10606 }, { "epoch": 15.87874251497006, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1451, "step": 10607 }, { "epoch": 15.880239520958083, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1447, "step": 10608 }, { "epoch": 15.881736526946108, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 10609 }, { "epoch": 15.883233532934131, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1382, "step": 10610 }, { "epoch": 15.884730538922156, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1435, "step": 10611 }, { "epoch": 15.886227544910179, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1397, "step": 10612 }, { "epoch": 15.887724550898204, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1476, "step": 10613 }, { "epoch": 15.889221556886227, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 10614 }, { "epoch": 15.890718562874252, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1405, "step": 10615 }, { "epoch": 15.892215568862275, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1447, "step": 10616 }, { "epoch": 15.8937125748503, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1406, "step": 10617 }, { "epoch": 15.895209580838323, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1403, "step": 10618 }, { "epoch": 15.896706586826348, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1463, "step": 10619 }, { "epoch": 15.89820359281437, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1433, "step": 10620 }, { "epoch": 15.899700598802395, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1395, "step": 10621 }, { "epoch": 15.901197604790418, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1409, "step": 10622 }, { "epoch": 15.902694610778443, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1541, "step": 10623 }, { "epoch": 15.904191616766466, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1415, "step": 10624 }, { "epoch": 15.905688622754491, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1414, "step": 10625 }, { "epoch": 15.907185628742514, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1463, "step": 10626 }, { "epoch": 15.908682634730539, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1441, "step": 10627 }, { "epoch": 15.910179640718562, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1447, "step": 10628 }, { "epoch": 15.911676646706587, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1458, "step": 10629 }, { "epoch": 15.91317365269461, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1417, "step": 10630 }, { "epoch": 15.914670658682635, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1402, "step": 10631 }, { "epoch": 15.91616766467066, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1455, "step": 10632 }, { "epoch": 15.917664670658683, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1464, "step": 10633 }, { "epoch": 15.919161676646706, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1459, "step": 10634 }, { "epoch": 15.92065868263473, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1429, "step": 10635 }, { "epoch": 15.922155688622755, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1394, "step": 10636 }, { "epoch": 15.923652694610778, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1403, "step": 10637 }, { "epoch": 15.925149700598803, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1402, "step": 10638 }, { "epoch": 15.926646706586826, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1415, "step": 10639 }, { "epoch": 15.928143712574851, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1451, "step": 10640 }, { "epoch": 15.929640718562874, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1415, "step": 10641 }, { "epoch": 15.931137724550899, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1486, "step": 10642 }, { "epoch": 15.932634730538922, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.142, "step": 10643 }, { "epoch": 15.934131736526947, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1456, "step": 10644 }, { "epoch": 15.93562874251497, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1415, "step": 10645 }, { "epoch": 15.937125748502995, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1392, "step": 10646 }, { "epoch": 15.938622754491018, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.145, "step": 10647 }, { "epoch": 15.940119760479043, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1437, "step": 10648 }, { "epoch": 15.941616766467066, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.146, "step": 10649 }, { "epoch": 15.94311377245509, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1403, "step": 10650 }, { "epoch": 15.944610778443113, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1396, "step": 10651 }, { "epoch": 15.946107784431138, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1418, "step": 10652 }, { "epoch": 15.947604790419161, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1378, "step": 10653 }, { "epoch": 15.949101796407186, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1401, "step": 10654 }, { "epoch": 15.95059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1404, "step": 10655 }, { "epoch": 15.952095808383234, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1467, "step": 10656 }, { "epoch": 15.953592814371257, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1413, "step": 10657 }, { "epoch": 15.955089820359282, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1416, "step": 10658 }, { "epoch": 15.956586826347305, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1448, "step": 10659 }, { "epoch": 15.95808383233533, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1424, "step": 10660 }, { "epoch": 15.959580838323353, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1465, "step": 10661 }, { "epoch": 15.961077844311378, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1386, "step": 10662 }, { "epoch": 15.9625748502994, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1406, "step": 10663 }, { "epoch": 15.964071856287426, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1465, "step": 10664 }, { "epoch": 15.965568862275449, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.15, "step": 10665 }, { "epoch": 15.967065868263473, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1451, "step": 10666 }, { "epoch": 15.968562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1533, "step": 10667 }, { "epoch": 15.970059880239521, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1398, "step": 10668 }, { "epoch": 15.971556886227544, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1452, "step": 10669 }, { "epoch": 15.97305389221557, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1424, "step": 10670 }, { "epoch": 15.974550898203592, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1396, "step": 10671 }, { "epoch": 15.976047904191617, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1436, "step": 10672 }, { "epoch": 15.97754491017964, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1422, "step": 10673 }, { "epoch": 15.979041916167665, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1449, "step": 10674 }, { "epoch": 15.980538922155688, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1474, "step": 10675 }, { "epoch": 15.982035928143713, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1429, "step": 10676 }, { "epoch": 15.983532934131736, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1468, "step": 10677 }, { "epoch": 15.98502994011976, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 10678 }, { "epoch": 15.986526946107784, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1439, "step": 10679 }, { "epoch": 15.988023952095809, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1412, "step": 10680 }, { "epoch": 15.989520958083832, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1386, "step": 10681 }, { "epoch": 15.991017964071856, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1407, "step": 10682 }, { "epoch": 15.99251497005988, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1413, "step": 10683 }, { "epoch": 15.994011976047904, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1432, "step": 10684 }, { "epoch": 15.995508982035929, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1387, "step": 10685 }, { "epoch": 15.997005988023952, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 10686 }, { "epoch": 15.998502994011975, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1474, "step": 10687 }, { "epoch": 16.0, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1321, "step": 10688 }, { "epoch": 16.001497005988025, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1448, "step": 10689 }, { "epoch": 16.00299401197605, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1479, "step": 10690 }, { "epoch": 16.00449101796407, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1438, "step": 10691 }, { "epoch": 16.005988023952096, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 10692 }, { "epoch": 16.00748502994012, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1416, "step": 10693 }, { "epoch": 16.008982035928145, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1441, "step": 10694 }, { "epoch": 16.010479041916167, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1349, "step": 10695 }, { "epoch": 16.01197604790419, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1458, "step": 10696 }, { "epoch": 16.013473053892216, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1412, "step": 10697 }, { "epoch": 16.01497005988024, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1404, "step": 10698 }, { "epoch": 16.016467065868262, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1421, "step": 10699 }, { "epoch": 16.017964071856287, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1451, "step": 10700 }, { "epoch": 16.019461077844312, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1326, "step": 10701 }, { "epoch": 16.020958083832337, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1461, "step": 10702 }, { "epoch": 16.022455089820358, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1412, "step": 10703 }, { "epoch": 16.023952095808383, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1389, "step": 10704 }, { "epoch": 16.025449101796408, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1423, "step": 10705 }, { "epoch": 16.026946107784433, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 10706 }, { "epoch": 16.028443113772454, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1456, "step": 10707 }, { "epoch": 16.02994011976048, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1443, "step": 10708 }, { "epoch": 16.031437125748504, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1365, "step": 10709 }, { "epoch": 16.03293413173653, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1433, "step": 10710 }, { "epoch": 16.03443113772455, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1467, "step": 10711 }, { "epoch": 16.035928143712574, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1368, "step": 10712 }, { "epoch": 16.0374251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1381, "step": 10713 }, { "epoch": 16.038922155688624, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1377, "step": 10714 }, { "epoch": 16.040419161676645, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1397, "step": 10715 }, { "epoch": 16.04191616766467, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1478, "step": 10716 }, { "epoch": 16.043413173652695, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1367, "step": 10717 }, { "epoch": 16.04491017964072, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 10718 }, { "epoch": 16.04640718562874, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1407, "step": 10719 }, { "epoch": 16.047904191616766, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 10720 }, { "epoch": 16.04940119760479, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1515, "step": 10721 }, { "epoch": 16.050898203592816, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1311, "step": 10722 }, { "epoch": 16.052395209580837, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1478, "step": 10723 }, { "epoch": 16.05389221556886, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.144, "step": 10724 }, { "epoch": 16.055389221556887, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1446, "step": 10725 }, { "epoch": 16.05688622754491, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.14, "step": 10726 }, { "epoch": 16.058383233532933, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1355, "step": 10727 }, { "epoch": 16.059880239520957, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1506, "step": 10728 }, { "epoch": 16.061377245508982, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1406, "step": 10729 }, { "epoch": 16.062874251497007, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1409, "step": 10730 }, { "epoch": 16.06437125748503, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.138, "step": 10731 }, { "epoch": 16.065868263473053, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1374, "step": 10732 }, { "epoch": 16.067365269461078, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1445, "step": 10733 }, { "epoch": 16.068862275449103, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1405, "step": 10734 }, { "epoch": 16.070359281437124, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1448, "step": 10735 }, { "epoch": 16.07185628742515, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1468, "step": 10736 }, { "epoch": 16.073353293413174, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1402, "step": 10737 }, { "epoch": 16.0748502994012, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1392, "step": 10738 }, { "epoch": 16.07634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1448, "step": 10739 }, { "epoch": 16.077844311377245, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1392, "step": 10740 }, { "epoch": 16.07934131736527, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1496, "step": 10741 }, { "epoch": 16.080838323353294, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1402, "step": 10742 }, { "epoch": 16.082335329341316, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 10743 }, { "epoch": 16.08383233532934, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1439, "step": 10744 }, { "epoch": 16.085329341317365, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1489, "step": 10745 }, { "epoch": 16.08682634730539, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.138, "step": 10746 }, { "epoch": 16.088323353293415, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 10747 }, { "epoch": 16.089820359281436, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1441, "step": 10748 }, { "epoch": 16.09131736526946, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 10749 }, { "epoch": 16.092814371257486, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.144, "step": 10750 }, { "epoch": 16.09431137724551, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1432, "step": 10751 }, { "epoch": 16.095808383233532, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 10752 }, { "epoch": 16.097305389221557, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1472, "step": 10753 }, { "epoch": 16.09880239520958, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.139, "step": 10754 }, { "epoch": 16.100299401197606, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1392, "step": 10755 }, { "epoch": 16.101796407185628, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.141, "step": 10756 }, { "epoch": 16.103293413173652, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1406, "step": 10757 }, { "epoch": 16.104790419161677, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1428, "step": 10758 }, { "epoch": 16.106287425149702, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 10759 }, { "epoch": 16.107784431137723, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1398, "step": 10760 }, { "epoch": 16.10928143712575, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 10761 }, { "epoch": 16.110778443113773, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1336, "step": 10762 }, { "epoch": 16.112275449101798, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1359, "step": 10763 }, { "epoch": 16.11377245508982, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1391, "step": 10764 }, { "epoch": 16.115269461077844, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1373, "step": 10765 }, { "epoch": 16.11676646706587, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 10766 }, { "epoch": 16.118263473053894, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1413, "step": 10767 }, { "epoch": 16.119760479041915, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.138, "step": 10768 }, { "epoch": 16.12125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1433, "step": 10769 }, { "epoch": 16.122754491017965, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1449, "step": 10770 }, { "epoch": 16.12425149700599, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.144, "step": 10771 }, { "epoch": 16.12574850299401, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.138, "step": 10772 }, { "epoch": 16.127245508982035, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.147, "step": 10773 }, { "epoch": 16.12874251497006, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1412, "step": 10774 }, { "epoch": 16.130239520958085, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1498, "step": 10775 }, { "epoch": 16.131736526946106, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1453, "step": 10776 }, { "epoch": 16.13323353293413, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1427, "step": 10777 }, { "epoch": 16.134730538922156, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1375, "step": 10778 }, { "epoch": 16.13622754491018, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1417, "step": 10779 }, { "epoch": 16.137724550898202, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1387, "step": 10780 }, { "epoch": 16.139221556886227, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1418, "step": 10781 }, { "epoch": 16.14071856287425, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 10782 }, { "epoch": 16.142215568862277, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1365, "step": 10783 }, { "epoch": 16.143712574850298, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1415, "step": 10784 }, { "epoch": 16.145209580838323, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1421, "step": 10785 }, { "epoch": 16.146706586826348, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1426, "step": 10786 }, { "epoch": 16.148203592814372, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1434, "step": 10787 }, { "epoch": 16.149700598802394, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1425, "step": 10788 }, { "epoch": 16.15119760479042, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1405, "step": 10789 }, { "epoch": 16.152694610778443, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1359, "step": 10790 }, { "epoch": 16.154191616766468, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1385, "step": 10791 }, { "epoch": 16.15568862275449, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.149, "step": 10792 }, { "epoch": 16.157185628742514, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 10793 }, { "epoch": 16.15868263473054, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1379, "step": 10794 }, { "epoch": 16.160179640718564, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1367, "step": 10795 }, { "epoch": 16.161676646706585, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.137, "step": 10796 }, { "epoch": 16.16317365269461, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1392, "step": 10797 }, { "epoch": 16.164670658682635, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1403, "step": 10798 }, { "epoch": 16.16616766467066, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1461, "step": 10799 }, { "epoch": 16.16766467065868, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.144, "step": 10800 }, { "epoch": 16.169161676646706, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1455, "step": 10801 }, { "epoch": 16.17065868263473, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1462, "step": 10802 }, { "epoch": 16.172155688622755, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1431, "step": 10803 }, { "epoch": 16.17365269461078, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 10804 }, { "epoch": 16.1751497005988, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 10805 }, { "epoch": 16.176646706586826, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1411, "step": 10806 }, { "epoch": 16.17814371257485, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1404, "step": 10807 }, { "epoch": 16.179640718562876, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1462, "step": 10808 }, { "epoch": 16.181137724550897, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1365, "step": 10809 }, { "epoch": 16.182634730538922, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1456, "step": 10810 }, { "epoch": 16.184131736526947, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1396, "step": 10811 }, { "epoch": 16.18562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1328, "step": 10812 }, { "epoch": 16.187125748502993, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1422, "step": 10813 }, { "epoch": 16.188622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1459, "step": 10814 }, { "epoch": 16.190119760479043, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1377, "step": 10815 }, { "epoch": 16.191616766467067, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1434, "step": 10816 }, { "epoch": 16.19311377245509, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1438, "step": 10817 }, { "epoch": 16.194610778443113, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1421, "step": 10818 }, { "epoch": 16.19610778443114, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1432, "step": 10819 }, { "epoch": 16.197604790419163, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1334, "step": 10820 }, { "epoch": 16.199101796407184, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1375, "step": 10821 }, { "epoch": 16.20059880239521, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1469, "step": 10822 }, { "epoch": 16.202095808383234, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1396, "step": 10823 }, { "epoch": 16.20359281437126, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1413, "step": 10824 }, { "epoch": 16.20508982035928, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1357, "step": 10825 }, { "epoch": 16.206586826347305, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1424, "step": 10826 }, { "epoch": 16.20808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1381, "step": 10827 }, { "epoch": 16.209580838323355, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1425, "step": 10828 }, { "epoch": 16.211077844311376, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1414, "step": 10829 }, { "epoch": 16.2125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.141, "step": 10830 }, { "epoch": 16.214071856287426, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1378, "step": 10831 }, { "epoch": 16.21556886227545, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1392, "step": 10832 }, { "epoch": 16.21706586826347, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1394, "step": 10833 }, { "epoch": 16.218562874251496, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1386, "step": 10834 }, { "epoch": 16.22005988023952, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1452, "step": 10835 }, { "epoch": 16.221556886227546, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1364, "step": 10836 }, { "epoch": 16.223053892215567, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 10837 }, { "epoch": 16.224550898203592, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1399, "step": 10838 }, { "epoch": 16.226047904191617, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1431, "step": 10839 }, { "epoch": 16.227544910179642, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1415, "step": 10840 }, { "epoch": 16.229041916167663, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1398, "step": 10841 }, { "epoch": 16.230538922155688, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1439, "step": 10842 }, { "epoch": 16.232035928143713, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 10843 }, { "epoch": 16.233532934131738, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1428, "step": 10844 }, { "epoch": 16.23502994011976, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1428, "step": 10845 }, { "epoch": 16.236526946107784, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 10846 }, { "epoch": 16.23802395209581, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1382, "step": 10847 }, { "epoch": 16.239520958083833, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 10848 }, { "epoch": 16.241017964071855, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1383, "step": 10849 }, { "epoch": 16.24251497005988, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1428, "step": 10850 }, { "epoch": 16.244011976047904, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1427, "step": 10851 }, { "epoch": 16.24550898203593, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1366, "step": 10852 }, { "epoch": 16.24700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1395, "step": 10853 }, { "epoch": 16.248502994011975, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.142, "step": 10854 }, { "epoch": 16.25, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1414, "step": 10855 }, { "epoch": 16.251497005988025, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1407, "step": 10856 }, { "epoch": 16.25299401197605, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.143, "step": 10857 }, { "epoch": 16.25449101796407, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1436, "step": 10858 }, { "epoch": 16.255988023952096, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1441, "step": 10859 }, { "epoch": 16.25748502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1414, "step": 10860 }, { "epoch": 16.258982035928145, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1419, "step": 10861 }, { "epoch": 16.260479041916167, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1445, "step": 10862 }, { "epoch": 16.26197604790419, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1452, "step": 10863 }, { "epoch": 16.263473053892216, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1428, "step": 10864 }, { "epoch": 16.26497005988024, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1461, "step": 10865 }, { "epoch": 16.266467065868262, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1398, "step": 10866 }, { "epoch": 16.267964071856287, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1433, "step": 10867 }, { "epoch": 16.269461077844312, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1384, "step": 10868 }, { "epoch": 16.270958083832337, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1386, "step": 10869 }, { "epoch": 16.272455089820358, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1432, "step": 10870 }, { "epoch": 16.273952095808383, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1323, "step": 10871 }, { "epoch": 16.275449101796408, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.139, "step": 10872 }, { "epoch": 16.276946107784433, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1387, "step": 10873 }, { "epoch": 16.278443113772454, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1439, "step": 10874 }, { "epoch": 16.27994011976048, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1432, "step": 10875 }, { "epoch": 16.281437125748504, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1375, "step": 10876 }, { "epoch": 16.28293413173653, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.143, "step": 10877 }, { "epoch": 16.28443113772455, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1443, "step": 10878 }, { "epoch": 16.285928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1379, "step": 10879 }, { "epoch": 16.2874251497006, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1429, "step": 10880 }, { "epoch": 16.288922155688624, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 10881 }, { "epoch": 16.290419161676645, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1431, "step": 10882 }, { "epoch": 16.29191616766467, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1411, "step": 10883 }, { "epoch": 16.293413173652695, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1387, "step": 10884 }, { "epoch": 16.29491017964072, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1463, "step": 10885 }, { "epoch": 16.29640718562874, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1456, "step": 10886 }, { "epoch": 16.297904191616766, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1412, "step": 10887 }, { "epoch": 16.29940119760479, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1405, "step": 10888 }, { "epoch": 16.300898203592816, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1389, "step": 10889 }, { "epoch": 16.302395209580837, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1403, "step": 10890 }, { "epoch": 16.30389221556886, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.136, "step": 10891 }, { "epoch": 16.305389221556887, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1428, "step": 10892 }, { "epoch": 16.30688622754491, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.139, "step": 10893 }, { "epoch": 16.308383233532933, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1366, "step": 10894 }, { "epoch": 16.309880239520957, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1415, "step": 10895 }, { "epoch": 16.311377245508982, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1383, "step": 10896 }, { "epoch": 16.312874251497007, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1418, "step": 10897 }, { "epoch": 16.31437125748503, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 10898 }, { "epoch": 16.315868263473053, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1465, "step": 10899 }, { "epoch": 16.317365269461078, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1322, "step": 10900 }, { "epoch": 16.318862275449103, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1456, "step": 10901 }, { "epoch": 16.320359281437124, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1425, "step": 10902 }, { "epoch": 16.32185628742515, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1415, "step": 10903 }, { "epoch": 16.323353293413174, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 10904 }, { "epoch": 16.3248502994012, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 10905 }, { "epoch": 16.32634730538922, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1446, "step": 10906 }, { "epoch": 16.327844311377245, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1373, "step": 10907 }, { "epoch": 16.32934131736527, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1389, "step": 10908 }, { "epoch": 16.330838323353294, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1437, "step": 10909 }, { "epoch": 16.33233532934132, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1382, "step": 10910 }, { "epoch": 16.33383233532934, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1418, "step": 10911 }, { "epoch": 16.335329341317365, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1402, "step": 10912 }, { "epoch": 16.33682634730539, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1366, "step": 10913 }, { "epoch": 16.338323353293415, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1353, "step": 10914 }, { "epoch": 16.339820359281436, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1389, "step": 10915 }, { "epoch": 16.34131736526946, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1375, "step": 10916 }, { "epoch": 16.342814371257486, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1444, "step": 10917 }, { "epoch": 16.34431137724551, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1398, "step": 10918 }, { "epoch": 16.345808383233532, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1431, "step": 10919 }, { "epoch": 16.347305389221557, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1371, "step": 10920 }, { "epoch": 16.34880239520958, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1441, "step": 10921 }, { "epoch": 16.350299401197606, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1459, "step": 10922 }, { "epoch": 16.351796407185628, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1415, "step": 10923 }, { "epoch": 16.353293413173652, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1375, "step": 10924 }, { "epoch": 16.354790419161677, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1433, "step": 10925 }, { "epoch": 16.356287425149702, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 10926 }, { "epoch": 16.357784431137723, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1357, "step": 10927 }, { "epoch": 16.35928143712575, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1507, "step": 10928 }, { "epoch": 16.360778443113773, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1341, "step": 10929 }, { "epoch": 16.362275449101798, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1432, "step": 10930 }, { "epoch": 16.36377245508982, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1373, "step": 10931 }, { "epoch": 16.365269461077844, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1403, "step": 10932 }, { "epoch": 16.36676646706587, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1397, "step": 10933 }, { "epoch": 16.368263473053894, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1324, "step": 10934 }, { "epoch": 16.369760479041915, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1408, "step": 10935 }, { "epoch": 16.37125748502994, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1402, "step": 10936 }, { "epoch": 16.372754491017965, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1441, "step": 10937 }, { "epoch": 16.37425149700599, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1475, "step": 10938 }, { "epoch": 16.37574850299401, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1335, "step": 10939 }, { "epoch": 16.377245508982035, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 10940 }, { "epoch": 16.37874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1394, "step": 10941 }, { "epoch": 16.380239520958085, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1458, "step": 10942 }, { "epoch": 16.381736526946106, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1424, "step": 10943 }, { "epoch": 16.38323353293413, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1408, "step": 10944 }, { "epoch": 16.384730538922156, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.137, "step": 10945 }, { "epoch": 16.38622754491018, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1411, "step": 10946 }, { "epoch": 16.387724550898202, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 10947 }, { "epoch": 16.389221556886227, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1427, "step": 10948 }, { "epoch": 16.39071856287425, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1422, "step": 10949 }, { "epoch": 16.392215568862277, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 10950 }, { "epoch": 16.393712574850298, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.137, "step": 10951 }, { "epoch": 16.395209580838323, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.138, "step": 10952 }, { "epoch": 16.396706586826348, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.135, "step": 10953 }, { "epoch": 16.398203592814372, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1416, "step": 10954 }, { "epoch": 16.399700598802394, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1368, "step": 10955 }, { "epoch": 16.40119760479042, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1437, "step": 10956 }, { "epoch": 16.402694610778443, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1473, "step": 10957 }, { "epoch": 16.404191616766468, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 10958 }, { "epoch": 16.40568862275449, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1449, "step": 10959 }, { "epoch": 16.407185628742514, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1343, "step": 10960 }, { "epoch": 16.40868263473054, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1388, "step": 10961 }, { "epoch": 16.410179640718564, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1428, "step": 10962 }, { "epoch": 16.411676646706585, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1391, "step": 10963 }, { "epoch": 16.41317365269461, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1457, "step": 10964 }, { "epoch": 16.414670658682635, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.136, "step": 10965 }, { "epoch": 16.41616766467066, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1456, "step": 10966 }, { "epoch": 16.41766467065868, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1467, "step": 10967 }, { "epoch": 16.419161676646706, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1455, "step": 10968 }, { "epoch": 16.42065868263473, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1359, "step": 10969 }, { "epoch": 16.422155688622755, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.148, "step": 10970 }, { "epoch": 16.42365269461078, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1379, "step": 10971 }, { "epoch": 16.4251497005988, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.135, "step": 10972 }, { "epoch": 16.426646706586826, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 10973 }, { "epoch": 16.42814371257485, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1383, "step": 10974 }, { "epoch": 16.429640718562876, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 10975 }, { "epoch": 16.431137724550897, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1423, "step": 10976 }, { "epoch": 16.432634730538922, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1449, "step": 10977 }, { "epoch": 16.434131736526947, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1464, "step": 10978 }, { "epoch": 16.43562874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1439, "step": 10979 }, { "epoch": 16.437125748502993, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1469, "step": 10980 }, { "epoch": 16.438622754491018, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.142, "step": 10981 }, { "epoch": 16.440119760479043, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1417, "step": 10982 }, { "epoch": 16.441616766467067, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1452, "step": 10983 }, { "epoch": 16.44311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1384, "step": 10984 }, { "epoch": 16.444610778443113, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.14, "step": 10985 }, { "epoch": 16.44610778443114, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1406, "step": 10986 }, { "epoch": 16.447604790419163, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1361, "step": 10987 }, { "epoch": 16.449101796407184, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 10988 }, { "epoch": 16.45059880239521, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1428, "step": 10989 }, { "epoch": 16.452095808383234, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1384, "step": 10990 }, { "epoch": 16.45359281437126, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1423, "step": 10991 }, { "epoch": 16.45508982035928, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.145, "step": 10992 }, { "epoch": 16.456586826347305, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1333, "step": 10993 }, { "epoch": 16.45808383233533, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 10994 }, { "epoch": 16.459580838323355, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1362, "step": 10995 }, { "epoch": 16.461077844311376, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1435, "step": 10996 }, { "epoch": 16.4625748502994, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1343, "step": 10997 }, { "epoch": 16.464071856287426, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1448, "step": 10998 }, { "epoch": 16.46556886227545, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1449, "step": 10999 }, { "epoch": 16.46706586826347, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1373, "step": 11000 }, { "epoch": 16.468562874251496, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1537, "step": 11001 }, { "epoch": 16.47005988023952, "grad_norm": 0.5078125, "learning_rate": 0.0008, "loss": 1.2339, "step": 11002 }, { "epoch": 16.471556886227546, "grad_norm": 0.8203125, "learning_rate": 0.0008, "loss": 1.2478, "step": 11003 }, { "epoch": 16.473053892215567, "grad_norm": 0.69921875, "learning_rate": 0.0008, "loss": 1.2196, "step": 11004 }, { "epoch": 16.474550898203592, "grad_norm": 1.0390625, "learning_rate": 0.0008, "loss": 1.2444, "step": 11005 }, { "epoch": 16.476047904191617, "grad_norm": 0.9453125, "learning_rate": 0.0008, "loss": 1.2377, "step": 11006 }, { "epoch": 16.477544910179642, "grad_norm": 0.90234375, "learning_rate": 0.0008, "loss": 1.2631, "step": 11007 }, { "epoch": 16.479041916167663, "grad_norm": 0.73046875, "learning_rate": 0.0008, "loss": 1.2588, "step": 11008 }, { "epoch": 16.480538922155688, "grad_norm": 0.9921875, "learning_rate": 0.0008, "loss": 1.283, "step": 11009 }, { "epoch": 16.482035928143713, "grad_norm": 0.86328125, "learning_rate": 0.0008, "loss": 1.2862, "step": 11010 }, { "epoch": 16.483532934131738, "grad_norm": 0.45703125, "learning_rate": 0.0008, "loss": 1.2847, "step": 11011 }, { "epoch": 16.48502994011976, "grad_norm": 0.54296875, "learning_rate": 0.0008, "loss": 1.2827, "step": 11012 }, { "epoch": 16.486526946107784, "grad_norm": 0.90234375, "learning_rate": 0.0008, "loss": 1.2914, "step": 11013 }, { "epoch": 16.48802395209581, "grad_norm": 0.84375, "learning_rate": 0.0008, "loss": 1.2901, "step": 11014 }, { "epoch": 16.489520958083833, "grad_norm": 0.76171875, "learning_rate": 0.0008, "loss": 1.2746, "step": 11015 }, { "epoch": 16.491017964071855, "grad_norm": 0.703125, "learning_rate": 0.0008, "loss": 1.2701, "step": 11016 }, { "epoch": 16.49251497005988, "grad_norm": 1.015625, "learning_rate": 0.0008, "loss": 1.2886, "step": 11017 }, { "epoch": 16.494011976047904, "grad_norm": 1.4609375, "learning_rate": 0.0008, "loss": 1.3552, "step": 11018 }, { "epoch": 16.49550898203593, "grad_norm": 0.72265625, "learning_rate": 0.0008, "loss": 1.2829, "step": 11019 }, { "epoch": 16.49700598802395, "grad_norm": 1.0078125, "learning_rate": 0.0008, "loss": 1.2804, "step": 11020 }, { "epoch": 16.498502994011975, "grad_norm": 0.77734375, "learning_rate": 0.0008, "loss": 1.276, "step": 11021 }, { "epoch": 16.5, "grad_norm": 0.57421875, "learning_rate": 0.0008, "loss": 1.2676, "step": 11022 }, { "epoch": 16.501497005988025, "grad_norm": 0.671875, "learning_rate": 0.0008, "loss": 1.2626, "step": 11023 }, { "epoch": 16.50299401197605, "grad_norm": 0.4140625, "learning_rate": 0.0008, "loss": 1.2453, "step": 11024 }, { "epoch": 16.50449101796407, "grad_norm": 0.484375, "learning_rate": 0.0008, "loss": 1.248, "step": 11025 }, { "epoch": 16.505988023952096, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.2363, "step": 11026 }, { "epoch": 16.50748502994012, "grad_norm": 0.27734375, "learning_rate": 0.0008, "loss": 1.236, "step": 11027 }, { "epoch": 16.508982035928145, "grad_norm": 0.26171875, "learning_rate": 0.0008, "loss": 1.228, "step": 11028 }, { "epoch": 16.510479041916167, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.2281, "step": 11029 }, { "epoch": 16.51197604790419, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.2243, "step": 11030 }, { "epoch": 16.513473053892216, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.2231, "step": 11031 }, { "epoch": 16.51497005988024, "grad_norm": 0.203125, "learning_rate": 0.0008, "loss": 1.2208, "step": 11032 }, { "epoch": 16.516467065868262, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.2153, "step": 11033 }, { "epoch": 16.517964071856287, "grad_norm": 0.396484375, "learning_rate": 0.0008, "loss": 1.2161, "step": 11034 }, { "epoch": 16.519461077844312, "grad_norm": 0.53515625, "learning_rate": 0.0008, "loss": 1.2174, "step": 11035 }, { "epoch": 16.520958083832337, "grad_norm": 0.84375, "learning_rate": 0.0008, "loss": 1.2215, "step": 11036 }, { "epoch": 16.522455089820358, "grad_norm": 0.921875, "learning_rate": 0.0008, "loss": 1.2288, "step": 11037 }, { "epoch": 16.523952095808383, "grad_norm": 0.5703125, "learning_rate": 0.0008, "loss": 1.2081, "step": 11038 }, { "epoch": 16.525449101796408, "grad_norm": 0.3046875, "learning_rate": 0.0008, "loss": 1.2025, "step": 11039 }, { "epoch": 16.526946107784433, "grad_norm": 0.67578125, "learning_rate": 0.0008, "loss": 1.2166, "step": 11040 }, { "epoch": 16.528443113772454, "grad_norm": 0.408203125, "learning_rate": 0.0008, "loss": 1.2124, "step": 11041 }, { "epoch": 16.52994011976048, "grad_norm": 0.408203125, "learning_rate": 0.0008, "loss": 1.2155, "step": 11042 }, { "epoch": 16.531437125748504, "grad_norm": 0.32421875, "learning_rate": 0.0008, "loss": 1.2082, "step": 11043 }, { "epoch": 16.53293413173653, "grad_norm": 0.248046875, "learning_rate": 0.0008, "loss": 1.1996, "step": 11044 }, { "epoch": 16.53443113772455, "grad_norm": 0.26171875, "learning_rate": 0.0008, "loss": 1.191, "step": 11045 }, { "epoch": 16.535928143712574, "grad_norm": 0.2314453125, "learning_rate": 0.0008, "loss": 1.1928, "step": 11046 }, { "epoch": 16.5374251497006, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.1906, "step": 11047 }, { "epoch": 16.538922155688624, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1906, "step": 11048 }, { "epoch": 16.540419161676645, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1896, "step": 11049 }, { "epoch": 16.54191616766467, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1849, "step": 11050 }, { "epoch": 16.543413173652695, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1869, "step": 11051 }, { "epoch": 16.54491017964072, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1876, "step": 11052 }, { "epoch": 16.54640718562874, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1822, "step": 11053 }, { "epoch": 16.547904191616766, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1865, "step": 11054 }, { "epoch": 16.54940119760479, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1717, "step": 11055 }, { "epoch": 16.550898203592816, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.1808, "step": 11056 }, { "epoch": 16.552395209580837, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.175, "step": 11057 }, { "epoch": 16.55389221556886, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1804, "step": 11058 }, { "epoch": 16.555389221556887, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1745, "step": 11059 }, { "epoch": 16.55688622754491, "grad_norm": 0.2021484375, "learning_rate": 0.0008, "loss": 1.1694, "step": 11060 }, { "epoch": 16.558383233532933, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.1768, "step": 11061 }, { "epoch": 16.559880239520957, "grad_norm": 0.205078125, "learning_rate": 0.0008, "loss": 1.171, "step": 11062 }, { "epoch": 16.561377245508982, "grad_norm": 0.291015625, "learning_rate": 0.0008, "loss": 1.1725, "step": 11063 }, { "epoch": 16.562874251497007, "grad_norm": 0.48046875, "learning_rate": 0.0008, "loss": 1.1779, "step": 11064 }, { "epoch": 16.56437125748503, "grad_norm": 0.79296875, "learning_rate": 0.0008, "loss": 1.1795, "step": 11065 }, { "epoch": 16.565868263473053, "grad_norm": 0.8515625, "learning_rate": 0.0008, "loss": 1.1998, "step": 11066 }, { "epoch": 16.567365269461078, "grad_norm": 0.314453125, "learning_rate": 0.0008, "loss": 1.1738, "step": 11067 }, { "epoch": 16.568862275449103, "grad_norm": 0.4375, "learning_rate": 0.0008, "loss": 1.1762, "step": 11068 }, { "epoch": 16.570359281437124, "grad_norm": 0.443359375, "learning_rate": 0.0008, "loss": 1.1845, "step": 11069 }, { "epoch": 16.57185628742515, "grad_norm": 0.330078125, "learning_rate": 0.0008, "loss": 1.1764, "step": 11070 }, { "epoch": 16.573353293413174, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.1823, "step": 11071 }, { "epoch": 16.5748502994012, "grad_norm": 0.208984375, "learning_rate": 0.0008, "loss": 1.1664, "step": 11072 }, { "epoch": 16.57634730538922, "grad_norm": 0.2294921875, "learning_rate": 0.0008, "loss": 1.1715, "step": 11073 }, { "epoch": 16.577844311377245, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1694, "step": 11074 }, { "epoch": 16.57934131736527, "grad_norm": 0.2255859375, "learning_rate": 0.0008, "loss": 1.1633, "step": 11075 }, { "epoch": 16.580838323353294, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1594, "step": 11076 }, { "epoch": 16.58233532934132, "grad_norm": 0.2490234375, "learning_rate": 0.0008, "loss": 1.1658, "step": 11077 }, { "epoch": 16.58383233532934, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1665, "step": 11078 }, { "epoch": 16.585329341317365, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1682, "step": 11079 }, { "epoch": 16.58682634730539, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1667, "step": 11080 }, { "epoch": 16.58832335329341, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1605, "step": 11081 }, { "epoch": 16.589820359281436, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1633, "step": 11082 }, { "epoch": 16.59131736526946, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1626, "step": 11083 }, { "epoch": 16.592814371257486, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1626, "step": 11084 }, { "epoch": 16.59431137724551, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1606, "step": 11085 }, { "epoch": 16.595808383233532, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1572, "step": 11086 }, { "epoch": 16.597305389221557, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1589, "step": 11087 }, { "epoch": 16.59880239520958, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1535, "step": 11088 }, { "epoch": 16.600299401197606, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1563, "step": 11089 }, { "epoch": 16.601796407185628, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1593, "step": 11090 }, { "epoch": 16.603293413173652, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1585, "step": 11091 }, { "epoch": 16.604790419161677, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1593, "step": 11092 }, { "epoch": 16.606287425149702, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.163, "step": 11093 }, { "epoch": 16.607784431137723, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1582, "step": 11094 }, { "epoch": 16.60928143712575, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1586, "step": 11095 }, { "epoch": 16.610778443113773, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1556, "step": 11096 }, { "epoch": 16.612275449101798, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.155, "step": 11097 }, { "epoch": 16.61377245508982, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1638, "step": 11098 }, { "epoch": 16.615269461077844, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1563, "step": 11099 }, { "epoch": 16.61676646706587, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1597, "step": 11100 }, { "epoch": 16.618263473053894, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1577, "step": 11101 }, { "epoch": 16.619760479041915, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1585, "step": 11102 }, { "epoch": 16.62125748502994, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1597, "step": 11103 }, { "epoch": 16.622754491017965, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1537, "step": 11104 }, { "epoch": 16.62425149700599, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1552, "step": 11105 }, { "epoch": 16.62574850299401, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1551, "step": 11106 }, { "epoch": 16.627245508982035, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1477, "step": 11107 }, { "epoch": 16.62874251497006, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1582, "step": 11108 }, { "epoch": 16.630239520958085, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.157, "step": 11109 }, { "epoch": 16.631736526946106, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1535, "step": 11110 }, { "epoch": 16.63323353293413, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1575, "step": 11111 }, { "epoch": 16.634730538922156, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1559, "step": 11112 }, { "epoch": 16.63622754491018, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1557, "step": 11113 }, { "epoch": 16.637724550898202, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1539, "step": 11114 }, { "epoch": 16.639221556886227, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1605, "step": 11115 }, { "epoch": 16.64071856287425, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1561, "step": 11116 }, { "epoch": 16.642215568862277, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1545, "step": 11117 }, { "epoch": 16.643712574850298, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1518, "step": 11118 }, { "epoch": 16.645209580838323, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1506, "step": 11119 }, { "epoch": 16.646706586826348, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1522, "step": 11120 }, { "epoch": 16.648203592814372, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1531, "step": 11121 }, { "epoch": 16.649700598802394, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1532, "step": 11122 }, { "epoch": 16.65119760479042, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1592, "step": 11123 }, { "epoch": 16.652694610778443, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1598, "step": 11124 }, { "epoch": 16.654191616766468, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1529, "step": 11125 }, { "epoch": 16.65568862275449, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1574, "step": 11126 }, { "epoch": 16.657185628742514, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1462, "step": 11127 }, { "epoch": 16.65868263473054, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1524, "step": 11128 }, { "epoch": 16.660179640718564, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1567, "step": 11129 }, { "epoch": 16.66167664670659, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1529, "step": 11130 }, { "epoch": 16.66317365269461, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1506, "step": 11131 }, { "epoch": 16.664670658682635, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1445, "step": 11132 }, { "epoch": 16.66616766467066, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1503, "step": 11133 }, { "epoch": 16.66766467065868, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1453, "step": 11134 }, { "epoch": 16.669161676646706, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1542, "step": 11135 }, { "epoch": 16.67065868263473, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1519, "step": 11136 }, { "epoch": 16.672155688622755, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1602, "step": 11137 }, { "epoch": 16.67365269461078, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1489, "step": 11138 }, { "epoch": 16.6751497005988, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1504, "step": 11139 }, { "epoch": 16.676646706586826, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1483, "step": 11140 }, { "epoch": 16.67814371257485, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.153, "step": 11141 }, { "epoch": 16.679640718562876, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1476, "step": 11142 }, { "epoch": 16.681137724550897, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1546, "step": 11143 }, { "epoch": 16.682634730538922, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1521, "step": 11144 }, { "epoch": 16.684131736526947, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1565, "step": 11145 }, { "epoch": 16.68562874251497, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1555, "step": 11146 }, { "epoch": 16.687125748502993, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1517, "step": 11147 }, { "epoch": 16.688622754491018, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1492, "step": 11148 }, { "epoch": 16.690119760479043, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1494, "step": 11149 }, { "epoch": 16.691616766467067, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1535, "step": 11150 }, { "epoch": 16.69311377245509, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1471, "step": 11151 }, { "epoch": 16.694610778443113, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1564, "step": 11152 }, { "epoch": 16.69610778443114, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1487, "step": 11153 }, { "epoch": 16.697604790419163, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1528, "step": 11154 }, { "epoch": 16.699101796407184, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1518, "step": 11155 }, { "epoch": 16.70059880239521, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1453, "step": 11156 }, { "epoch": 16.702095808383234, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.149, "step": 11157 }, { "epoch": 16.70359281437126, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1488, "step": 11158 }, { "epoch": 16.70508982035928, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1471, "step": 11159 }, { "epoch": 16.706586826347305, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1465, "step": 11160 }, { "epoch": 16.70808383233533, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1531, "step": 11161 }, { "epoch": 16.709580838323355, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1481, "step": 11162 }, { "epoch": 16.711077844311376, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.147, "step": 11163 }, { "epoch": 16.7125748502994, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1505, "step": 11164 }, { "epoch": 16.714071856287426, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.145, "step": 11165 }, { "epoch": 16.71556886227545, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1524, "step": 11166 }, { "epoch": 16.71706586826347, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1476, "step": 11167 }, { "epoch": 16.718562874251496, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1575, "step": 11168 }, { "epoch": 16.72005988023952, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1608, "step": 11169 }, { "epoch": 16.721556886227546, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1537, "step": 11170 }, { "epoch": 16.723053892215567, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1419, "step": 11171 }, { "epoch": 16.724550898203592, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1466, "step": 11172 }, { "epoch": 16.726047904191617, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1478, "step": 11173 }, { "epoch": 16.727544910179642, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1502, "step": 11174 }, { "epoch": 16.729041916167663, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1529, "step": 11175 }, { "epoch": 16.730538922155688, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1506, "step": 11176 }, { "epoch": 16.732035928143713, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1526, "step": 11177 }, { "epoch": 16.733532934131738, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1471, "step": 11178 }, { "epoch": 16.73502994011976, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.153, "step": 11179 }, { "epoch": 16.736526946107784, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1528, "step": 11180 }, { "epoch": 16.73802395209581, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1478, "step": 11181 }, { "epoch": 16.739520958083833, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1463, "step": 11182 }, { "epoch": 16.741017964071855, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1509, "step": 11183 }, { "epoch": 16.74251497005988, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1508, "step": 11184 }, { "epoch": 16.744011976047904, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1459, "step": 11185 }, { "epoch": 16.74550898203593, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1425, "step": 11186 }, { "epoch": 16.74700598802395, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1492, "step": 11187 }, { "epoch": 16.748502994011975, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1478, "step": 11188 }, { "epoch": 16.75, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1528, "step": 11189 }, { "epoch": 16.751497005988025, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1529, "step": 11190 }, { "epoch": 16.75299401197605, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1525, "step": 11191 }, { "epoch": 16.75449101796407, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1474, "step": 11192 }, { "epoch": 16.755988023952096, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.147, "step": 11193 }, { "epoch": 16.75748502994012, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1531, "step": 11194 }, { "epoch": 16.758982035928145, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1559, "step": 11195 }, { "epoch": 16.760479041916167, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1571, "step": 11196 }, { "epoch": 16.76197604790419, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1498, "step": 11197 }, { "epoch": 16.763473053892216, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1517, "step": 11198 }, { "epoch": 16.76497005988024, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1444, "step": 11199 }, { "epoch": 16.766467065868262, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1451, "step": 11200 }, { "epoch": 16.767964071856287, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1509, "step": 11201 }, { "epoch": 16.769461077844312, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1484, "step": 11202 }, { "epoch": 16.770958083832337, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.152, "step": 11203 }, { "epoch": 16.772455089820358, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1491, "step": 11204 }, { "epoch": 16.773952095808383, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1484, "step": 11205 }, { "epoch": 16.775449101796408, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1542, "step": 11206 }, { "epoch": 16.776946107784433, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1435, "step": 11207 }, { "epoch": 16.778443113772454, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1429, "step": 11208 }, { "epoch": 16.77994011976048, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.145, "step": 11209 }, { "epoch": 16.781437125748504, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1484, "step": 11210 }, { "epoch": 16.78293413173653, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1487, "step": 11211 }, { "epoch": 16.78443113772455, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1456, "step": 11212 }, { "epoch": 16.785928143712574, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1449, "step": 11213 }, { "epoch": 16.7874251497006, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1424, "step": 11214 }, { "epoch": 16.788922155688624, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1531, "step": 11215 }, { "epoch": 16.790419161676645, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1436, "step": 11216 }, { "epoch": 16.79191616766467, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.149, "step": 11217 }, { "epoch": 16.793413173652695, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1478, "step": 11218 }, { "epoch": 16.79491017964072, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1505, "step": 11219 }, { "epoch": 16.79640718562874, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11220 }, { "epoch": 16.797904191616766, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1485, "step": 11221 }, { "epoch": 16.79940119760479, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1436, "step": 11222 }, { "epoch": 16.800898203592816, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1495, "step": 11223 }, { "epoch": 16.802395209580837, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1493, "step": 11224 }, { "epoch": 16.80389221556886, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1457, "step": 11225 }, { "epoch": 16.805389221556887, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1476, "step": 11226 }, { "epoch": 16.80688622754491, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1419, "step": 11227 }, { "epoch": 16.808383233532933, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1507, "step": 11228 }, { "epoch": 16.809880239520957, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1493, "step": 11229 }, { "epoch": 16.811377245508982, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1504, "step": 11230 }, { "epoch": 16.812874251497007, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1458, "step": 11231 }, { "epoch": 16.81437125748503, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1539, "step": 11232 }, { "epoch": 16.815868263473053, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1446, "step": 11233 }, { "epoch": 16.817365269461078, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1474, "step": 11234 }, { "epoch": 16.818862275449103, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1456, "step": 11235 }, { "epoch": 16.820359281437124, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1461, "step": 11236 }, { "epoch": 16.82185628742515, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1512, "step": 11237 }, { "epoch": 16.823353293413174, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1448, "step": 11238 }, { "epoch": 16.8248502994012, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1449, "step": 11239 }, { "epoch": 16.82634730538922, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1424, "step": 11240 }, { "epoch": 16.827844311377245, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1427, "step": 11241 }, { "epoch": 16.82934131736527, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1468, "step": 11242 }, { "epoch": 16.830838323353294, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1461, "step": 11243 }, { "epoch": 16.83233532934132, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1524, "step": 11244 }, { "epoch": 16.83383233532934, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1468, "step": 11245 }, { "epoch": 16.835329341317365, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1479, "step": 11246 }, { "epoch": 16.83682634730539, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1439, "step": 11247 }, { "epoch": 16.83832335329341, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1467, "step": 11248 }, { "epoch": 16.839820359281436, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1446, "step": 11249 }, { "epoch": 16.84131736526946, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1484, "step": 11250 }, { "epoch": 16.842814371257486, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1464, "step": 11251 }, { "epoch": 16.84431137724551, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1446, "step": 11252 }, { "epoch": 16.845808383233532, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1432, "step": 11253 }, { "epoch": 16.847305389221557, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.149, "step": 11254 }, { "epoch": 16.84880239520958, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1431, "step": 11255 }, { "epoch": 16.850299401197606, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1398, "step": 11256 }, { "epoch": 16.851796407185628, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1449, "step": 11257 }, { "epoch": 16.853293413173652, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1494, "step": 11258 }, { "epoch": 16.854790419161677, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1467, "step": 11259 }, { "epoch": 16.856287425149702, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1436, "step": 11260 }, { "epoch": 16.857784431137723, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1465, "step": 11261 }, { "epoch": 16.85928143712575, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1444, "step": 11262 }, { "epoch": 16.860778443113773, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1486, "step": 11263 }, { "epoch": 16.862275449101798, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.1518, "step": 11264 }, { "epoch": 16.86377245508982, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1393, "step": 11265 }, { "epoch": 16.865269461077844, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.148, "step": 11266 }, { "epoch": 16.86676646706587, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1459, "step": 11267 }, { "epoch": 16.868263473053894, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1498, "step": 11268 }, { "epoch": 16.869760479041915, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 11269 }, { "epoch": 16.87125748502994, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1456, "step": 11270 }, { "epoch": 16.872754491017965, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1509, "step": 11271 }, { "epoch": 16.87425149700599, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.1422, "step": 11272 }, { "epoch": 16.87574850299401, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1411, "step": 11273 }, { "epoch": 16.877245508982035, "grad_norm": 0.1953125, "learning_rate": 0.0008, "loss": 1.143, "step": 11274 }, { "epoch": 16.87874251497006, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1429, "step": 11275 }, { "epoch": 16.880239520958085, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1466, "step": 11276 }, { "epoch": 16.881736526946106, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1496, "step": 11277 }, { "epoch": 16.88323353293413, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1468, "step": 11278 }, { "epoch": 16.884730538922156, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1445, "step": 11279 }, { "epoch": 16.88622754491018, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1487, "step": 11280 }, { "epoch": 16.887724550898202, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1458, "step": 11281 }, { "epoch": 16.889221556886227, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1482, "step": 11282 }, { "epoch": 16.89071856287425, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1458, "step": 11283 }, { "epoch": 16.892215568862277, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1422, "step": 11284 }, { "epoch": 16.893712574850298, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1462, "step": 11285 }, { "epoch": 16.895209580838323, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1491, "step": 11286 }, { "epoch": 16.896706586826348, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1524, "step": 11287 }, { "epoch": 16.898203592814372, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1512, "step": 11288 }, { "epoch": 16.899700598802394, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1432, "step": 11289 }, { "epoch": 16.90119760479042, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1506, "step": 11290 }, { "epoch": 16.902694610778443, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1422, "step": 11291 }, { "epoch": 16.904191616766468, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1445, "step": 11292 }, { "epoch": 16.90568862275449, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1445, "step": 11293 }, { "epoch": 16.907185628742514, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1508, "step": 11294 }, { "epoch": 16.90868263473054, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1524, "step": 11295 }, { "epoch": 16.910179640718564, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1459, "step": 11296 }, { "epoch": 16.91167664670659, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1465, "step": 11297 }, { "epoch": 16.91317365269461, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.143, "step": 11298 }, { "epoch": 16.914670658682635, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11299 }, { "epoch": 16.91616766467066, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1471, "step": 11300 }, { "epoch": 16.91766467065868, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1412, "step": 11301 }, { "epoch": 16.919161676646706, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1435, "step": 11302 }, { "epoch": 16.92065868263473, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1458, "step": 11303 }, { "epoch": 16.922155688622755, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1416, "step": 11304 }, { "epoch": 16.92365269461078, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1398, "step": 11305 }, { "epoch": 16.9251497005988, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1488, "step": 11306 }, { "epoch": 16.926646706586826, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1461, "step": 11307 }, { "epoch": 16.92814371257485, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1397, "step": 11308 }, { "epoch": 16.929640718562876, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1462, "step": 11309 }, { "epoch": 16.931137724550897, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1495, "step": 11310 }, { "epoch": 16.932634730538922, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1491, "step": 11311 }, { "epoch": 16.934131736526947, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1457, "step": 11312 }, { "epoch": 16.93562874251497, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1431, "step": 11313 }, { "epoch": 16.937125748502993, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1447, "step": 11314 }, { "epoch": 16.938622754491018, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1422, "step": 11315 }, { "epoch": 16.940119760479043, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1453, "step": 11316 }, { "epoch": 16.941616766467067, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1533, "step": 11317 }, { "epoch": 16.94311377245509, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1452, "step": 11318 }, { "epoch": 16.944610778443113, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1473, "step": 11319 }, { "epoch": 16.94610778443114, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1475, "step": 11320 }, { "epoch": 16.947604790419163, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.147, "step": 11321 }, { "epoch": 16.949101796407184, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1459, "step": 11322 }, { "epoch": 16.95059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1453, "step": 11323 }, { "epoch": 16.952095808383234, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1378, "step": 11324 }, { "epoch": 16.95359281437126, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1416, "step": 11325 }, { "epoch": 16.95508982035928, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1418, "step": 11326 }, { "epoch": 16.956586826347305, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1424, "step": 11327 }, { "epoch": 16.95808383233533, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1419, "step": 11328 }, { "epoch": 16.959580838323355, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1409, "step": 11329 }, { "epoch": 16.961077844311376, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1411, "step": 11330 }, { "epoch": 16.9625748502994, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.149, "step": 11331 }, { "epoch": 16.964071856287426, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1442, "step": 11332 }, { "epoch": 16.96556886227545, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1432, "step": 11333 }, { "epoch": 16.96706586826347, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1507, "step": 11334 }, { "epoch": 16.968562874251496, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1437, "step": 11335 }, { "epoch": 16.97005988023952, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1459, "step": 11336 }, { "epoch": 16.971556886227546, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1466, "step": 11337 }, { "epoch": 16.973053892215567, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1386, "step": 11338 }, { "epoch": 16.974550898203592, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1506, "step": 11339 }, { "epoch": 16.976047904191617, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1466, "step": 11340 }, { "epoch": 16.977544910179642, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1467, "step": 11341 }, { "epoch": 16.979041916167663, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1459, "step": 11342 }, { "epoch": 16.980538922155688, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.144, "step": 11343 }, { "epoch": 16.982035928143713, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1409, "step": 11344 }, { "epoch": 16.983532934131738, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1483, "step": 11345 }, { "epoch": 16.98502994011976, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1467, "step": 11346 }, { "epoch": 16.986526946107784, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1493, "step": 11347 }, { "epoch": 16.98802395209581, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1338, "step": 11348 }, { "epoch": 16.989520958083833, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1453, "step": 11349 }, { "epoch": 16.991017964071855, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1452, "step": 11350 }, { "epoch": 16.99251497005988, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 11351 }, { "epoch": 16.994011976047904, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1516, "step": 11352 }, { "epoch": 16.99550898203593, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1423, "step": 11353 }, { "epoch": 16.99700598802395, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1444, "step": 11354 }, { "epoch": 16.998502994011975, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1388, "step": 11355 }, { "epoch": 17.0, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1399, "step": 11356 }, { "epoch": 17.001497005988025, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1538, "step": 11357 }, { "epoch": 17.00299401197605, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1396, "step": 11358 }, { "epoch": 17.00449101796407, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1425, "step": 11359 }, { "epoch": 17.005988023952096, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1519, "step": 11360 }, { "epoch": 17.00748502994012, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1427, "step": 11361 }, { "epoch": 17.008982035928145, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1473, "step": 11362 }, { "epoch": 17.010479041916167, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1442, "step": 11363 }, { "epoch": 17.01197604790419, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1489, "step": 11364 }, { "epoch": 17.013473053892216, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1472, "step": 11365 }, { "epoch": 17.01497005988024, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1414, "step": 11366 }, { "epoch": 17.016467065868262, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1466, "step": 11367 }, { "epoch": 17.017964071856287, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1392, "step": 11368 }, { "epoch": 17.019461077844312, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1482, "step": 11369 }, { "epoch": 17.020958083832337, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1505, "step": 11370 }, { "epoch": 17.022455089820358, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 11371 }, { "epoch": 17.023952095808383, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1438, "step": 11372 }, { "epoch": 17.025449101796408, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1431, "step": 11373 }, { "epoch": 17.026946107784433, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1469, "step": 11374 }, { "epoch": 17.028443113772454, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1388, "step": 11375 }, { "epoch": 17.02994011976048, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1483, "step": 11376 }, { "epoch": 17.031437125748504, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1466, "step": 11377 }, { "epoch": 17.03293413173653, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1397, "step": 11378 }, { "epoch": 17.03443113772455, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1482, "step": 11379 }, { "epoch": 17.035928143712574, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.141, "step": 11380 }, { "epoch": 17.0374251497006, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.146, "step": 11381 }, { "epoch": 17.038922155688624, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1474, "step": 11382 }, { "epoch": 17.040419161676645, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1398, "step": 11383 }, { "epoch": 17.04191616766467, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1442, "step": 11384 }, { "epoch": 17.043413173652695, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1484, "step": 11385 }, { "epoch": 17.04491017964072, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1421, "step": 11386 }, { "epoch": 17.04640718562874, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1474, "step": 11387 }, { "epoch": 17.047904191616766, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1446, "step": 11388 }, { "epoch": 17.04940119760479, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1424, "step": 11389 }, { "epoch": 17.050898203592816, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1436, "step": 11390 }, { "epoch": 17.052395209580837, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1481, "step": 11391 }, { "epoch": 17.05389221556886, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1428, "step": 11392 }, { "epoch": 17.055389221556887, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1457, "step": 11393 }, { "epoch": 17.05688622754491, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1391, "step": 11394 }, { "epoch": 17.058383233532933, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1416, "step": 11395 }, { "epoch": 17.059880239520957, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1469, "step": 11396 }, { "epoch": 17.061377245508982, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1495, "step": 11397 }, { "epoch": 17.062874251497007, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1405, "step": 11398 }, { "epoch": 17.06437125748503, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1394, "step": 11399 }, { "epoch": 17.065868263473053, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.144, "step": 11400 }, { "epoch": 17.067365269461078, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1394, "step": 11401 }, { "epoch": 17.068862275449103, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1446, "step": 11402 }, { "epoch": 17.070359281437124, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1422, "step": 11403 }, { "epoch": 17.07185628742515, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.138, "step": 11404 }, { "epoch": 17.073353293413174, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.143, "step": 11405 }, { "epoch": 17.0748502994012, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1447, "step": 11406 }, { "epoch": 17.07634730538922, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1341, "step": 11407 }, { "epoch": 17.077844311377245, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.136, "step": 11408 }, { "epoch": 17.07934131736527, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1488, "step": 11409 }, { "epoch": 17.080838323353294, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1499, "step": 11410 }, { "epoch": 17.082335329341316, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1437, "step": 11411 }, { "epoch": 17.08383233532934, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1475, "step": 11412 }, { "epoch": 17.085329341317365, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1447, "step": 11413 }, { "epoch": 17.08682634730539, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1433, "step": 11414 }, { "epoch": 17.088323353293415, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.147, "step": 11415 }, { "epoch": 17.089820359281436, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1433, "step": 11416 }, { "epoch": 17.09131736526946, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1469, "step": 11417 }, { "epoch": 17.092814371257486, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1478, "step": 11418 }, { "epoch": 17.09431137724551, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1415, "step": 11419 }, { "epoch": 17.095808383233532, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1461, "step": 11420 }, { "epoch": 17.097305389221557, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1418, "step": 11421 }, { "epoch": 17.09880239520958, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1336, "step": 11422 }, { "epoch": 17.100299401197606, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1369, "step": 11423 }, { "epoch": 17.101796407185628, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1433, "step": 11424 }, { "epoch": 17.103293413173652, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.15, "step": 11425 }, { "epoch": 17.104790419161677, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1447, "step": 11426 }, { "epoch": 17.106287425149702, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1381, "step": 11427 }, { "epoch": 17.107784431137723, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.145, "step": 11428 }, { "epoch": 17.10928143712575, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1414, "step": 11429 }, { "epoch": 17.110778443113773, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1458, "step": 11430 }, { "epoch": 17.112275449101798, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1451, "step": 11431 }, { "epoch": 17.11377245508982, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1433, "step": 11432 }, { "epoch": 17.115269461077844, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1415, "step": 11433 }, { "epoch": 17.11676646706587, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.15, "step": 11434 }, { "epoch": 17.118263473053894, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1393, "step": 11435 }, { "epoch": 17.119760479041915, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1401, "step": 11436 }, { "epoch": 17.12125748502994, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1448, "step": 11437 }, { "epoch": 17.122754491017965, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1447, "step": 11438 }, { "epoch": 17.12425149700599, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1395, "step": 11439 }, { "epoch": 17.12574850299401, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1415, "step": 11440 }, { "epoch": 17.127245508982035, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1456, "step": 11441 }, { "epoch": 17.12874251497006, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1412, "step": 11442 }, { "epoch": 17.130239520958085, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.15, "step": 11443 }, { "epoch": 17.131736526946106, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1458, "step": 11444 }, { "epoch": 17.13323353293413, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1403, "step": 11445 }, { "epoch": 17.134730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1398, "step": 11446 }, { "epoch": 17.13622754491018, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 11447 }, { "epoch": 17.137724550898202, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1432, "step": 11448 }, { "epoch": 17.139221556886227, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1443, "step": 11449 }, { "epoch": 17.14071856287425, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1463, "step": 11450 }, { "epoch": 17.142215568862277, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1405, "step": 11451 }, { "epoch": 17.143712574850298, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1407, "step": 11452 }, { "epoch": 17.145209580838323, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1424, "step": 11453 }, { "epoch": 17.146706586826348, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1414, "step": 11454 }, { "epoch": 17.148203592814372, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.143, "step": 11455 }, { "epoch": 17.149700598802394, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1389, "step": 11456 }, { "epoch": 17.15119760479042, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1384, "step": 11457 }, { "epoch": 17.152694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1361, "step": 11458 }, { "epoch": 17.154191616766468, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1387, "step": 11459 }, { "epoch": 17.15568862275449, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1453, "step": 11460 }, { "epoch": 17.157185628742514, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1358, "step": 11461 }, { "epoch": 17.15868263473054, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1484, "step": 11462 }, { "epoch": 17.160179640718564, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1515, "step": 11463 }, { "epoch": 17.161676646706585, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1465, "step": 11464 }, { "epoch": 17.16317365269461, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1406, "step": 11465 }, { "epoch": 17.164670658682635, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1382, "step": 11466 }, { "epoch": 17.16616766467066, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1379, "step": 11467 }, { "epoch": 17.16766467065868, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1491, "step": 11468 }, { "epoch": 17.169161676646706, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 11469 }, { "epoch": 17.17065868263473, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 11470 }, { "epoch": 17.172155688622755, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1452, "step": 11471 }, { "epoch": 17.17365269461078, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1416, "step": 11472 }, { "epoch": 17.1751497005988, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.14, "step": 11473 }, { "epoch": 17.176646706586826, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1447, "step": 11474 }, { "epoch": 17.17814371257485, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1348, "step": 11475 }, { "epoch": 17.179640718562876, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 11476 }, { "epoch": 17.181137724550897, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1356, "step": 11477 }, { "epoch": 17.182634730538922, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1406, "step": 11478 }, { "epoch": 17.184131736526947, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1405, "step": 11479 }, { "epoch": 17.18562874251497, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 11480 }, { "epoch": 17.187125748502993, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1387, "step": 11481 }, { "epoch": 17.188622754491018, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1467, "step": 11482 }, { "epoch": 17.190119760479043, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1462, "step": 11483 }, { "epoch": 17.191616766467067, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1467, "step": 11484 }, { "epoch": 17.19311377245509, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1421, "step": 11485 }, { "epoch": 17.194610778443113, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1403, "step": 11486 }, { "epoch": 17.19610778443114, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.142, "step": 11487 }, { "epoch": 17.197604790419163, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1372, "step": 11488 }, { "epoch": 17.199101796407184, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1405, "step": 11489 }, { "epoch": 17.20059880239521, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1456, "step": 11490 }, { "epoch": 17.202095808383234, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.139, "step": 11491 }, { "epoch": 17.20359281437126, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1372, "step": 11492 }, { "epoch": 17.20508982035928, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1336, "step": 11493 }, { "epoch": 17.206586826347305, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1468, "step": 11494 }, { "epoch": 17.20808383233533, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1402, "step": 11495 }, { "epoch": 17.209580838323355, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1452, "step": 11496 }, { "epoch": 17.211077844311376, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1386, "step": 11497 }, { "epoch": 17.2125748502994, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1437, "step": 11498 }, { "epoch": 17.214071856287426, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.143, "step": 11499 }, { "epoch": 17.21556886227545, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1481, "step": 11500 }, { "epoch": 17.21706586826347, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1451, "step": 11501 }, { "epoch": 17.218562874251496, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1463, "step": 11502 }, { "epoch": 17.22005988023952, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1445, "step": 11503 }, { "epoch": 17.221556886227546, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.141, "step": 11504 }, { "epoch": 17.223053892215567, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1438, "step": 11505 }, { "epoch": 17.224550898203592, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1374, "step": 11506 }, { "epoch": 17.226047904191617, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1394, "step": 11507 }, { "epoch": 17.227544910179642, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1414, "step": 11508 }, { "epoch": 17.229041916167663, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1476, "step": 11509 }, { "epoch": 17.230538922155688, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1388, "step": 11510 }, { "epoch": 17.232035928143713, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1468, "step": 11511 }, { "epoch": 17.233532934131738, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1422, "step": 11512 }, { "epoch": 17.23502994011976, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1322, "step": 11513 }, { "epoch": 17.236526946107784, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1394, "step": 11514 }, { "epoch": 17.23802395209581, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1445, "step": 11515 }, { "epoch": 17.239520958083833, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1449, "step": 11516 }, { "epoch": 17.241017964071855, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.134, "step": 11517 }, { "epoch": 17.24251497005988, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1474, "step": 11518 }, { "epoch": 17.244011976047904, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.139, "step": 11519 }, { "epoch": 17.24550898203593, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1461, "step": 11520 }, { "epoch": 17.24700598802395, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1444, "step": 11521 }, { "epoch": 17.248502994011975, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1369, "step": 11522 }, { "epoch": 17.25, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1421, "step": 11523 }, { "epoch": 17.251497005988025, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1441, "step": 11524 }, { "epoch": 17.25299401197605, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1386, "step": 11525 }, { "epoch": 17.25449101796407, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1386, "step": 11526 }, { "epoch": 17.255988023952096, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1424, "step": 11527 }, { "epoch": 17.25748502994012, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1417, "step": 11528 }, { "epoch": 17.258982035928145, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1456, "step": 11529 }, { "epoch": 17.260479041916167, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 11530 }, { "epoch": 17.26197604790419, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1435, "step": 11531 }, { "epoch": 17.263473053892216, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1443, "step": 11532 }, { "epoch": 17.26497005988024, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1437, "step": 11533 }, { "epoch": 17.266467065868262, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1445, "step": 11534 }, { "epoch": 17.267964071856287, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1428, "step": 11535 }, { "epoch": 17.269461077844312, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1448, "step": 11536 }, { "epoch": 17.270958083832337, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1407, "step": 11537 }, { "epoch": 17.272455089820358, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1463, "step": 11538 }, { "epoch": 17.273952095808383, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1421, "step": 11539 }, { "epoch": 17.275449101796408, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1449, "step": 11540 }, { "epoch": 17.276946107784433, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1418, "step": 11541 }, { "epoch": 17.278443113772454, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1402, "step": 11542 }, { "epoch": 17.27994011976048, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1364, "step": 11543 }, { "epoch": 17.281437125748504, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1423, "step": 11544 }, { "epoch": 17.28293413173653, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1444, "step": 11545 }, { "epoch": 17.28443113772455, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.137, "step": 11546 }, { "epoch": 17.285928143712574, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1472, "step": 11547 }, { "epoch": 17.2874251497006, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1432, "step": 11548 }, { "epoch": 17.288922155688624, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1405, "step": 11549 }, { "epoch": 17.290419161676645, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1319, "step": 11550 }, { "epoch": 17.29191616766467, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1334, "step": 11551 }, { "epoch": 17.293413173652695, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1431, "step": 11552 }, { "epoch": 17.29491017964072, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1405, "step": 11553 }, { "epoch": 17.29640718562874, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1412, "step": 11554 }, { "epoch": 17.297904191616766, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1441, "step": 11555 }, { "epoch": 17.29940119760479, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1407, "step": 11556 }, { "epoch": 17.300898203592816, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1394, "step": 11557 }, { "epoch": 17.302395209580837, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1404, "step": 11558 }, { "epoch": 17.30389221556886, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1374, "step": 11559 }, { "epoch": 17.305389221556887, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1357, "step": 11560 }, { "epoch": 17.30688622754491, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1421, "step": 11561 }, { "epoch": 17.308383233532933, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1426, "step": 11562 }, { "epoch": 17.309880239520957, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1374, "step": 11563 }, { "epoch": 17.311377245508982, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1467, "step": 11564 }, { "epoch": 17.312874251497007, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1426, "step": 11565 }, { "epoch": 17.31437125748503, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1393, "step": 11566 }, { "epoch": 17.315868263473053, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1401, "step": 11567 }, { "epoch": 17.317365269461078, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.14, "step": 11568 }, { "epoch": 17.318862275449103, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1464, "step": 11569 }, { "epoch": 17.320359281437124, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 11570 }, { "epoch": 17.32185628742515, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1439, "step": 11571 }, { "epoch": 17.323353293413174, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11572 }, { "epoch": 17.3248502994012, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1494, "step": 11573 }, { "epoch": 17.32634730538922, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1481, "step": 11574 }, { "epoch": 17.327844311377245, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.141, "step": 11575 }, { "epoch": 17.32934131736527, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1417, "step": 11576 }, { "epoch": 17.330838323353294, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1459, "step": 11577 }, { "epoch": 17.33233532934132, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1462, "step": 11578 }, { "epoch": 17.33383233532934, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1467, "step": 11579 }, { "epoch": 17.335329341317365, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.145, "step": 11580 }, { "epoch": 17.33682634730539, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1418, "step": 11581 }, { "epoch": 17.338323353293415, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1467, "step": 11582 }, { "epoch": 17.339820359281436, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 11583 }, { "epoch": 17.34131736526946, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1397, "step": 11584 }, { "epoch": 17.342814371257486, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1418, "step": 11585 }, { "epoch": 17.34431137724551, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1328, "step": 11586 }, { "epoch": 17.345808383233532, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1394, "step": 11587 }, { "epoch": 17.347305389221557, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1384, "step": 11588 }, { "epoch": 17.34880239520958, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1465, "step": 11589 }, { "epoch": 17.350299401197606, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1432, "step": 11590 }, { "epoch": 17.351796407185628, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1445, "step": 11591 }, { "epoch": 17.353293413173652, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1463, "step": 11592 }, { "epoch": 17.354790419161677, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1454, "step": 11593 }, { "epoch": 17.356287425149702, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1442, "step": 11594 }, { "epoch": 17.357784431137723, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1446, "step": 11595 }, { "epoch": 17.35928143712575, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1466, "step": 11596 }, { "epoch": 17.360778443113773, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 11597 }, { "epoch": 17.362275449101798, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1437, "step": 11598 }, { "epoch": 17.36377245508982, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1394, "step": 11599 }, { "epoch": 17.365269461077844, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 11600 }, { "epoch": 17.36676646706587, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1327, "step": 11601 }, { "epoch": 17.368263473053894, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1371, "step": 11602 }, { "epoch": 17.369760479041915, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1426, "step": 11603 }, { "epoch": 17.37125748502994, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1409, "step": 11604 }, { "epoch": 17.372754491017965, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1456, "step": 11605 }, { "epoch": 17.37425149700599, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1492, "step": 11606 }, { "epoch": 17.37574850299401, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1424, "step": 11607 }, { "epoch": 17.377245508982035, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1457, "step": 11608 }, { "epoch": 17.37874251497006, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1371, "step": 11609 }, { "epoch": 17.380239520958085, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1463, "step": 11610 }, { "epoch": 17.381736526946106, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1455, "step": 11611 }, { "epoch": 17.38323353293413, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1362, "step": 11612 }, { "epoch": 17.384730538922156, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1436, "step": 11613 }, { "epoch": 17.38622754491018, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1441, "step": 11614 }, { "epoch": 17.387724550898202, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1425, "step": 11615 }, { "epoch": 17.389221556886227, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1441, "step": 11616 }, { "epoch": 17.39071856287425, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1436, "step": 11617 }, { "epoch": 17.392215568862277, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1418, "step": 11618 }, { "epoch": 17.393712574850298, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1458, "step": 11619 }, { "epoch": 17.395209580838323, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1471, "step": 11620 }, { "epoch": 17.396706586826348, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1375, "step": 11621 }, { "epoch": 17.398203592814372, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1431, "step": 11622 }, { "epoch": 17.399700598802394, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1414, "step": 11623 }, { "epoch": 17.40119760479042, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1419, "step": 11624 }, { "epoch": 17.402694610778443, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1458, "step": 11625 }, { "epoch": 17.404191616766468, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1448, "step": 11626 }, { "epoch": 17.40568862275449, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1432, "step": 11627 }, { "epoch": 17.407185628742514, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1416, "step": 11628 }, { "epoch": 17.40868263473054, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1344, "step": 11629 }, { "epoch": 17.410179640718564, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1398, "step": 11630 }, { "epoch": 17.411676646706585, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1404, "step": 11631 }, { "epoch": 17.41317365269461, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1413, "step": 11632 }, { "epoch": 17.414670658682635, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1462, "step": 11633 }, { "epoch": 17.41616766467066, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1414, "step": 11634 }, { "epoch": 17.41766467065868, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1318, "step": 11635 }, { "epoch": 17.419161676646706, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1388, "step": 11636 }, { "epoch": 17.42065868263473, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1468, "step": 11637 }, { "epoch": 17.422155688622755, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11638 }, { "epoch": 17.42365269461078, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1397, "step": 11639 }, { "epoch": 17.4251497005988, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11640 }, { "epoch": 17.426646706586826, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1398, "step": 11641 }, { "epoch": 17.42814371257485, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1437, "step": 11642 }, { "epoch": 17.429640718562876, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 11643 }, { "epoch": 17.431137724550897, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1402, "step": 11644 }, { "epoch": 17.432634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1479, "step": 11645 }, { "epoch": 17.434131736526947, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1357, "step": 11646 }, { "epoch": 17.43562874251497, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.139, "step": 11647 }, { "epoch": 17.437125748502993, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1476, "step": 11648 }, { "epoch": 17.438622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 11649 }, { "epoch": 17.440119760479043, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1485, "step": 11650 }, { "epoch": 17.441616766467067, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1331, "step": 11651 }, { "epoch": 17.44311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1435, "step": 11652 }, { "epoch": 17.444610778443113, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1416, "step": 11653 }, { "epoch": 17.44610778443114, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1427, "step": 11654 }, { "epoch": 17.447604790419163, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1459, "step": 11655 }, { "epoch": 17.449101796407184, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1379, "step": 11656 }, { "epoch": 17.45059880239521, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1441, "step": 11657 }, { "epoch": 17.452095808383234, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1427, "step": 11658 }, { "epoch": 17.45359281437126, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 11659 }, { "epoch": 17.45508982035928, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1371, "step": 11660 }, { "epoch": 17.456586826347305, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1462, "step": 11661 }, { "epoch": 17.45808383233533, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1395, "step": 11662 }, { "epoch": 17.459580838323355, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.137, "step": 11663 }, { "epoch": 17.461077844311376, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.144, "step": 11664 }, { "epoch": 17.4625748502994, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1407, "step": 11665 }, { "epoch": 17.464071856287426, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1449, "step": 11666 }, { "epoch": 17.46556886227545, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.133, "step": 11667 }, { "epoch": 17.46706586826347, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 11668 }, { "epoch": 17.468562874251496, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1432, "step": 11669 }, { "epoch": 17.47005988023952, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1456, "step": 11670 }, { "epoch": 17.471556886227546, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1379, "step": 11671 }, { "epoch": 17.473053892215567, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1359, "step": 11672 }, { "epoch": 17.474550898203592, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 11673 }, { "epoch": 17.476047904191617, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1443, "step": 11674 }, { "epoch": 17.477544910179642, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1387, "step": 11675 }, { "epoch": 17.479041916167663, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1435, "step": 11676 }, { "epoch": 17.480538922155688, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1427, "step": 11677 }, { "epoch": 17.482035928143713, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1383, "step": 11678 }, { "epoch": 17.483532934131738, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1392, "step": 11679 }, { "epoch": 17.48502994011976, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1412, "step": 11680 }, { "epoch": 17.486526946107784, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1433, "step": 11681 }, { "epoch": 17.48802395209581, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1455, "step": 11682 }, { "epoch": 17.489520958083833, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1405, "step": 11683 }, { "epoch": 17.491017964071855, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1393, "step": 11684 }, { "epoch": 17.49251497005988, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1357, "step": 11685 }, { "epoch": 17.494011976047904, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.144, "step": 11686 }, { "epoch": 17.49550898203593, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1452, "step": 11687 }, { "epoch": 17.49700598802395, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1385, "step": 11688 }, { "epoch": 17.498502994011975, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 11689 }, { "epoch": 17.5, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1373, "step": 11690 }, { "epoch": 17.501497005988025, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.138, "step": 11691 }, { "epoch": 17.50299401197605, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1384, "step": 11692 }, { "epoch": 17.50449101796407, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.141, "step": 11693 }, { "epoch": 17.505988023952096, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1401, "step": 11694 }, { "epoch": 17.50748502994012, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1441, "step": 11695 }, { "epoch": 17.508982035928145, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1442, "step": 11696 }, { "epoch": 17.510479041916167, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1423, "step": 11697 }, { "epoch": 17.51197604790419, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1378, "step": 11698 }, { "epoch": 17.513473053892216, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1423, "step": 11699 }, { "epoch": 17.51497005988024, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.139, "step": 11700 }, { "epoch": 17.516467065868262, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1389, "step": 11701 }, { "epoch": 17.517964071856287, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1341, "step": 11702 }, { "epoch": 17.519461077844312, "grad_norm": 0.1748046875, "learning_rate": 0.0008, "loss": 1.1485, "step": 11703 }, { "epoch": 17.520958083832337, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1405, "step": 11704 }, { "epoch": 17.522455089820358, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1368, "step": 11705 }, { "epoch": 17.523952095808383, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1447, "step": 11706 }, { "epoch": 17.525449101796408, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1414, "step": 11707 }, { "epoch": 17.526946107784433, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1456, "step": 11708 }, { "epoch": 17.528443113772454, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1384, "step": 11709 }, { "epoch": 17.52994011976048, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.14, "step": 11710 }, { "epoch": 17.531437125748504, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1357, "step": 11711 }, { "epoch": 17.53293413173653, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1425, "step": 11712 }, { "epoch": 17.53443113772455, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1429, "step": 11713 }, { "epoch": 17.535928143712574, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1435, "step": 11714 }, { "epoch": 17.5374251497006, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1443, "step": 11715 }, { "epoch": 17.538922155688624, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1411, "step": 11716 }, { "epoch": 17.540419161676645, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1442, "step": 11717 }, { "epoch": 17.54191616766467, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1409, "step": 11718 }, { "epoch": 17.543413173652695, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1448, "step": 11719 }, { "epoch": 17.54491017964072, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.143, "step": 11720 }, { "epoch": 17.54640718562874, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1525, "step": 11721 }, { "epoch": 17.547904191616766, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1354, "step": 11722 }, { "epoch": 17.54940119760479, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1451, "step": 11723 }, { "epoch": 17.550898203592816, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1436, "step": 11724 }, { "epoch": 17.552395209580837, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1402, "step": 11725 }, { "epoch": 17.55389221556886, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1417, "step": 11726 }, { "epoch": 17.555389221556887, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1335, "step": 11727 }, { "epoch": 17.55688622754491, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.134, "step": 11728 }, { "epoch": 17.558383233532933, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1426, "step": 11729 }, { "epoch": 17.559880239520957, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1323, "step": 11730 }, { "epoch": 17.561377245508982, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1483, "step": 11731 }, { "epoch": 17.562874251497007, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1402, "step": 11732 }, { "epoch": 17.56437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 11733 }, { "epoch": 17.565868263473053, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1403, "step": 11734 }, { "epoch": 17.567365269461078, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1455, "step": 11735 }, { "epoch": 17.568862275449103, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 11736 }, { "epoch": 17.570359281437124, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1472, "step": 11737 }, { "epoch": 17.57185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1457, "step": 11738 }, { "epoch": 17.573353293413174, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.141, "step": 11739 }, { "epoch": 17.5748502994012, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1395, "step": 11740 }, { "epoch": 17.57634730538922, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1363, "step": 11741 }, { "epoch": 17.577844311377245, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1415, "step": 11742 }, { "epoch": 17.57934131736527, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1367, "step": 11743 }, { "epoch": 17.580838323353294, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11744 }, { "epoch": 17.58233532934132, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1393, "step": 11745 }, { "epoch": 17.58383233532934, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1433, "step": 11746 }, { "epoch": 17.585329341317365, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1425, "step": 11747 }, { "epoch": 17.58682634730539, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1417, "step": 11748 }, { "epoch": 17.58832335329341, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.144, "step": 11749 }, { "epoch": 17.589820359281436, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 11750 }, { "epoch": 17.59131736526946, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.142, "step": 11751 }, { "epoch": 17.592814371257486, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1437, "step": 11752 }, { "epoch": 17.59431137724551, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1374, "step": 11753 }, { "epoch": 17.595808383233532, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1412, "step": 11754 }, { "epoch": 17.597305389221557, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1377, "step": 11755 }, { "epoch": 17.59880239520958, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.145, "step": 11756 }, { "epoch": 17.600299401197606, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1444, "step": 11757 }, { "epoch": 17.601796407185628, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1424, "step": 11758 }, { "epoch": 17.603293413173652, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1433, "step": 11759 }, { "epoch": 17.604790419161677, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1449, "step": 11760 }, { "epoch": 17.606287425149702, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1445, "step": 11761 }, { "epoch": 17.607784431137723, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1439, "step": 11762 }, { "epoch": 17.60928143712575, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1406, "step": 11763 }, { "epoch": 17.610778443113773, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1419, "step": 11764 }, { "epoch": 17.612275449101798, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1445, "step": 11765 }, { "epoch": 17.61377245508982, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1397, "step": 11766 }, { "epoch": 17.615269461077844, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1462, "step": 11767 }, { "epoch": 17.61676646706587, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1399, "step": 11768 }, { "epoch": 17.618263473053894, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1403, "step": 11769 }, { "epoch": 17.619760479041915, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.142, "step": 11770 }, { "epoch": 17.62125748502994, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1379, "step": 11771 }, { "epoch": 17.622754491017965, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1453, "step": 11772 }, { "epoch": 17.62425149700599, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1414, "step": 11773 }, { "epoch": 17.62574850299401, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1484, "step": 11774 }, { "epoch": 17.627245508982035, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1365, "step": 11775 }, { "epoch": 17.62874251497006, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1423, "step": 11776 }, { "epoch": 17.630239520958085, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1426, "step": 11777 }, { "epoch": 17.631736526946106, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1474, "step": 11778 }, { "epoch": 17.63323353293413, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 11779 }, { "epoch": 17.634730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1436, "step": 11780 }, { "epoch": 17.63622754491018, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1447, "step": 11781 }, { "epoch": 17.637724550898202, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.142, "step": 11782 }, { "epoch": 17.639221556886227, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1345, "step": 11783 }, { "epoch": 17.64071856287425, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 11784 }, { "epoch": 17.642215568862277, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 11785 }, { "epoch": 17.643712574850298, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1361, "step": 11786 }, { "epoch": 17.645209580838323, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1397, "step": 11787 }, { "epoch": 17.646706586826348, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1386, "step": 11788 }, { "epoch": 17.648203592814372, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.138, "step": 11789 }, { "epoch": 17.649700598802394, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1397, "step": 11790 }, { "epoch": 17.65119760479042, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1479, "step": 11791 }, { "epoch": 17.652694610778443, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1419, "step": 11792 }, { "epoch": 17.654191616766468, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.136, "step": 11793 }, { "epoch": 17.65568862275449, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.133, "step": 11794 }, { "epoch": 17.657185628742514, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1366, "step": 11795 }, { "epoch": 17.65868263473054, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1419, "step": 11796 }, { "epoch": 17.660179640718564, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1401, "step": 11797 }, { "epoch": 17.66167664670659, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1343, "step": 11798 }, { "epoch": 17.66317365269461, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.141, "step": 11799 }, { "epoch": 17.664670658682635, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1377, "step": 11800 }, { "epoch": 17.66616766467066, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 11801 }, { "epoch": 17.66766467065868, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1435, "step": 11802 }, { "epoch": 17.669161676646706, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.143, "step": 11803 }, { "epoch": 17.67065868263473, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1429, "step": 11804 }, { "epoch": 17.672155688622755, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1374, "step": 11805 }, { "epoch": 17.67365269461078, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.143, "step": 11806 }, { "epoch": 17.6751497005988, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1367, "step": 11807 }, { "epoch": 17.676646706586826, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1452, "step": 11808 }, { "epoch": 17.67814371257485, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1485, "step": 11809 }, { "epoch": 17.679640718562876, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 11810 }, { "epoch": 17.681137724550897, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.141, "step": 11811 }, { "epoch": 17.682634730538922, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1413, "step": 11812 }, { "epoch": 17.684131736526947, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1412, "step": 11813 }, { "epoch": 17.68562874251497, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1405, "step": 11814 }, { "epoch": 17.687125748502993, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1441, "step": 11815 }, { "epoch": 17.688622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1371, "step": 11816 }, { "epoch": 17.690119760479043, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1413, "step": 11817 }, { "epoch": 17.691616766467067, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.139, "step": 11818 }, { "epoch": 17.69311377245509, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1386, "step": 11819 }, { "epoch": 17.694610778443113, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1365, "step": 11820 }, { "epoch": 17.69610778443114, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 11821 }, { "epoch": 17.697604790419163, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1483, "step": 11822 }, { "epoch": 17.699101796407184, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1472, "step": 11823 }, { "epoch": 17.70059880239521, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1424, "step": 11824 }, { "epoch": 17.702095808383234, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1385, "step": 11825 }, { "epoch": 17.70359281437126, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1431, "step": 11826 }, { "epoch": 17.70508982035928, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 11827 }, { "epoch": 17.706586826347305, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1442, "step": 11828 }, { "epoch": 17.70808383233533, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1396, "step": 11829 }, { "epoch": 17.709580838323355, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1435, "step": 11830 }, { "epoch": 17.711077844311376, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1449, "step": 11831 }, { "epoch": 17.7125748502994, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1408, "step": 11832 }, { "epoch": 17.714071856287426, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1403, "step": 11833 }, { "epoch": 17.71556886227545, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1418, "step": 11834 }, { "epoch": 17.71706586826347, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1461, "step": 11835 }, { "epoch": 17.718562874251496, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1414, "step": 11836 }, { "epoch": 17.72005988023952, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1428, "step": 11837 }, { "epoch": 17.721556886227546, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 11838 }, { "epoch": 17.723053892215567, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1523, "step": 11839 }, { "epoch": 17.724550898203592, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 11840 }, { "epoch": 17.726047904191617, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1419, "step": 11841 }, { "epoch": 17.727544910179642, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1323, "step": 11842 }, { "epoch": 17.729041916167663, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1363, "step": 11843 }, { "epoch": 17.730538922155688, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1448, "step": 11844 }, { "epoch": 17.732035928143713, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1399, "step": 11845 }, { "epoch": 17.733532934131738, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1402, "step": 11846 }, { "epoch": 17.73502994011976, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1455, "step": 11847 }, { "epoch": 17.736526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 11848 }, { "epoch": 17.73802395209581, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1397, "step": 11849 }, { "epoch": 17.739520958083833, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1407, "step": 11850 }, { "epoch": 17.741017964071855, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1408, "step": 11851 }, { "epoch": 17.74251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 11852 }, { "epoch": 17.744011976047904, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1383, "step": 11853 }, { "epoch": 17.74550898203593, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1421, "step": 11854 }, { "epoch": 17.74700598802395, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1396, "step": 11855 }, { "epoch": 17.748502994011975, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1388, "step": 11856 }, { "epoch": 17.75, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.143, "step": 11857 }, { "epoch": 17.751497005988025, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1376, "step": 11858 }, { "epoch": 17.75299401197605, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1403, "step": 11859 }, { "epoch": 17.75449101796407, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1482, "step": 11860 }, { "epoch": 17.755988023952096, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 11861 }, { "epoch": 17.75748502994012, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 11862 }, { "epoch": 17.758982035928145, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 11863 }, { "epoch": 17.760479041916167, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1433, "step": 11864 }, { "epoch": 17.76197604790419, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1449, "step": 11865 }, { "epoch": 17.763473053892216, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1351, "step": 11866 }, { "epoch": 17.76497005988024, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1396, "step": 11867 }, { "epoch": 17.766467065868262, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1434, "step": 11868 }, { "epoch": 17.767964071856287, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1394, "step": 11869 }, { "epoch": 17.769461077844312, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1479, "step": 11870 }, { "epoch": 17.770958083832337, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1407, "step": 11871 }, { "epoch": 17.772455089820358, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1424, "step": 11872 }, { "epoch": 17.773952095808383, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1383, "step": 11873 }, { "epoch": 17.775449101796408, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1451, "step": 11874 }, { "epoch": 17.776946107784433, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1344, "step": 11875 }, { "epoch": 17.778443113772454, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 11876 }, { "epoch": 17.77994011976048, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 11877 }, { "epoch": 17.781437125748504, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1454, "step": 11878 }, { "epoch": 17.78293413173653, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1428, "step": 11879 }, { "epoch": 17.78443113772455, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1405, "step": 11880 }, { "epoch": 17.785928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1443, "step": 11881 }, { "epoch": 17.7874251497006, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 11882 }, { "epoch": 17.788922155688624, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1424, "step": 11883 }, { "epoch": 17.790419161676645, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.136, "step": 11884 }, { "epoch": 17.79191616766467, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1442, "step": 11885 }, { "epoch": 17.793413173652695, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.14, "step": 11886 }, { "epoch": 17.79491017964072, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1357, "step": 11887 }, { "epoch": 17.79640718562874, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1382, "step": 11888 }, { "epoch": 17.797904191616766, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1375, "step": 11889 }, { "epoch": 17.79940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 11890 }, { "epoch": 17.800898203592816, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1447, "step": 11891 }, { "epoch": 17.802395209580837, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1361, "step": 11892 }, { "epoch": 17.80389221556886, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1367, "step": 11893 }, { "epoch": 17.805389221556887, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1398, "step": 11894 }, { "epoch": 17.80688622754491, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1443, "step": 11895 }, { "epoch": 17.808383233532933, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.143, "step": 11896 }, { "epoch": 17.809880239520957, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.138, "step": 11897 }, { "epoch": 17.811377245508982, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1393, "step": 11898 }, { "epoch": 17.812874251497007, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.132, "step": 11899 }, { "epoch": 17.81437125748503, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1441, "step": 11900 }, { "epoch": 17.815868263473053, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1398, "step": 11901 }, { "epoch": 17.817365269461078, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.141, "step": 11902 }, { "epoch": 17.818862275449103, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1403, "step": 11903 }, { "epoch": 17.820359281437124, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1435, "step": 11904 }, { "epoch": 17.82185628742515, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1504, "step": 11905 }, { "epoch": 17.823353293413174, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1364, "step": 11906 }, { "epoch": 17.8248502994012, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1391, "step": 11907 }, { "epoch": 17.82634730538922, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1454, "step": 11908 }, { "epoch": 17.827844311377245, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1405, "step": 11909 }, { "epoch": 17.82934131736527, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1475, "step": 11910 }, { "epoch": 17.830838323353294, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1416, "step": 11911 }, { "epoch": 17.83233532934132, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1377, "step": 11912 }, { "epoch": 17.83383233532934, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1416, "step": 11913 }, { "epoch": 17.835329341317365, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1419, "step": 11914 }, { "epoch": 17.83682634730539, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1389, "step": 11915 }, { "epoch": 17.83832335329341, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1429, "step": 11916 }, { "epoch": 17.839820359281436, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1432, "step": 11917 }, { "epoch": 17.84131736526946, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1372, "step": 11918 }, { "epoch": 17.842814371257486, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1403, "step": 11919 }, { "epoch": 17.84431137724551, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1423, "step": 11920 }, { "epoch": 17.845808383233532, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1408, "step": 11921 }, { "epoch": 17.847305389221557, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 11922 }, { "epoch": 17.84880239520958, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1345, "step": 11923 }, { "epoch": 17.850299401197606, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1328, "step": 11924 }, { "epoch": 17.851796407185628, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1372, "step": 11925 }, { "epoch": 17.853293413173652, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1406, "step": 11926 }, { "epoch": 17.854790419161677, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1409, "step": 11927 }, { "epoch": 17.856287425149702, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1355, "step": 11928 }, { "epoch": 17.857784431137723, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 11929 }, { "epoch": 17.85928143712575, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.137, "step": 11930 }, { "epoch": 17.860778443113773, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1442, "step": 11931 }, { "epoch": 17.862275449101798, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1418, "step": 11932 }, { "epoch": 17.86377245508982, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1404, "step": 11933 }, { "epoch": 17.865269461077844, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1341, "step": 11934 }, { "epoch": 17.86676646706587, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1477, "step": 11935 }, { "epoch": 17.868263473053894, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1425, "step": 11936 }, { "epoch": 17.869760479041915, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1375, "step": 11937 }, { "epoch": 17.87125748502994, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.138, "step": 11938 }, { "epoch": 17.872754491017965, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.141, "step": 11939 }, { "epoch": 17.87425149700599, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1345, "step": 11940 }, { "epoch": 17.87574850299401, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1448, "step": 11941 }, { "epoch": 17.877245508982035, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1425, "step": 11942 }, { "epoch": 17.87874251497006, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1412, "step": 11943 }, { "epoch": 17.880239520958085, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.151, "step": 11944 }, { "epoch": 17.881736526946106, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1405, "step": 11945 }, { "epoch": 17.88323353293413, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 11946 }, { "epoch": 17.884730538922156, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1374, "step": 11947 }, { "epoch": 17.88622754491018, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1439, "step": 11948 }, { "epoch": 17.887724550898202, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1369, "step": 11949 }, { "epoch": 17.889221556886227, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 11950 }, { "epoch": 17.89071856287425, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1458, "step": 11951 }, { "epoch": 17.892215568862277, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1404, "step": 11952 }, { "epoch": 17.893712574850298, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.139, "step": 11953 }, { "epoch": 17.895209580838323, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1475, "step": 11954 }, { "epoch": 17.896706586826348, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1438, "step": 11955 }, { "epoch": 17.898203592814372, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.139, "step": 11956 }, { "epoch": 17.899700598802394, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1386, "step": 11957 }, { "epoch": 17.90119760479042, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1386, "step": 11958 }, { "epoch": 17.902694610778443, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1453, "step": 11959 }, { "epoch": 17.904191616766468, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.137, "step": 11960 }, { "epoch": 17.90568862275449, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1391, "step": 11961 }, { "epoch": 17.907185628742514, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1449, "step": 11962 }, { "epoch": 17.90868263473054, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1401, "step": 11963 }, { "epoch": 17.910179640718564, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1325, "step": 11964 }, { "epoch": 17.91167664670659, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.143, "step": 11965 }, { "epoch": 17.91317365269461, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 11966 }, { "epoch": 17.914670658682635, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1461, "step": 11967 }, { "epoch": 17.91616766467066, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1405, "step": 11968 }, { "epoch": 17.91766467065868, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.144, "step": 11969 }, { "epoch": 17.919161676646706, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1474, "step": 11970 }, { "epoch": 17.92065868263473, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1355, "step": 11971 }, { "epoch": 17.922155688622755, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1409, "step": 11972 }, { "epoch": 17.92365269461078, "grad_norm": 0.1787109375, "learning_rate": 0.0008, "loss": 1.1415, "step": 11973 }, { "epoch": 17.9251497005988, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1452, "step": 11974 }, { "epoch": 17.926646706586826, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1364, "step": 11975 }, { "epoch": 17.92814371257485, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1411, "step": 11976 }, { "epoch": 17.929640718562876, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1461, "step": 11977 }, { "epoch": 17.931137724550897, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1497, "step": 11978 }, { "epoch": 17.932634730538922, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 11979 }, { "epoch": 17.934131736526947, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 11980 }, { "epoch": 17.93562874251497, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1367, "step": 11981 }, { "epoch": 17.937125748502993, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1451, "step": 11982 }, { "epoch": 17.938622754491018, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1413, "step": 11983 }, { "epoch": 17.940119760479043, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1455, "step": 11984 }, { "epoch": 17.941616766467067, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1467, "step": 11985 }, { "epoch": 17.94311377245509, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1368, "step": 11986 }, { "epoch": 17.944610778443113, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1354, "step": 11987 }, { "epoch": 17.94610778443114, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1455, "step": 11988 }, { "epoch": 17.947604790419163, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1361, "step": 11989 }, { "epoch": 17.949101796407184, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.138, "step": 11990 }, { "epoch": 17.95059880239521, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1428, "step": 11991 }, { "epoch": 17.952095808383234, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1367, "step": 11992 }, { "epoch": 17.95359281437126, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1364, "step": 11993 }, { "epoch": 17.95508982035928, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1439, "step": 11994 }, { "epoch": 17.956586826347305, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1414, "step": 11995 }, { "epoch": 17.95808383233533, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.147, "step": 11996 }, { "epoch": 17.959580838323355, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1399, "step": 11997 }, { "epoch": 17.961077844311376, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 11998 }, { "epoch": 17.9625748502994, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1406, "step": 11999 }, { "epoch": 17.964071856287426, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.142, "step": 12000 }, { "epoch": 17.96556886227545, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1469, "step": 12001 }, { "epoch": 17.96706586826347, "grad_norm": 0.2314453125, "learning_rate": 0.0008, "loss": 1.1497, "step": 12002 }, { "epoch": 17.968562874251496, "grad_norm": 0.251953125, "learning_rate": 0.0008, "loss": 1.1484, "step": 12003 }, { "epoch": 17.97005988023952, "grad_norm": 0.337890625, "learning_rate": 0.0008, "loss": 1.1496, "step": 12004 }, { "epoch": 17.971556886227546, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.1488, "step": 12005 }, { "epoch": 17.973053892215567, "grad_norm": 0.2265625, "learning_rate": 0.0008, "loss": 1.1412, "step": 12006 }, { "epoch": 17.974550898203592, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1473, "step": 12007 }, { "epoch": 17.976047904191617, "grad_norm": 0.2197265625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12008 }, { "epoch": 17.977544910179642, "grad_norm": 0.2265625, "learning_rate": 0.0008, "loss": 1.1439, "step": 12009 }, { "epoch": 17.979041916167663, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.142, "step": 12010 }, { "epoch": 17.980538922155688, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.1466, "step": 12011 }, { "epoch": 17.982035928143713, "grad_norm": 0.201171875, "learning_rate": 0.0008, "loss": 1.1401, "step": 12012 }, { "epoch": 17.983532934131738, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1367, "step": 12013 }, { "epoch": 17.98502994011976, "grad_norm": 0.189453125, "learning_rate": 0.0008, "loss": 1.1391, "step": 12014 }, { "epoch": 17.986526946107784, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1421, "step": 12015 }, { "epoch": 17.98802395209581, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1431, "step": 12016 }, { "epoch": 17.989520958083833, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1421, "step": 12017 }, { "epoch": 17.991017964071855, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1448, "step": 12018 }, { "epoch": 17.99251497005988, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1355, "step": 12019 }, { "epoch": 17.994011976047904, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 12020 }, { "epoch": 17.99550898203593, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1451, "step": 12021 }, { "epoch": 17.99700598802395, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1428, "step": 12022 }, { "epoch": 17.998502994011975, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1452, "step": 12023 }, { "epoch": 18.0, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1394, "step": 12024 }, { "epoch": 18.001497005988025, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1368, "step": 12025 }, { "epoch": 18.00299401197605, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1372, "step": 12026 }, { "epoch": 18.00449101796407, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 12027 }, { "epoch": 18.005988023952096, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1422, "step": 12028 }, { "epoch": 18.00748502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1406, "step": 12029 }, { "epoch": 18.008982035928145, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1355, "step": 12030 }, { "epoch": 18.010479041916167, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1426, "step": 12031 }, { "epoch": 18.01197604790419, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.14, "step": 12032 }, { "epoch": 18.013473053892216, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12033 }, { "epoch": 18.01497005988024, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1406, "step": 12034 }, { "epoch": 18.016467065868262, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12035 }, { "epoch": 18.017964071856287, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12036 }, { "epoch": 18.019461077844312, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.141, "step": 12037 }, { "epoch": 18.020958083832337, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1366, "step": 12038 }, { "epoch": 18.022455089820358, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1398, "step": 12039 }, { "epoch": 18.023952095808383, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.139, "step": 12040 }, { "epoch": 18.025449101796408, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.137, "step": 12041 }, { "epoch": 18.026946107784433, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1414, "step": 12042 }, { "epoch": 18.028443113772454, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1476, "step": 12043 }, { "epoch": 18.02994011976048, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1429, "step": 12044 }, { "epoch": 18.031437125748504, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12045 }, { "epoch": 18.03293413173653, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1427, "step": 12046 }, { "epoch": 18.03443113772455, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1393, "step": 12047 }, { "epoch": 18.035928143712574, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1379, "step": 12048 }, { "epoch": 18.0374251497006, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1355, "step": 12049 }, { "epoch": 18.038922155688624, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1358, "step": 12050 }, { "epoch": 18.040419161676645, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1424, "step": 12051 }, { "epoch": 18.04191616766467, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1419, "step": 12052 }, { "epoch": 18.043413173652695, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1399, "step": 12053 }, { "epoch": 18.04491017964072, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1433, "step": 12054 }, { "epoch": 18.04640718562874, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1385, "step": 12055 }, { "epoch": 18.047904191616766, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12056 }, { "epoch": 18.04940119760479, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.141, "step": 12057 }, { "epoch": 18.050898203592816, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1454, "step": 12058 }, { "epoch": 18.052395209580837, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1404, "step": 12059 }, { "epoch": 18.05389221556886, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1385, "step": 12060 }, { "epoch": 18.055389221556887, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12061 }, { "epoch": 18.05688622754491, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1398, "step": 12062 }, { "epoch": 18.058383233532933, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1418, "step": 12063 }, { "epoch": 18.059880239520957, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1348, "step": 12064 }, { "epoch": 18.061377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1368, "step": 12065 }, { "epoch": 18.062874251497007, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1433, "step": 12066 }, { "epoch": 18.06437125748503, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12067 }, { "epoch": 18.065868263473053, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1358, "step": 12068 }, { "epoch": 18.067365269461078, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 12069 }, { "epoch": 18.068862275449103, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1376, "step": 12070 }, { "epoch": 18.070359281437124, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1371, "step": 12071 }, { "epoch": 18.07185628742515, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1449, "step": 12072 }, { "epoch": 18.073353293413174, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1457, "step": 12073 }, { "epoch": 18.0748502994012, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1335, "step": 12074 }, { "epoch": 18.07634730538922, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1328, "step": 12075 }, { "epoch": 18.077844311377245, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1344, "step": 12076 }, { "epoch": 18.07934131736527, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1406, "step": 12077 }, { "epoch": 18.080838323353294, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1464, "step": 12078 }, { "epoch": 18.082335329341316, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.142, "step": 12079 }, { "epoch": 18.08383233532934, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1406, "step": 12080 }, { "epoch": 18.085329341317365, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1425, "step": 12081 }, { "epoch": 18.08682634730539, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1456, "step": 12082 }, { "epoch": 18.088323353293415, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12083 }, { "epoch": 18.089820359281436, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1448, "step": 12084 }, { "epoch": 18.09131736526946, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12085 }, { "epoch": 18.092814371257486, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.145, "step": 12086 }, { "epoch": 18.09431137724551, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1419, "step": 12087 }, { "epoch": 18.095808383233532, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1455, "step": 12088 }, { "epoch": 18.097305389221557, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1411, "step": 12089 }, { "epoch": 18.09880239520958, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1504, "step": 12090 }, { "epoch": 18.100299401197606, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1392, "step": 12091 }, { "epoch": 18.101796407185628, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1395, "step": 12092 }, { "epoch": 18.103293413173652, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1332, "step": 12093 }, { "epoch": 18.104790419161677, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12094 }, { "epoch": 18.106287425149702, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 12095 }, { "epoch": 18.107784431137723, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 12096 }, { "epoch": 18.10928143712575, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1362, "step": 12097 }, { "epoch": 18.110778443113773, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1383, "step": 12098 }, { "epoch": 18.112275449101798, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1367, "step": 12099 }, { "epoch": 18.11377245508982, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1411, "step": 12100 }, { "epoch": 18.115269461077844, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12101 }, { "epoch": 18.11676646706587, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1348, "step": 12102 }, { "epoch": 18.118263473053894, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1358, "step": 12103 }, { "epoch": 18.119760479041915, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1419, "step": 12104 }, { "epoch": 18.12125748502994, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 12105 }, { "epoch": 18.122754491017965, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.134, "step": 12106 }, { "epoch": 18.12425149700599, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 12107 }, { "epoch": 18.12574850299401, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1442, "step": 12108 }, { "epoch": 18.127245508982035, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1359, "step": 12109 }, { "epoch": 18.12874251497006, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1359, "step": 12110 }, { "epoch": 18.130239520958085, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.14, "step": 12111 }, { "epoch": 18.131736526946106, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1382, "step": 12112 }, { "epoch": 18.13323353293413, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.14, "step": 12113 }, { "epoch": 18.134730538922156, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1388, "step": 12114 }, { "epoch": 18.13622754491018, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1437, "step": 12115 }, { "epoch": 18.137724550898202, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12116 }, { "epoch": 18.139221556886227, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1467, "step": 12117 }, { "epoch": 18.14071856287425, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1397, "step": 12118 }, { "epoch": 18.142215568862277, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.136, "step": 12119 }, { "epoch": 18.143712574850298, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1436, "step": 12120 }, { "epoch": 18.145209580838323, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1411, "step": 12121 }, { "epoch": 18.146706586826348, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1374, "step": 12122 }, { "epoch": 18.148203592814372, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1395, "step": 12123 }, { "epoch": 18.149700598802394, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12124 }, { "epoch": 18.15119760479042, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 12125 }, { "epoch": 18.152694610778443, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.137, "step": 12126 }, { "epoch": 18.154191616766468, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1366, "step": 12127 }, { "epoch": 18.15568862275449, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1444, "step": 12128 }, { "epoch": 18.157185628742514, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1381, "step": 12129 }, { "epoch": 18.15868263473054, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1337, "step": 12130 }, { "epoch": 18.160179640718564, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1397, "step": 12131 }, { "epoch": 18.161676646706585, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.137, "step": 12132 }, { "epoch": 18.16317365269461, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1381, "step": 12133 }, { "epoch": 18.164670658682635, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1362, "step": 12134 }, { "epoch": 18.16616766467066, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1467, "step": 12135 }, { "epoch": 18.16766467065868, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1398, "step": 12136 }, { "epoch": 18.169161676646706, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1384, "step": 12137 }, { "epoch": 18.17065868263473, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1383, "step": 12138 }, { "epoch": 18.172155688622755, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1386, "step": 12139 }, { "epoch": 18.17365269461078, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1381, "step": 12140 }, { "epoch": 18.1751497005988, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1378, "step": 12141 }, { "epoch": 18.176646706586826, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.143, "step": 12142 }, { "epoch": 18.17814371257485, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1396, "step": 12143 }, { "epoch": 18.179640718562876, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1411, "step": 12144 }, { "epoch": 18.181137724550897, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 12145 }, { "epoch": 18.182634730538922, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1402, "step": 12146 }, { "epoch": 18.184131736526947, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1385, "step": 12147 }, { "epoch": 18.18562874251497, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1359, "step": 12148 }, { "epoch": 18.187125748502993, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12149 }, { "epoch": 18.188622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1357, "step": 12150 }, { "epoch": 18.190119760479043, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1323, "step": 12151 }, { "epoch": 18.191616766467067, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1394, "step": 12152 }, { "epoch": 18.19311377245509, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1432, "step": 12153 }, { "epoch": 18.194610778443113, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1442, "step": 12154 }, { "epoch": 18.19610778443114, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1355, "step": 12155 }, { "epoch": 18.197604790419163, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.143, "step": 12156 }, { "epoch": 18.199101796407184, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 12157 }, { "epoch": 18.20059880239521, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1424, "step": 12158 }, { "epoch": 18.202095808383234, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1346, "step": 12159 }, { "epoch": 18.20359281437126, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1398, "step": 12160 }, { "epoch": 18.20508982035928, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1436, "step": 12161 }, { "epoch": 18.206586826347305, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1482, "step": 12162 }, { "epoch": 18.20808383233533, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1398, "step": 12163 }, { "epoch": 18.209580838323355, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12164 }, { "epoch": 18.211077844311376, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1397, "step": 12165 }, { "epoch": 18.2125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1417, "step": 12166 }, { "epoch": 18.214071856287426, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1341, "step": 12167 }, { "epoch": 18.21556886227545, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1302, "step": 12168 }, { "epoch": 18.21706586826347, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1421, "step": 12169 }, { "epoch": 18.218562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1416, "step": 12170 }, { "epoch": 18.22005988023952, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1408, "step": 12171 }, { "epoch": 18.221556886227546, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1438, "step": 12172 }, { "epoch": 18.223053892215567, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.149, "step": 12173 }, { "epoch": 18.224550898203592, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1396, "step": 12174 }, { "epoch": 18.226047904191617, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1426, "step": 12175 }, { "epoch": 18.227544910179642, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1388, "step": 12176 }, { "epoch": 18.229041916167663, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1394, "step": 12177 }, { "epoch": 18.230538922155688, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1406, "step": 12178 }, { "epoch": 18.232035928143713, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.137, "step": 12179 }, { "epoch": 18.233532934131738, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1409, "step": 12180 }, { "epoch": 18.23502994011976, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1486, "step": 12181 }, { "epoch": 18.236526946107784, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1332, "step": 12182 }, { "epoch": 18.23802395209581, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1437, "step": 12183 }, { "epoch": 18.239520958083833, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1444, "step": 12184 }, { "epoch": 18.241017964071855, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12185 }, { "epoch": 18.24251497005988, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12186 }, { "epoch": 18.244011976047904, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.14, "step": 12187 }, { "epoch": 18.24550898203593, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1442, "step": 12188 }, { "epoch": 18.24700598802395, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1401, "step": 12189 }, { "epoch": 18.248502994011975, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1315, "step": 12190 }, { "epoch": 18.25, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1415, "step": 12191 }, { "epoch": 18.251497005988025, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.136, "step": 12192 }, { "epoch": 18.25299401197605, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1454, "step": 12193 }, { "epoch": 18.25449101796407, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1307, "step": 12194 }, { "epoch": 18.255988023952096, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1358, "step": 12195 }, { "epoch": 18.25748502994012, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1415, "step": 12196 }, { "epoch": 18.258982035928145, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1426, "step": 12197 }, { "epoch": 18.260479041916167, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1419, "step": 12198 }, { "epoch": 18.26197604790419, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1491, "step": 12199 }, { "epoch": 18.263473053892216, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1406, "step": 12200 }, { "epoch": 18.26497005988024, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1447, "step": 12201 }, { "epoch": 18.266467065868262, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 12202 }, { "epoch": 18.267964071856287, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12203 }, { "epoch": 18.269461077844312, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1403, "step": 12204 }, { "epoch": 18.270958083832337, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1398, "step": 12205 }, { "epoch": 18.272455089820358, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1366, "step": 12206 }, { "epoch": 18.273952095808383, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1428, "step": 12207 }, { "epoch": 18.275449101796408, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1382, "step": 12208 }, { "epoch": 18.276946107784433, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1429, "step": 12209 }, { "epoch": 18.278443113772454, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1388, "step": 12210 }, { "epoch": 18.27994011976048, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1305, "step": 12211 }, { "epoch": 18.281437125748504, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1359, "step": 12212 }, { "epoch": 18.28293413173653, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1344, "step": 12213 }, { "epoch": 18.28443113772455, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1374, "step": 12214 }, { "epoch": 18.285928143712574, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 12215 }, { "epoch": 18.2874251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1416, "step": 12216 }, { "epoch": 18.288922155688624, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1424, "step": 12217 }, { "epoch": 18.290419161676645, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1369, "step": 12218 }, { "epoch": 18.29191616766467, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1392, "step": 12219 }, { "epoch": 18.293413173652695, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12220 }, { "epoch": 18.29491017964072, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1425, "step": 12221 }, { "epoch": 18.29640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1389, "step": 12222 }, { "epoch": 18.297904191616766, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1418, "step": 12223 }, { "epoch": 18.29940119760479, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1375, "step": 12224 }, { "epoch": 18.300898203592816, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1429, "step": 12225 }, { "epoch": 18.302395209580837, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1361, "step": 12226 }, { "epoch": 18.30389221556886, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1462, "step": 12227 }, { "epoch": 18.305389221556887, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.137, "step": 12228 }, { "epoch": 18.30688622754491, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.138, "step": 12229 }, { "epoch": 18.308383233532933, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1348, "step": 12230 }, { "epoch": 18.309880239520957, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1347, "step": 12231 }, { "epoch": 18.311377245508982, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1361, "step": 12232 }, { "epoch": 18.312874251497007, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.139, "step": 12233 }, { "epoch": 18.31437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1372, "step": 12234 }, { "epoch": 18.315868263473053, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1423, "step": 12235 }, { "epoch": 18.317365269461078, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1355, "step": 12236 }, { "epoch": 18.318862275449103, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12237 }, { "epoch": 18.320359281437124, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1394, "step": 12238 }, { "epoch": 18.32185628742515, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1414, "step": 12239 }, { "epoch": 18.323353293413174, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1376, "step": 12240 }, { "epoch": 18.3248502994012, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1374, "step": 12241 }, { "epoch": 18.32634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12242 }, { "epoch": 18.327844311377245, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.136, "step": 12243 }, { "epoch": 18.32934131736527, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.146, "step": 12244 }, { "epoch": 18.330838323353294, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1415, "step": 12245 }, { "epoch": 18.33233532934132, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1413, "step": 12246 }, { "epoch": 18.33383233532934, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1393, "step": 12247 }, { "epoch": 18.335329341317365, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1331, "step": 12248 }, { "epoch": 18.33682634730539, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1374, "step": 12249 }, { "epoch": 18.338323353293415, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 12250 }, { "epoch": 18.339820359281436, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.138, "step": 12251 }, { "epoch": 18.34131736526946, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1368, "step": 12252 }, { "epoch": 18.342814371257486, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1351, "step": 12253 }, { "epoch": 18.34431137724551, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1359, "step": 12254 }, { "epoch": 18.345808383233532, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 12255 }, { "epoch": 18.347305389221557, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1407, "step": 12256 }, { "epoch": 18.34880239520958, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1418, "step": 12257 }, { "epoch": 18.350299401197606, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 12258 }, { "epoch": 18.351796407185628, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1356, "step": 12259 }, { "epoch": 18.353293413173652, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1424, "step": 12260 }, { "epoch": 18.354790419161677, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1407, "step": 12261 }, { "epoch": 18.356287425149702, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1344, "step": 12262 }, { "epoch": 18.357784431137723, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1406, "step": 12263 }, { "epoch": 18.35928143712575, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1444, "step": 12264 }, { "epoch": 18.360778443113773, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 12265 }, { "epoch": 18.362275449101798, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1421, "step": 12266 }, { "epoch": 18.36377245508982, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.138, "step": 12267 }, { "epoch": 18.365269461077844, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.138, "step": 12268 }, { "epoch": 18.36676646706587, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1353, "step": 12269 }, { "epoch": 18.368263473053894, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1341, "step": 12270 }, { "epoch": 18.369760479041915, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1368, "step": 12271 }, { "epoch": 18.37125748502994, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1391, "step": 12272 }, { "epoch": 18.372754491017965, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1442, "step": 12273 }, { "epoch": 18.37425149700599, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.132, "step": 12274 }, { "epoch": 18.37574850299401, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1354, "step": 12275 }, { "epoch": 18.377245508982035, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1351, "step": 12276 }, { "epoch": 18.37874251497006, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.1424, "step": 12277 }, { "epoch": 18.380239520958085, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1438, "step": 12278 }, { "epoch": 18.381736526946106, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1345, "step": 12279 }, { "epoch": 18.38323353293413, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1439, "step": 12280 }, { "epoch": 18.384730538922156, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1347, "step": 12281 }, { "epoch": 18.38622754491018, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1374, "step": 12282 }, { "epoch": 18.387724550898202, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12283 }, { "epoch": 18.389221556886227, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.1306, "step": 12284 }, { "epoch": 18.39071856287425, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1385, "step": 12285 }, { "epoch": 18.392215568862277, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1415, "step": 12286 }, { "epoch": 18.393712574850298, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1379, "step": 12287 }, { "epoch": 18.395209580838323, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.1376, "step": 12288 }, { "epoch": 18.396706586826348, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.142, "step": 12289 }, { "epoch": 18.398203592814372, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 12290 }, { "epoch": 18.399700598802394, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1399, "step": 12291 }, { "epoch": 18.40119760479042, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1474, "step": 12292 }, { "epoch": 18.402694610778443, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1399, "step": 12293 }, { "epoch": 18.404191616766468, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.145, "step": 12294 }, { "epoch": 18.40568862275449, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 12295 }, { "epoch": 18.407185628742514, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1342, "step": 12296 }, { "epoch": 18.40868263473054, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1411, "step": 12297 }, { "epoch": 18.410179640718564, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1404, "step": 12298 }, { "epoch": 18.411676646706585, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1435, "step": 12299 }, { "epoch": 18.41317365269461, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1361, "step": 12300 }, { "epoch": 18.414670658682635, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1332, "step": 12301 }, { "epoch": 18.41616766467066, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1384, "step": 12302 }, { "epoch": 18.41766467065868, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1423, "step": 12303 }, { "epoch": 18.419161676646706, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.136, "step": 12304 }, { "epoch": 18.42065868263473, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1423, "step": 12305 }, { "epoch": 18.422155688622755, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1435, "step": 12306 }, { "epoch": 18.42365269461078, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1339, "step": 12307 }, { "epoch": 18.4251497005988, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1395, "step": 12308 }, { "epoch": 18.426646706586826, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1447, "step": 12309 }, { "epoch": 18.42814371257485, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1365, "step": 12310 }, { "epoch": 18.429640718562876, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1436, "step": 12311 }, { "epoch": 18.431137724550897, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1375, "step": 12312 }, { "epoch": 18.432634730538922, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1393, "step": 12313 }, { "epoch": 18.434131736526947, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1382, "step": 12314 }, { "epoch": 18.43562874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1334, "step": 12315 }, { "epoch": 18.437125748502993, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1314, "step": 12316 }, { "epoch": 18.438622754491018, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1438, "step": 12317 }, { "epoch": 18.440119760479043, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1406, "step": 12318 }, { "epoch": 18.441616766467067, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1376, "step": 12319 }, { "epoch": 18.44311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.14, "step": 12320 }, { "epoch": 18.444610778443113, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 12321 }, { "epoch": 18.44610778443114, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1313, "step": 12322 }, { "epoch": 18.447604790419163, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1365, "step": 12323 }, { "epoch": 18.449101796407184, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1386, "step": 12324 }, { "epoch": 18.45059880239521, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1348, "step": 12325 }, { "epoch": 18.452095808383234, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1323, "step": 12326 }, { "epoch": 18.45359281437126, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1431, "step": 12327 }, { "epoch": 18.45508982035928, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1334, "step": 12328 }, { "epoch": 18.456586826347305, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1384, "step": 12329 }, { "epoch": 18.45808383233533, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1387, "step": 12330 }, { "epoch": 18.459580838323355, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 12331 }, { "epoch": 18.461077844311376, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1315, "step": 12332 }, { "epoch": 18.4625748502994, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.133, "step": 12333 }, { "epoch": 18.464071856287426, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.137, "step": 12334 }, { "epoch": 18.46556886227545, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1418, "step": 12335 }, { "epoch": 18.46706586826347, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1409, "step": 12336 }, { "epoch": 18.468562874251496, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1393, "step": 12337 }, { "epoch": 18.47005988023952, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1481, "step": 12338 }, { "epoch": 18.471556886227546, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12339 }, { "epoch": 18.473053892215567, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 12340 }, { "epoch": 18.474550898203592, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1446, "step": 12341 }, { "epoch": 18.476047904191617, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1411, "step": 12342 }, { "epoch": 18.477544910179642, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1351, "step": 12343 }, { "epoch": 18.479041916167663, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1333, "step": 12344 }, { "epoch": 18.480538922155688, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1407, "step": 12345 }, { "epoch": 18.482035928143713, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1368, "step": 12346 }, { "epoch": 18.483532934131738, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1384, "step": 12347 }, { "epoch": 18.48502994011976, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1417, "step": 12348 }, { "epoch": 18.486526946107784, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1377, "step": 12349 }, { "epoch": 18.48802395209581, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1351, "step": 12350 }, { "epoch": 18.489520958083833, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12351 }, { "epoch": 18.491017964071855, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1399, "step": 12352 }, { "epoch": 18.49251497005988, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1389, "step": 12353 }, { "epoch": 18.494011976047904, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1404, "step": 12354 }, { "epoch": 18.49550898203593, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.137, "step": 12355 }, { "epoch": 18.49700598802395, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1412, "step": 12356 }, { "epoch": 18.498502994011975, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12357 }, { "epoch": 18.5, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1447, "step": 12358 }, { "epoch": 18.501497005988025, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1414, "step": 12359 }, { "epoch": 18.50299401197605, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1506, "step": 12360 }, { "epoch": 18.50449101796407, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1373, "step": 12361 }, { "epoch": 18.505988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1368, "step": 12362 }, { "epoch": 18.50748502994012, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1385, "step": 12363 }, { "epoch": 18.508982035928145, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.142, "step": 12364 }, { "epoch": 18.510479041916167, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.134, "step": 12365 }, { "epoch": 18.51197604790419, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1456, "step": 12366 }, { "epoch": 18.513473053892216, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 12367 }, { "epoch": 18.51497005988024, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1355, "step": 12368 }, { "epoch": 18.516467065868262, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1408, "step": 12369 }, { "epoch": 18.517964071856287, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1432, "step": 12370 }, { "epoch": 18.519461077844312, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1357, "step": 12371 }, { "epoch": 18.520958083832337, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 12372 }, { "epoch": 18.522455089820358, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1382, "step": 12373 }, { "epoch": 18.523952095808383, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1357, "step": 12374 }, { "epoch": 18.525449101796408, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1407, "step": 12375 }, { "epoch": 18.526946107784433, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1403, "step": 12376 }, { "epoch": 18.528443113772454, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1371, "step": 12377 }, { "epoch": 18.52994011976048, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1381, "step": 12378 }, { "epoch": 18.531437125748504, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1358, "step": 12379 }, { "epoch": 18.53293413173653, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1362, "step": 12380 }, { "epoch": 18.53443113772455, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.14, "step": 12381 }, { "epoch": 18.535928143712574, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1373, "step": 12382 }, { "epoch": 18.5374251497006, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1433, "step": 12383 }, { "epoch": 18.538922155688624, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1377, "step": 12384 }, { "epoch": 18.540419161676645, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1343, "step": 12385 }, { "epoch": 18.54191616766467, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1456, "step": 12386 }, { "epoch": 18.543413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.136, "step": 12387 }, { "epoch": 18.54491017964072, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1373, "step": 12388 }, { "epoch": 18.54640718562874, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 12389 }, { "epoch": 18.547904191616766, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1401, "step": 12390 }, { "epoch": 18.54940119760479, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1395, "step": 12391 }, { "epoch": 18.550898203592816, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1347, "step": 12392 }, { "epoch": 18.552395209580837, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.137, "step": 12393 }, { "epoch": 18.55389221556886, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1462, "step": 12394 }, { "epoch": 18.555389221556887, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 12395 }, { "epoch": 18.55688622754491, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1419, "step": 12396 }, { "epoch": 18.558383233532933, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1436, "step": 12397 }, { "epoch": 18.559880239520957, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 12398 }, { "epoch": 18.561377245508982, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1377, "step": 12399 }, { "epoch": 18.562874251497007, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1364, "step": 12400 }, { "epoch": 18.56437125748503, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1379, "step": 12401 }, { "epoch": 18.565868263473053, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1425, "step": 12402 }, { "epoch": 18.567365269461078, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1423, "step": 12403 }, { "epoch": 18.568862275449103, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1378, "step": 12404 }, { "epoch": 18.570359281437124, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1354, "step": 12405 }, { "epoch": 18.57185628742515, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 12406 }, { "epoch": 18.573353293413174, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1465, "step": 12407 }, { "epoch": 18.5748502994012, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1293, "step": 12408 }, { "epoch": 18.57634730538922, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.137, "step": 12409 }, { "epoch": 18.577844311377245, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1396, "step": 12410 }, { "epoch": 18.57934131736527, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1351, "step": 12411 }, { "epoch": 18.580838323353294, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 12412 }, { "epoch": 18.58233532934132, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1408, "step": 12413 }, { "epoch": 18.58383233532934, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.138, "step": 12414 }, { "epoch": 18.585329341317365, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.133, "step": 12415 }, { "epoch": 18.58682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 12416 }, { "epoch": 18.58832335329341, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1402, "step": 12417 }, { "epoch": 18.589820359281436, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1353, "step": 12418 }, { "epoch": 18.59131736526946, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.142, "step": 12419 }, { "epoch": 18.592814371257486, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1389, "step": 12420 }, { "epoch": 18.59431137724551, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1394, "step": 12421 }, { "epoch": 18.595808383233532, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1412, "step": 12422 }, { "epoch": 18.597305389221557, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 12423 }, { "epoch": 18.59880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1358, "step": 12424 }, { "epoch": 18.600299401197606, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 12425 }, { "epoch": 18.601796407185628, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1423, "step": 12426 }, { "epoch": 18.603293413173652, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1402, "step": 12427 }, { "epoch": 18.604790419161677, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1319, "step": 12428 }, { "epoch": 18.606287425149702, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1383, "step": 12429 }, { "epoch": 18.607784431137723, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1396, "step": 12430 }, { "epoch": 18.60928143712575, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1355, "step": 12431 }, { "epoch": 18.610778443113773, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1307, "step": 12432 }, { "epoch": 18.612275449101798, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1423, "step": 12433 }, { "epoch": 18.61377245508982, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.142, "step": 12434 }, { "epoch": 18.615269461077844, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1443, "step": 12435 }, { "epoch": 18.61676646706587, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1384, "step": 12436 }, { "epoch": 18.618263473053894, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1384, "step": 12437 }, { "epoch": 18.619760479041915, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12438 }, { "epoch": 18.62125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1402, "step": 12439 }, { "epoch": 18.622754491017965, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1334, "step": 12440 }, { "epoch": 18.62425149700599, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.136, "step": 12441 }, { "epoch": 18.62574850299401, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1399, "step": 12442 }, { "epoch": 18.627245508982035, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1422, "step": 12443 }, { "epoch": 18.62874251497006, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.138, "step": 12444 }, { "epoch": 18.630239520958085, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1391, "step": 12445 }, { "epoch": 18.631736526946106, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1332, "step": 12446 }, { "epoch": 18.63323353293413, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1333, "step": 12447 }, { "epoch": 18.634730538922156, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1396, "step": 12448 }, { "epoch": 18.63622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1355, "step": 12449 }, { "epoch": 18.637724550898202, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1399, "step": 12450 }, { "epoch": 18.639221556886227, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1432, "step": 12451 }, { "epoch": 18.64071856287425, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1403, "step": 12452 }, { "epoch": 18.642215568862277, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1366, "step": 12453 }, { "epoch": 18.643712574850298, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1429, "step": 12454 }, { "epoch": 18.645209580838323, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1398, "step": 12455 }, { "epoch": 18.646706586826348, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1444, "step": 12456 }, { "epoch": 18.648203592814372, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1387, "step": 12457 }, { "epoch": 18.649700598802394, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1345, "step": 12458 }, { "epoch": 18.65119760479042, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1365, "step": 12459 }, { "epoch": 18.652694610778443, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1364, "step": 12460 }, { "epoch": 18.654191616766468, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1405, "step": 12461 }, { "epoch": 18.65568862275449, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1408, "step": 12462 }, { "epoch": 18.657185628742514, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1325, "step": 12463 }, { "epoch": 18.65868263473054, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1365, "step": 12464 }, { "epoch": 18.660179640718564, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12465 }, { "epoch": 18.66167664670659, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1362, "step": 12466 }, { "epoch": 18.66317365269461, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1363, "step": 12467 }, { "epoch": 18.664670658682635, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 12468 }, { "epoch": 18.66616766467066, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.142, "step": 12469 }, { "epoch": 18.66766467065868, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1445, "step": 12470 }, { "epoch": 18.669161676646706, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1374, "step": 12471 }, { "epoch": 18.67065868263473, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1423, "step": 12472 }, { "epoch": 18.672155688622755, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.134, "step": 12473 }, { "epoch": 18.67365269461078, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1393, "step": 12474 }, { "epoch": 18.6751497005988, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1383, "step": 12475 }, { "epoch": 18.676646706586826, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1358, "step": 12476 }, { "epoch": 18.67814371257485, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1314, "step": 12477 }, { "epoch": 18.679640718562876, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1412, "step": 12478 }, { "epoch": 18.681137724550897, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 12479 }, { "epoch": 18.682634730538922, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1294, "step": 12480 }, { "epoch": 18.684131736526947, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1412, "step": 12481 }, { "epoch": 18.68562874251497, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1401, "step": 12482 }, { "epoch": 18.687125748502993, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1393, "step": 12483 }, { "epoch": 18.688622754491018, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1392, "step": 12484 }, { "epoch": 18.690119760479043, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1333, "step": 12485 }, { "epoch": 18.691616766467067, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1422, "step": 12486 }, { "epoch": 18.69311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1368, "step": 12487 }, { "epoch": 18.694610778443113, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12488 }, { "epoch": 18.69610778443114, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1398, "step": 12489 }, { "epoch": 18.697604790419163, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1375, "step": 12490 }, { "epoch": 18.699101796407184, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 12491 }, { "epoch": 18.70059880239521, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12492 }, { "epoch": 18.702095808383234, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1363, "step": 12493 }, { "epoch": 18.70359281437126, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1358, "step": 12494 }, { "epoch": 18.70508982035928, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 12495 }, { "epoch": 18.706586826347305, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.142, "step": 12496 }, { "epoch": 18.70808383233533, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1386, "step": 12497 }, { "epoch": 18.709580838323355, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1385, "step": 12498 }, { "epoch": 18.711077844311376, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 12499 }, { "epoch": 18.7125748502994, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1452, "step": 12500 }, { "epoch": 18.714071856287426, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1319, "step": 12501 }, { "epoch": 18.71556886227545, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1358, "step": 12502 }, { "epoch": 18.71706586826347, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12503 }, { "epoch": 18.718562874251496, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12504 }, { "epoch": 18.72005988023952, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1425, "step": 12505 }, { "epoch": 18.721556886227546, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1495, "step": 12506 }, { "epoch": 18.723053892215567, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 12507 }, { "epoch": 18.724550898203592, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1364, "step": 12508 }, { "epoch": 18.726047904191617, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1412, "step": 12509 }, { "epoch": 18.727544910179642, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1415, "step": 12510 }, { "epoch": 18.729041916167663, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1382, "step": 12511 }, { "epoch": 18.730538922155688, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1422, "step": 12512 }, { "epoch": 18.732035928143713, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1397, "step": 12513 }, { "epoch": 18.733532934131738, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1359, "step": 12514 }, { "epoch": 18.73502994011976, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.14, "step": 12515 }, { "epoch": 18.736526946107784, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1386, "step": 12516 }, { "epoch": 18.73802395209581, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1329, "step": 12517 }, { "epoch": 18.739520958083833, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1421, "step": 12518 }, { "epoch": 18.741017964071855, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1397, "step": 12519 }, { "epoch": 18.74251497005988, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1448, "step": 12520 }, { "epoch": 18.744011976047904, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1403, "step": 12521 }, { "epoch": 18.74550898203593, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1392, "step": 12522 }, { "epoch": 18.74700598802395, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1343, "step": 12523 }, { "epoch": 18.748502994011975, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1431, "step": 12524 }, { "epoch": 18.75, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1365, "step": 12525 }, { "epoch": 18.751497005988025, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1433, "step": 12526 }, { "epoch": 18.75299401197605, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1388, "step": 12527 }, { "epoch": 18.75449101796407, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1375, "step": 12528 }, { "epoch": 18.755988023952096, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.142, "step": 12529 }, { "epoch": 18.75748502994012, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1375, "step": 12530 }, { "epoch": 18.758982035928145, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.142, "step": 12531 }, { "epoch": 18.760479041916167, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1347, "step": 12532 }, { "epoch": 18.76197604790419, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1423, "step": 12533 }, { "epoch": 18.763473053892216, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1374, "step": 12534 }, { "epoch": 18.76497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1405, "step": 12535 }, { "epoch": 18.766467065868262, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1347, "step": 12536 }, { "epoch": 18.767964071856287, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12537 }, { "epoch": 18.769461077844312, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 12538 }, { "epoch": 18.770958083832337, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1399, "step": 12539 }, { "epoch": 18.772455089820358, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 12540 }, { "epoch": 18.773952095808383, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1422, "step": 12541 }, { "epoch": 18.775449101796408, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1371, "step": 12542 }, { "epoch": 18.776946107784433, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.143, "step": 12543 }, { "epoch": 18.778443113772454, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 12544 }, { "epoch": 18.77994011976048, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1382, "step": 12545 }, { "epoch": 18.781437125748504, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1366, "step": 12546 }, { "epoch": 18.78293413173653, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1395, "step": 12547 }, { "epoch": 18.78443113772455, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.14, "step": 12548 }, { "epoch": 18.785928143712574, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1396, "step": 12549 }, { "epoch": 18.7874251497006, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1478, "step": 12550 }, { "epoch": 18.788922155688624, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1444, "step": 12551 }, { "epoch": 18.790419161676645, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1365, "step": 12552 }, { "epoch": 18.79191616766467, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 12553 }, { "epoch": 18.793413173652695, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1391, "step": 12554 }, { "epoch": 18.79491017964072, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1373, "step": 12555 }, { "epoch": 18.79640718562874, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1368, "step": 12556 }, { "epoch": 18.797904191616766, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12557 }, { "epoch": 18.79940119760479, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1384, "step": 12558 }, { "epoch": 18.800898203592816, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1445, "step": 12559 }, { "epoch": 18.802395209580837, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.141, "step": 12560 }, { "epoch": 18.80389221556886, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1441, "step": 12561 }, { "epoch": 18.805389221556887, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1445, "step": 12562 }, { "epoch": 18.80688622754491, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 12563 }, { "epoch": 18.808383233532933, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1352, "step": 12564 }, { "epoch": 18.809880239520957, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1488, "step": 12565 }, { "epoch": 18.811377245508982, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 12566 }, { "epoch": 18.812874251497007, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1359, "step": 12567 }, { "epoch": 18.81437125748503, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1345, "step": 12568 }, { "epoch": 18.815868263473053, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 12569 }, { "epoch": 18.817365269461078, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1396, "step": 12570 }, { "epoch": 18.818862275449103, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1334, "step": 12571 }, { "epoch": 18.820359281437124, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 12572 }, { "epoch": 18.82185628742515, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1361, "step": 12573 }, { "epoch": 18.823353293413174, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 12574 }, { "epoch": 18.8248502994012, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1392, "step": 12575 }, { "epoch": 18.82634730538922, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1364, "step": 12576 }, { "epoch": 18.827844311377245, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1426, "step": 12577 }, { "epoch": 18.82934131736527, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12578 }, { "epoch": 18.830838323353294, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1391, "step": 12579 }, { "epoch": 18.83233532934132, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1406, "step": 12580 }, { "epoch": 18.83383233532934, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1459, "step": 12581 }, { "epoch": 18.835329341317365, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1379, "step": 12582 }, { "epoch": 18.83682634730539, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1346, "step": 12583 }, { "epoch": 18.83832335329341, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1383, "step": 12584 }, { "epoch": 18.839820359281436, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1327, "step": 12585 }, { "epoch": 18.84131736526946, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 12586 }, { "epoch": 18.842814371257486, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1444, "step": 12587 }, { "epoch": 18.84431137724551, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1335, "step": 12588 }, { "epoch": 18.845808383233532, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1431, "step": 12589 }, { "epoch": 18.847305389221557, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1292, "step": 12590 }, { "epoch": 18.84880239520958, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1429, "step": 12591 }, { "epoch": 18.850299401197606, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1395, "step": 12592 }, { "epoch": 18.851796407185628, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1348, "step": 12593 }, { "epoch": 18.853293413173652, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1337, "step": 12594 }, { "epoch": 18.854790419161677, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1456, "step": 12595 }, { "epoch": 18.856287425149702, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1366, "step": 12596 }, { "epoch": 18.857784431137723, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1374, "step": 12597 }, { "epoch": 18.85928143712575, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1385, "step": 12598 }, { "epoch": 18.860778443113773, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1398, "step": 12599 }, { "epoch": 18.862275449101798, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1336, "step": 12600 }, { "epoch": 18.86377245508982, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1402, "step": 12601 }, { "epoch": 18.865269461077844, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1313, "step": 12602 }, { "epoch": 18.86676646706587, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1413, "step": 12603 }, { "epoch": 18.868263473053894, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1344, "step": 12604 }, { "epoch": 18.869760479041915, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.132, "step": 12605 }, { "epoch": 18.87125748502994, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1423, "step": 12606 }, { "epoch": 18.872754491017965, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1418, "step": 12607 }, { "epoch": 18.87425149700599, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1396, "step": 12608 }, { "epoch": 18.87574850299401, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1354, "step": 12609 }, { "epoch": 18.877245508982035, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 12610 }, { "epoch": 18.87874251497006, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1353, "step": 12611 }, { "epoch": 18.880239520958085, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1385, "step": 12612 }, { "epoch": 18.881736526946106, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1357, "step": 12613 }, { "epoch": 18.88323353293413, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1461, "step": 12614 }, { "epoch": 18.884730538922156, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1348, "step": 12615 }, { "epoch": 18.88622754491018, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1371, "step": 12616 }, { "epoch": 18.887724550898202, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1408, "step": 12617 }, { "epoch": 18.889221556886227, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.133, "step": 12618 }, { "epoch": 18.89071856287425, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1378, "step": 12619 }, { "epoch": 18.892215568862277, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1448, "step": 12620 }, { "epoch": 18.893712574850298, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.139, "step": 12621 }, { "epoch": 18.895209580838323, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 12622 }, { "epoch": 18.896706586826348, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1385, "step": 12623 }, { "epoch": 18.898203592814372, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1362, "step": 12624 }, { "epoch": 18.899700598802394, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1354, "step": 12625 }, { "epoch": 18.90119760479042, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1411, "step": 12626 }, { "epoch": 18.902694610778443, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1399, "step": 12627 }, { "epoch": 18.904191616766468, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1296, "step": 12628 }, { "epoch": 18.90568862275449, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1343, "step": 12629 }, { "epoch": 18.907185628742514, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1392, "step": 12630 }, { "epoch": 18.90868263473054, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1379, "step": 12631 }, { "epoch": 18.910179640718564, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1377, "step": 12632 }, { "epoch": 18.91167664670659, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1409, "step": 12633 }, { "epoch": 18.91317365269461, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1373, "step": 12634 }, { "epoch": 18.914670658682635, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1351, "step": 12635 }, { "epoch": 18.91616766467066, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 12636 }, { "epoch": 18.91766467065868, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1377, "step": 12637 }, { "epoch": 18.919161676646706, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1439, "step": 12638 }, { "epoch": 18.92065868263473, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1394, "step": 12639 }, { "epoch": 18.922155688622755, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 12640 }, { "epoch": 18.92365269461078, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1384, "step": 12641 }, { "epoch": 18.9251497005988, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1425, "step": 12642 }, { "epoch": 18.926646706586826, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1379, "step": 12643 }, { "epoch": 18.92814371257485, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1407, "step": 12644 }, { "epoch": 18.929640718562876, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1399, "step": 12645 }, { "epoch": 18.931137724550897, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1368, "step": 12646 }, { "epoch": 18.932634730538922, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1345, "step": 12647 }, { "epoch": 18.934131736526947, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.144, "step": 12648 }, { "epoch": 18.93562874251497, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1379, "step": 12649 }, { "epoch": 18.937125748502993, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 12650 }, { "epoch": 18.938622754491018, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1402, "step": 12651 }, { "epoch": 18.940119760479043, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1366, "step": 12652 }, { "epoch": 18.941616766467067, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1398, "step": 12653 }, { "epoch": 18.94311377245509, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1438, "step": 12654 }, { "epoch": 18.944610778443113, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.135, "step": 12655 }, { "epoch": 18.94610778443114, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1381, "step": 12656 }, { "epoch": 18.947604790419163, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1388, "step": 12657 }, { "epoch": 18.949101796407184, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1348, "step": 12658 }, { "epoch": 18.95059880239521, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1378, "step": 12659 }, { "epoch": 18.952095808383234, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1403, "step": 12660 }, { "epoch": 18.95359281437126, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1392, "step": 12661 }, { "epoch": 18.95508982035928, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 12662 }, { "epoch": 18.956586826347305, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1406, "step": 12663 }, { "epoch": 18.95808383233533, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1374, "step": 12664 }, { "epoch": 18.959580838323355, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1342, "step": 12665 }, { "epoch": 18.961077844311376, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.137, "step": 12666 }, { "epoch": 18.9625748502994, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1344, "step": 12667 }, { "epoch": 18.964071856287426, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1369, "step": 12668 }, { "epoch": 18.96556886227545, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1376, "step": 12669 }, { "epoch": 18.96706586826347, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1397, "step": 12670 }, { "epoch": 18.968562874251496, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.14, "step": 12671 }, { "epoch": 18.97005988023952, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 12672 }, { "epoch": 18.971556886227546, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1317, "step": 12673 }, { "epoch": 18.973053892215567, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1408, "step": 12674 }, { "epoch": 18.974550898203592, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1361, "step": 12675 }, { "epoch": 18.976047904191617, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1359, "step": 12676 }, { "epoch": 18.977544910179642, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1315, "step": 12677 }, { "epoch": 18.979041916167663, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1398, "step": 12678 }, { "epoch": 18.980538922155688, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1362, "step": 12679 }, { "epoch": 18.982035928143713, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1385, "step": 12680 }, { "epoch": 18.983532934131738, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1338, "step": 12681 }, { "epoch": 18.98502994011976, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1391, "step": 12682 }, { "epoch": 18.986526946107784, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1404, "step": 12683 }, { "epoch": 18.98802395209581, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12684 }, { "epoch": 18.989520958083833, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.14, "step": 12685 }, { "epoch": 18.991017964071855, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1371, "step": 12686 }, { "epoch": 18.99251497005988, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1408, "step": 12687 }, { "epoch": 18.994011976047904, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1363, "step": 12688 }, { "epoch": 18.99550898203593, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1381, "step": 12689 }, { "epoch": 18.99700598802395, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 12690 }, { "epoch": 18.998502994011975, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1448, "step": 12691 }, { "epoch": 19.0, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.136, "step": 12692 }, { "epoch": 19.001497005988025, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1364, "step": 12693 }, { "epoch": 19.00299401197605, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1417, "step": 12694 }, { "epoch": 19.00449101796407, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1391, "step": 12695 }, { "epoch": 19.005988023952096, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 12696 }, { "epoch": 19.00748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1401, "step": 12697 }, { "epoch": 19.008982035928145, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1413, "step": 12698 }, { "epoch": 19.010479041916167, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 12699 }, { "epoch": 19.01197604790419, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1359, "step": 12700 }, { "epoch": 19.013473053892216, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 12701 }, { "epoch": 19.01497005988024, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1365, "step": 12702 }, { "epoch": 19.016467065868262, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1345, "step": 12703 }, { "epoch": 19.017964071856287, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1309, "step": 12704 }, { "epoch": 19.019461077844312, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12705 }, { "epoch": 19.020958083832337, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1362, "step": 12706 }, { "epoch": 19.022455089820358, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1342, "step": 12707 }, { "epoch": 19.023952095808383, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1406, "step": 12708 }, { "epoch": 19.025449101796408, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1389, "step": 12709 }, { "epoch": 19.026946107784433, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1393, "step": 12710 }, { "epoch": 19.028443113772454, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1334, "step": 12711 }, { "epoch": 19.02994011976048, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1373, "step": 12712 }, { "epoch": 19.031437125748504, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 12713 }, { "epoch": 19.03293413173653, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1329, "step": 12714 }, { "epoch": 19.03443113772455, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1385, "step": 12715 }, { "epoch": 19.035928143712574, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1331, "step": 12716 }, { "epoch": 19.0374251497006, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1321, "step": 12717 }, { "epoch": 19.038922155688624, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.139, "step": 12718 }, { "epoch": 19.040419161676645, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1374, "step": 12719 }, { "epoch": 19.04191616766467, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1347, "step": 12720 }, { "epoch": 19.043413173652695, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1321, "step": 12721 }, { "epoch": 19.04491017964072, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1403, "step": 12722 }, { "epoch": 19.04640718562874, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1429, "step": 12723 }, { "epoch": 19.047904191616766, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1372, "step": 12724 }, { "epoch": 19.04940119760479, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1381, "step": 12725 }, { "epoch": 19.050898203592816, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1353, "step": 12726 }, { "epoch": 19.052395209580837, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1349, "step": 12727 }, { "epoch": 19.05389221556886, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1418, "step": 12728 }, { "epoch": 19.055389221556887, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1365, "step": 12729 }, { "epoch": 19.05688622754491, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1351, "step": 12730 }, { "epoch": 19.058383233532933, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1392, "step": 12731 }, { "epoch": 19.059880239520957, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1373, "step": 12732 }, { "epoch": 19.061377245508982, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.137, "step": 12733 }, { "epoch": 19.062874251497007, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1426, "step": 12734 }, { "epoch": 19.06437125748503, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1445, "step": 12735 }, { "epoch": 19.065868263473053, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1313, "step": 12736 }, { "epoch": 19.067365269461078, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1398, "step": 12737 }, { "epoch": 19.068862275449103, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1417, "step": 12738 }, { "epoch": 19.070359281437124, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1472, "step": 12739 }, { "epoch": 19.07185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1315, "step": 12740 }, { "epoch": 19.073353293413174, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1376, "step": 12741 }, { "epoch": 19.0748502994012, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1362, "step": 12742 }, { "epoch": 19.07634730538922, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1367, "step": 12743 }, { "epoch": 19.077844311377245, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.14, "step": 12744 }, { "epoch": 19.07934131736527, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.142, "step": 12745 }, { "epoch": 19.080838323353294, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.138, "step": 12746 }, { "epoch": 19.082335329341316, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1347, "step": 12747 }, { "epoch": 19.08383233532934, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1349, "step": 12748 }, { "epoch": 19.085329341317365, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1446, "step": 12749 }, { "epoch": 19.08682634730539, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 12750 }, { "epoch": 19.088323353293415, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1408, "step": 12751 }, { "epoch": 19.089820359281436, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 12752 }, { "epoch": 19.09131736526946, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1417, "step": 12753 }, { "epoch": 19.092814371257486, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1417, "step": 12754 }, { "epoch": 19.09431137724551, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1368, "step": 12755 }, { "epoch": 19.095808383233532, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1398, "step": 12756 }, { "epoch": 19.097305389221557, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1296, "step": 12757 }, { "epoch": 19.09880239520958, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1474, "step": 12758 }, { "epoch": 19.100299401197606, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1361, "step": 12759 }, { "epoch": 19.101796407185628, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1438, "step": 12760 }, { "epoch": 19.103293413173652, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1421, "step": 12761 }, { "epoch": 19.104790419161677, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.141, "step": 12762 }, { "epoch": 19.106287425149702, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1375, "step": 12763 }, { "epoch": 19.107784431137723, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.132, "step": 12764 }, { "epoch": 19.10928143712575, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1426, "step": 12765 }, { "epoch": 19.110778443113773, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1439, "step": 12766 }, { "epoch": 19.112275449101798, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1346, "step": 12767 }, { "epoch": 19.11377245508982, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1383, "step": 12768 }, { "epoch": 19.115269461077844, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.132, "step": 12769 }, { "epoch": 19.11676646706587, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1365, "step": 12770 }, { "epoch": 19.118263473053894, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1334, "step": 12771 }, { "epoch": 19.119760479041915, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1397, "step": 12772 }, { "epoch": 19.12125748502994, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1396, "step": 12773 }, { "epoch": 19.122754491017965, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12774 }, { "epoch": 19.12425149700599, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1337, "step": 12775 }, { "epoch": 19.12574850299401, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 12776 }, { "epoch": 19.127245508982035, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1387, "step": 12777 }, { "epoch": 19.12874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1334, "step": 12778 }, { "epoch": 19.130239520958085, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.137, "step": 12779 }, { "epoch": 19.131736526946106, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1375, "step": 12780 }, { "epoch": 19.13323353293413, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 12781 }, { "epoch": 19.134730538922156, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.132, "step": 12782 }, { "epoch": 19.13622754491018, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1378, "step": 12783 }, { "epoch": 19.137724550898202, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1291, "step": 12784 }, { "epoch": 19.139221556886227, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1356, "step": 12785 }, { "epoch": 19.14071856287425, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1371, "step": 12786 }, { "epoch": 19.142215568862277, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1386, "step": 12787 }, { "epoch": 19.143712574850298, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1375, "step": 12788 }, { "epoch": 19.145209580838323, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1373, "step": 12789 }, { "epoch": 19.146706586826348, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1349, "step": 12790 }, { "epoch": 19.148203592814372, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1402, "step": 12791 }, { "epoch": 19.149700598802394, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12792 }, { "epoch": 19.15119760479042, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 12793 }, { "epoch": 19.152694610778443, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1286, "step": 12794 }, { "epoch": 19.154191616766468, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1401, "step": 12795 }, { "epoch": 19.15568862275449, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1355, "step": 12796 }, { "epoch": 19.157185628742514, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1358, "step": 12797 }, { "epoch": 19.15868263473054, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1445, "step": 12798 }, { "epoch": 19.160179640718564, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1429, "step": 12799 }, { "epoch": 19.161676646706585, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1361, "step": 12800 }, { "epoch": 19.16317365269461, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1369, "step": 12801 }, { "epoch": 19.164670658682635, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.127, "step": 12802 }, { "epoch": 19.16616766467066, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1312, "step": 12803 }, { "epoch": 19.16766467065868, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1376, "step": 12804 }, { "epoch": 19.169161676646706, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1325, "step": 12805 }, { "epoch": 19.17065868263473, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12806 }, { "epoch": 19.172155688622755, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.136, "step": 12807 }, { "epoch": 19.17365269461078, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1336, "step": 12808 }, { "epoch": 19.1751497005988, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1363, "step": 12809 }, { "epoch": 19.176646706586826, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1401, "step": 12810 }, { "epoch": 19.17814371257485, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1329, "step": 12811 }, { "epoch": 19.179640718562876, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 12812 }, { "epoch": 19.181137724550897, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1409, "step": 12813 }, { "epoch": 19.182634730538922, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1431, "step": 12814 }, { "epoch": 19.184131736526947, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1379, "step": 12815 }, { "epoch": 19.18562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1342, "step": 12816 }, { "epoch": 19.187125748502993, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 12817 }, { "epoch": 19.188622754491018, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1396, "step": 12818 }, { "epoch": 19.190119760479043, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1389, "step": 12819 }, { "epoch": 19.191616766467067, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1407, "step": 12820 }, { "epoch": 19.19311377245509, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1374, "step": 12821 }, { "epoch": 19.194610778443113, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1308, "step": 12822 }, { "epoch": 19.19610778443114, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 12823 }, { "epoch": 19.197604790419163, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1387, "step": 12824 }, { "epoch": 19.199101796407184, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1389, "step": 12825 }, { "epoch": 19.20059880239521, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1452, "step": 12826 }, { "epoch": 19.202095808383234, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12827 }, { "epoch": 19.20359281437126, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 12828 }, { "epoch": 19.20508982035928, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1352, "step": 12829 }, { "epoch": 19.206586826347305, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1414, "step": 12830 }, { "epoch": 19.20808383233533, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1445, "step": 12831 }, { "epoch": 19.209580838323355, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1319, "step": 12832 }, { "epoch": 19.211077844311376, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1394, "step": 12833 }, { "epoch": 19.2125748502994, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1439, "step": 12834 }, { "epoch": 19.214071856287426, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1367, "step": 12835 }, { "epoch": 19.21556886227545, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1304, "step": 12836 }, { "epoch": 19.21706586826347, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1363, "step": 12837 }, { "epoch": 19.218562874251496, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1401, "step": 12838 }, { "epoch": 19.22005988023952, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1371, "step": 12839 }, { "epoch": 19.221556886227546, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1364, "step": 12840 }, { "epoch": 19.223053892215567, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 12841 }, { "epoch": 19.224550898203592, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 12842 }, { "epoch": 19.226047904191617, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1392, "step": 12843 }, { "epoch": 19.227544910179642, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1373, "step": 12844 }, { "epoch": 19.229041916167663, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1289, "step": 12845 }, { "epoch": 19.230538922155688, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 12846 }, { "epoch": 19.232035928143713, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 12847 }, { "epoch": 19.233532934131738, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1357, "step": 12848 }, { "epoch": 19.23502994011976, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1346, "step": 12849 }, { "epoch": 19.236526946107784, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 12850 }, { "epoch": 19.23802395209581, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12851 }, { "epoch": 19.239520958083833, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1332, "step": 12852 }, { "epoch": 19.241017964071855, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.135, "step": 12853 }, { "epoch": 19.24251497005988, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1342, "step": 12854 }, { "epoch": 19.244011976047904, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1382, "step": 12855 }, { "epoch": 19.24550898203593, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1384, "step": 12856 }, { "epoch": 19.24700598802395, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1262, "step": 12857 }, { "epoch": 19.248502994011975, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 12858 }, { "epoch": 19.25, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1383, "step": 12859 }, { "epoch": 19.251497005988025, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1367, "step": 12860 }, { "epoch": 19.25299401197605, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1422, "step": 12861 }, { "epoch": 19.25449101796407, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1341, "step": 12862 }, { "epoch": 19.255988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1373, "step": 12863 }, { "epoch": 19.25748502994012, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12864 }, { "epoch": 19.258982035928145, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1335, "step": 12865 }, { "epoch": 19.260479041916167, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.134, "step": 12866 }, { "epoch": 19.26197604790419, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1322, "step": 12867 }, { "epoch": 19.263473053892216, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1383, "step": 12868 }, { "epoch": 19.26497005988024, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1388, "step": 12869 }, { "epoch": 19.266467065868262, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1364, "step": 12870 }, { "epoch": 19.267964071856287, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.145, "step": 12871 }, { "epoch": 19.269461077844312, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1376, "step": 12872 }, { "epoch": 19.270958083832337, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.129, "step": 12873 }, { "epoch": 19.272455089820358, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1341, "step": 12874 }, { "epoch": 19.273952095808383, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1406, "step": 12875 }, { "epoch": 19.275449101796408, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1277, "step": 12876 }, { "epoch": 19.276946107784433, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 12877 }, { "epoch": 19.278443113772454, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1434, "step": 12878 }, { "epoch": 19.27994011976048, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1386, "step": 12879 }, { "epoch": 19.281437125748504, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 12880 }, { "epoch": 19.28293413173653, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1361, "step": 12881 }, { "epoch": 19.28443113772455, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1371, "step": 12882 }, { "epoch": 19.285928143712574, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1359, "step": 12883 }, { "epoch": 19.2874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1382, "step": 12884 }, { "epoch": 19.288922155688624, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1421, "step": 12885 }, { "epoch": 19.290419161676645, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1377, "step": 12886 }, { "epoch": 19.29191616766467, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1357, "step": 12887 }, { "epoch": 19.293413173652695, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1362, "step": 12888 }, { "epoch": 19.29491017964072, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1331, "step": 12889 }, { "epoch": 19.29640718562874, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1463, "step": 12890 }, { "epoch": 19.297904191616766, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1434, "step": 12891 }, { "epoch": 19.29940119760479, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1402, "step": 12892 }, { "epoch": 19.300898203592816, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1392, "step": 12893 }, { "epoch": 19.302395209580837, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1415, "step": 12894 }, { "epoch": 19.30389221556886, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1362, "step": 12895 }, { "epoch": 19.305389221556887, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.141, "step": 12896 }, { "epoch": 19.30688622754491, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.134, "step": 12897 }, { "epoch": 19.308383233532933, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1454, "step": 12898 }, { "epoch": 19.309880239520957, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.134, "step": 12899 }, { "epoch": 19.311377245508982, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1376, "step": 12900 }, { "epoch": 19.312874251497007, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1446, "step": 12901 }, { "epoch": 19.31437125748503, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1372, "step": 12902 }, { "epoch": 19.315868263473053, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1331, "step": 12903 }, { "epoch": 19.317365269461078, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 12904 }, { "epoch": 19.318862275449103, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1404, "step": 12905 }, { "epoch": 19.320359281437124, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1394, "step": 12906 }, { "epoch": 19.32185628742515, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12907 }, { "epoch": 19.323353293413174, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1411, "step": 12908 }, { "epoch": 19.3248502994012, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.136, "step": 12909 }, { "epoch": 19.32634730538922, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1378, "step": 12910 }, { "epoch": 19.327844311377245, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.133, "step": 12911 }, { "epoch": 19.32934131736527, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12912 }, { "epoch": 19.330838323353294, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1266, "step": 12913 }, { "epoch": 19.33233532934132, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1405, "step": 12914 }, { "epoch": 19.33383233532934, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1359, "step": 12915 }, { "epoch": 19.335329341317365, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1397, "step": 12916 }, { "epoch": 19.33682634730539, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 12917 }, { "epoch": 19.338323353293415, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1312, "step": 12918 }, { "epoch": 19.339820359281436, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1379, "step": 12919 }, { "epoch": 19.34131736526946, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.137, "step": 12920 }, { "epoch": 19.342814371257486, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 12921 }, { "epoch": 19.34431137724551, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1387, "step": 12922 }, { "epoch": 19.345808383233532, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1388, "step": 12923 }, { "epoch": 19.347305389221557, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1377, "step": 12924 }, { "epoch": 19.34880239520958, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1404, "step": 12925 }, { "epoch": 19.350299401197606, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1408, "step": 12926 }, { "epoch": 19.351796407185628, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1364, "step": 12927 }, { "epoch": 19.353293413173652, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.141, "step": 12928 }, { "epoch": 19.354790419161677, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1349, "step": 12929 }, { "epoch": 19.356287425149702, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1401, "step": 12930 }, { "epoch": 19.357784431137723, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1373, "step": 12931 }, { "epoch": 19.35928143712575, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1389, "step": 12932 }, { "epoch": 19.360778443113773, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1371, "step": 12933 }, { "epoch": 19.362275449101798, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1369, "step": 12934 }, { "epoch": 19.36377245508982, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1368, "step": 12935 }, { "epoch": 19.365269461077844, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1336, "step": 12936 }, { "epoch": 19.36676646706587, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1336, "step": 12937 }, { "epoch": 19.368263473053894, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 12938 }, { "epoch": 19.369760479041915, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12939 }, { "epoch": 19.37125748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12940 }, { "epoch": 19.372754491017965, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.139, "step": 12941 }, { "epoch": 19.37425149700599, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1417, "step": 12942 }, { "epoch": 19.37574850299401, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1356, "step": 12943 }, { "epoch": 19.377245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.135, "step": 12944 }, { "epoch": 19.37874251497006, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1343, "step": 12945 }, { "epoch": 19.380239520958085, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1367, "step": 12946 }, { "epoch": 19.381736526946106, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.141, "step": 12947 }, { "epoch": 19.38323353293413, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1461, "step": 12948 }, { "epoch": 19.384730538922156, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 12949 }, { "epoch": 19.38622754491018, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1311, "step": 12950 }, { "epoch": 19.387724550898202, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1437, "step": 12951 }, { "epoch": 19.389221556886227, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1384, "step": 12952 }, { "epoch": 19.39071856287425, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1408, "step": 12953 }, { "epoch": 19.392215568862277, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1393, "step": 12954 }, { "epoch": 19.393712574850298, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1405, "step": 12955 }, { "epoch": 19.395209580838323, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1389, "step": 12956 }, { "epoch": 19.396706586826348, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.134, "step": 12957 }, { "epoch": 19.398203592814372, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1331, "step": 12958 }, { "epoch": 19.399700598802394, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1368, "step": 12959 }, { "epoch": 19.40119760479042, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1328, "step": 12960 }, { "epoch": 19.402694610778443, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1458, "step": 12961 }, { "epoch": 19.404191616766468, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1418, "step": 12962 }, { "epoch": 19.40568862275449, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1361, "step": 12963 }, { "epoch": 19.407185628742514, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1343, "step": 12964 }, { "epoch": 19.40868263473054, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1425, "step": 12965 }, { "epoch": 19.410179640718564, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1323, "step": 12966 }, { "epoch": 19.411676646706585, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1365, "step": 12967 }, { "epoch": 19.41317365269461, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1372, "step": 12968 }, { "epoch": 19.414670658682635, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1469, "step": 12969 }, { "epoch": 19.41616766467066, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12970 }, { "epoch": 19.41766467065868, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 12971 }, { "epoch": 19.419161676646706, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.137, "step": 12972 }, { "epoch": 19.42065868263473, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1302, "step": 12973 }, { "epoch": 19.422155688622755, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1397, "step": 12974 }, { "epoch": 19.42365269461078, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 12975 }, { "epoch": 19.4251497005988, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1404, "step": 12976 }, { "epoch": 19.426646706586826, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1344, "step": 12977 }, { "epoch": 19.42814371257485, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1354, "step": 12978 }, { "epoch": 19.429640718562876, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1338, "step": 12979 }, { "epoch": 19.431137724550897, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 12980 }, { "epoch": 19.432634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1393, "step": 12981 }, { "epoch": 19.434131736526947, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1395, "step": 12982 }, { "epoch": 19.43562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1431, "step": 12983 }, { "epoch": 19.437125748502993, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.14, "step": 12984 }, { "epoch": 19.438622754491018, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1405, "step": 12985 }, { "epoch": 19.440119760479043, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1347, "step": 12986 }, { "epoch": 19.441616766467067, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 12987 }, { "epoch": 19.44311377245509, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 12988 }, { "epoch": 19.444610778443113, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1392, "step": 12989 }, { "epoch": 19.44610778443114, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1336, "step": 12990 }, { "epoch": 19.447604790419163, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.131, "step": 12991 }, { "epoch": 19.449101796407184, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1363, "step": 12992 }, { "epoch": 19.45059880239521, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12993 }, { "epoch": 19.452095808383234, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12994 }, { "epoch": 19.45359281437126, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1437, "step": 12995 }, { "epoch": 19.45508982035928, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1347, "step": 12996 }, { "epoch": 19.456586826347305, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1349, "step": 12997 }, { "epoch": 19.45808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1314, "step": 12998 }, { "epoch": 19.459580838323355, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1362, "step": 12999 }, { "epoch": 19.461077844311376, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13000 }, { "epoch": 19.4625748502994, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1384, "step": 13001 }, { "epoch": 19.464071856287426, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13002 }, { "epoch": 19.46556886227545, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 13003 }, { "epoch": 19.46706586826347, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1371, "step": 13004 }, { "epoch": 19.468562874251496, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1314, "step": 13005 }, { "epoch": 19.47005988023952, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1388, "step": 13006 }, { "epoch": 19.471556886227546, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1409, "step": 13007 }, { "epoch": 19.473053892215567, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1328, "step": 13008 }, { "epoch": 19.474550898203592, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1416, "step": 13009 }, { "epoch": 19.476047904191617, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1387, "step": 13010 }, { "epoch": 19.477544910179642, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1391, "step": 13011 }, { "epoch": 19.479041916167663, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1372, "step": 13012 }, { "epoch": 19.480538922155688, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1303, "step": 13013 }, { "epoch": 19.482035928143713, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1347, "step": 13014 }, { "epoch": 19.483532934131738, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1417, "step": 13015 }, { "epoch": 19.48502994011976, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1365, "step": 13016 }, { "epoch": 19.486526946107784, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1371, "step": 13017 }, { "epoch": 19.48802395209581, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.135, "step": 13018 }, { "epoch": 19.489520958083833, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 13019 }, { "epoch": 19.491017964071855, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1293, "step": 13020 }, { "epoch": 19.49251497005988, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1392, "step": 13021 }, { "epoch": 19.494011976047904, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1396, "step": 13022 }, { "epoch": 19.49550898203593, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1334, "step": 13023 }, { "epoch": 19.49700598802395, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1369, "step": 13024 }, { "epoch": 19.498502994011975, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1428, "step": 13025 }, { "epoch": 19.5, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1361, "step": 13026 }, { "epoch": 19.501497005988025, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1398, "step": 13027 }, { "epoch": 19.50299401197605, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1323, "step": 13028 }, { "epoch": 19.50449101796407, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1376, "step": 13029 }, { "epoch": 19.505988023952096, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.14, "step": 13030 }, { "epoch": 19.50748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1315, "step": 13031 }, { "epoch": 19.508982035928145, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1286, "step": 13032 }, { "epoch": 19.510479041916167, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1345, "step": 13033 }, { "epoch": 19.51197604790419, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1339, "step": 13034 }, { "epoch": 19.513473053892216, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1382, "step": 13035 }, { "epoch": 19.51497005988024, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.142, "step": 13036 }, { "epoch": 19.516467065868262, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1397, "step": 13037 }, { "epoch": 19.517964071856287, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1405, "step": 13038 }, { "epoch": 19.519461077844312, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1409, "step": 13039 }, { "epoch": 19.520958083832337, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1329, "step": 13040 }, { "epoch": 19.522455089820358, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13041 }, { "epoch": 19.523952095808383, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1364, "step": 13042 }, { "epoch": 19.525449101796408, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1451, "step": 13043 }, { "epoch": 19.526946107784433, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1428, "step": 13044 }, { "epoch": 19.528443113772454, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1289, "step": 13045 }, { "epoch": 19.52994011976048, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1357, "step": 13046 }, { "epoch": 19.531437125748504, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1376, "step": 13047 }, { "epoch": 19.53293413173653, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1375, "step": 13048 }, { "epoch": 19.53443113772455, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1363, "step": 13049 }, { "epoch": 19.535928143712574, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1339, "step": 13050 }, { "epoch": 19.5374251497006, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1401, "step": 13051 }, { "epoch": 19.538922155688624, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 13052 }, { "epoch": 19.540419161676645, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.142, "step": 13053 }, { "epoch": 19.54191616766467, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.135, "step": 13054 }, { "epoch": 19.543413173652695, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1332, "step": 13055 }, { "epoch": 19.54491017964072, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1392, "step": 13056 }, { "epoch": 19.54640718562874, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.144, "step": 13057 }, { "epoch": 19.547904191616766, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.131, "step": 13058 }, { "epoch": 19.54940119760479, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13059 }, { "epoch": 19.550898203592816, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1378, "step": 13060 }, { "epoch": 19.552395209580837, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1435, "step": 13061 }, { "epoch": 19.55389221556886, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1412, "step": 13062 }, { "epoch": 19.555389221556887, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.134, "step": 13063 }, { "epoch": 19.55688622754491, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1332, "step": 13064 }, { "epoch": 19.558383233532933, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13065 }, { "epoch": 19.559880239520957, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1399, "step": 13066 }, { "epoch": 19.561377245508982, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 13067 }, { "epoch": 19.562874251497007, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1381, "step": 13068 }, { "epoch": 19.56437125748503, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1338, "step": 13069 }, { "epoch": 19.565868263473053, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.138, "step": 13070 }, { "epoch": 19.567365269461078, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1324, "step": 13071 }, { "epoch": 19.568862275449103, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1394, "step": 13072 }, { "epoch": 19.570359281437124, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1372, "step": 13073 }, { "epoch": 19.57185628742515, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1348, "step": 13074 }, { "epoch": 19.573353293413174, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1313, "step": 13075 }, { "epoch": 19.5748502994012, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1395, "step": 13076 }, { "epoch": 19.57634730538922, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1307, "step": 13077 }, { "epoch": 19.577844311377245, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1398, "step": 13078 }, { "epoch": 19.57934131736527, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1406, "step": 13079 }, { "epoch": 19.580838323353294, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1377, "step": 13080 }, { "epoch": 19.58233532934132, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1352, "step": 13081 }, { "epoch": 19.58383233532934, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1322, "step": 13082 }, { "epoch": 19.585329341317365, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1382, "step": 13083 }, { "epoch": 19.58682634730539, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1369, "step": 13084 }, { "epoch": 19.58832335329341, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1363, "step": 13085 }, { "epoch": 19.589820359281436, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13086 }, { "epoch": 19.59131736526946, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1336, "step": 13087 }, { "epoch": 19.592814371257486, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1319, "step": 13088 }, { "epoch": 19.59431137724551, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1388, "step": 13089 }, { "epoch": 19.595808383233532, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1364, "step": 13090 }, { "epoch": 19.597305389221557, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1378, "step": 13091 }, { "epoch": 19.59880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1351, "step": 13092 }, { "epoch": 19.600299401197606, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1323, "step": 13093 }, { "epoch": 19.601796407185628, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.133, "step": 13094 }, { "epoch": 19.603293413173652, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1384, "step": 13095 }, { "epoch": 19.604790419161677, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1414, "step": 13096 }, { "epoch": 19.606287425149702, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 13097 }, { "epoch": 19.607784431137723, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1338, "step": 13098 }, { "epoch": 19.60928143712575, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1389, "step": 13099 }, { "epoch": 19.610778443113773, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1384, "step": 13100 }, { "epoch": 19.612275449101798, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1397, "step": 13101 }, { "epoch": 19.61377245508982, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1385, "step": 13102 }, { "epoch": 19.615269461077844, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 13103 }, { "epoch": 19.61676646706587, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13104 }, { "epoch": 19.618263473053894, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.137, "step": 13105 }, { "epoch": 19.619760479041915, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1338, "step": 13106 }, { "epoch": 19.62125748502994, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.138, "step": 13107 }, { "epoch": 19.622754491017965, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1362, "step": 13108 }, { "epoch": 19.62425149700599, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.14, "step": 13109 }, { "epoch": 19.62574850299401, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1359, "step": 13110 }, { "epoch": 19.627245508982035, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1366, "step": 13111 }, { "epoch": 19.62874251497006, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1337, "step": 13112 }, { "epoch": 19.630239520958085, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 13113 }, { "epoch": 19.631736526946106, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1344, "step": 13114 }, { "epoch": 19.63323353293413, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13115 }, { "epoch": 19.634730538922156, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1382, "step": 13116 }, { "epoch": 19.63622754491018, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13117 }, { "epoch": 19.637724550898202, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1421, "step": 13118 }, { "epoch": 19.639221556886227, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1353, "step": 13119 }, { "epoch": 19.64071856287425, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1397, "step": 13120 }, { "epoch": 19.642215568862277, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1454, "step": 13121 }, { "epoch": 19.643712574850298, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1369, "step": 13122 }, { "epoch": 19.645209580838323, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1436, "step": 13123 }, { "epoch": 19.646706586826348, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1349, "step": 13124 }, { "epoch": 19.648203592814372, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1333, "step": 13125 }, { "epoch": 19.649700598802394, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1368, "step": 13126 }, { "epoch": 19.65119760479042, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1352, "step": 13127 }, { "epoch": 19.652694610778443, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1369, "step": 13128 }, { "epoch": 19.654191616766468, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1363, "step": 13129 }, { "epoch": 19.65568862275449, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1405, "step": 13130 }, { "epoch": 19.657185628742514, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1376, "step": 13131 }, { "epoch": 19.65868263473054, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1422, "step": 13132 }, { "epoch": 19.660179640718564, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.139, "step": 13133 }, { "epoch": 19.66167664670659, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1342, "step": 13134 }, { "epoch": 19.66317365269461, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1422, "step": 13135 }, { "epoch": 19.664670658682635, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1434, "step": 13136 }, { "epoch": 19.66616766467066, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1429, "step": 13137 }, { "epoch": 19.66766467065868, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13138 }, { "epoch": 19.669161676646706, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13139 }, { "epoch": 19.67065868263473, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1332, "step": 13140 }, { "epoch": 19.672155688622755, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1354, "step": 13141 }, { "epoch": 19.67365269461078, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1424, "step": 13142 }, { "epoch": 19.6751497005988, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1389, "step": 13143 }, { "epoch": 19.676646706586826, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 13144 }, { "epoch": 19.67814371257485, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1304, "step": 13145 }, { "epoch": 19.679640718562876, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1484, "step": 13146 }, { "epoch": 19.681137724550897, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1421, "step": 13147 }, { "epoch": 19.682634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1358, "step": 13148 }, { "epoch": 19.684131736526947, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1361, "step": 13149 }, { "epoch": 19.68562874251497, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 13150 }, { "epoch": 19.687125748502993, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 13151 }, { "epoch": 19.688622754491018, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1327, "step": 13152 }, { "epoch": 19.690119760479043, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1439, "step": 13153 }, { "epoch": 19.691616766467067, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1361, "step": 13154 }, { "epoch": 19.69311377245509, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.139, "step": 13155 }, { "epoch": 19.694610778443113, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1366, "step": 13156 }, { "epoch": 19.69610778443114, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1409, "step": 13157 }, { "epoch": 19.697604790419163, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1325, "step": 13158 }, { "epoch": 19.699101796407184, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1409, "step": 13159 }, { "epoch": 19.70059880239521, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1439, "step": 13160 }, { "epoch": 19.702095808383234, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13161 }, { "epoch": 19.70359281437126, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.14, "step": 13162 }, { "epoch": 19.70508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.134, "step": 13163 }, { "epoch": 19.706586826347305, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1345, "step": 13164 }, { "epoch": 19.70808383233533, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1325, "step": 13165 }, { "epoch": 19.709580838323355, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1413, "step": 13166 }, { "epoch": 19.711077844311376, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1456, "step": 13167 }, { "epoch": 19.7125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1321, "step": 13168 }, { "epoch": 19.714071856287426, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1437, "step": 13169 }, { "epoch": 19.71556886227545, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1365, "step": 13170 }, { "epoch": 19.71706586826347, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1412, "step": 13171 }, { "epoch": 19.718562874251496, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.138, "step": 13172 }, { "epoch": 19.72005988023952, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13173 }, { "epoch": 19.721556886227546, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1365, "step": 13174 }, { "epoch": 19.723053892215567, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1375, "step": 13175 }, { "epoch": 19.724550898203592, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1353, "step": 13176 }, { "epoch": 19.726047904191617, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1293, "step": 13177 }, { "epoch": 19.727544910179642, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 13178 }, { "epoch": 19.729041916167663, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1343, "step": 13179 }, { "epoch": 19.730538922155688, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 13180 }, { "epoch": 19.732035928143713, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1388, "step": 13181 }, { "epoch": 19.733532934131738, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1326, "step": 13182 }, { "epoch": 19.73502994011976, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1403, "step": 13183 }, { "epoch": 19.736526946107784, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13184 }, { "epoch": 19.73802395209581, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1411, "step": 13185 }, { "epoch": 19.739520958083833, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13186 }, { "epoch": 19.741017964071855, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1356, "step": 13187 }, { "epoch": 19.74251497005988, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.137, "step": 13188 }, { "epoch": 19.744011976047904, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1366, "step": 13189 }, { "epoch": 19.74550898203593, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.138, "step": 13190 }, { "epoch": 19.74700598802395, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1361, "step": 13191 }, { "epoch": 19.748502994011975, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1332, "step": 13192 }, { "epoch": 19.75, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1381, "step": 13193 }, { "epoch": 19.751497005988025, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1345, "step": 13194 }, { "epoch": 19.75299401197605, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1383, "step": 13195 }, { "epoch": 19.75449101796407, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1385, "step": 13196 }, { "epoch": 19.755988023952096, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1381, "step": 13197 }, { "epoch": 19.75748502994012, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1398, "step": 13198 }, { "epoch": 19.758982035928145, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1463, "step": 13199 }, { "epoch": 19.760479041916167, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1422, "step": 13200 }, { "epoch": 19.76197604790419, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1364, "step": 13201 }, { "epoch": 19.763473053892216, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1395, "step": 13202 }, { "epoch": 19.76497005988024, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 13203 }, { "epoch": 19.766467065868262, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.137, "step": 13204 }, { "epoch": 19.767964071856287, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1366, "step": 13205 }, { "epoch": 19.769461077844312, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.149, "step": 13206 }, { "epoch": 19.770958083832337, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.135, "step": 13207 }, { "epoch": 19.772455089820358, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1329, "step": 13208 }, { "epoch": 19.773952095808383, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1394, "step": 13209 }, { "epoch": 19.775449101796408, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.139, "step": 13210 }, { "epoch": 19.776946107784433, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1387, "step": 13211 }, { "epoch": 19.778443113772454, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.138, "step": 13212 }, { "epoch": 19.77994011976048, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 13213 }, { "epoch": 19.781437125748504, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1308, "step": 13214 }, { "epoch": 19.78293413173653, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1376, "step": 13215 }, { "epoch": 19.78443113772455, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1304, "step": 13216 }, { "epoch": 19.785928143712574, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1373, "step": 13217 }, { "epoch": 19.7874251497006, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1445, "step": 13218 }, { "epoch": 19.788922155688624, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1369, "step": 13219 }, { "epoch": 19.790419161676645, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1343, "step": 13220 }, { "epoch": 19.79191616766467, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1307, "step": 13221 }, { "epoch": 19.793413173652695, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.136, "step": 13222 }, { "epoch": 19.79491017964072, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.136, "step": 13223 }, { "epoch": 19.79640718562874, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1431, "step": 13224 }, { "epoch": 19.797904191616766, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1393, "step": 13225 }, { "epoch": 19.79940119760479, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1386, "step": 13226 }, { "epoch": 19.800898203592816, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.141, "step": 13227 }, { "epoch": 19.802395209580837, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 13228 }, { "epoch": 19.80389221556886, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1372, "step": 13229 }, { "epoch": 19.805389221556887, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1375, "step": 13230 }, { "epoch": 19.80688622754491, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1396, "step": 13231 }, { "epoch": 19.808383233532933, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1336, "step": 13232 }, { "epoch": 19.809880239520957, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13233 }, { "epoch": 19.811377245508982, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1389, "step": 13234 }, { "epoch": 19.812874251497007, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1372, "step": 13235 }, { "epoch": 19.81437125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1342, "step": 13236 }, { "epoch": 19.815868263473053, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1375, "step": 13237 }, { "epoch": 19.817365269461078, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1339, "step": 13238 }, { "epoch": 19.818862275449103, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1433, "step": 13239 }, { "epoch": 19.820359281437124, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1327, "step": 13240 }, { "epoch": 19.82185628742515, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1367, "step": 13241 }, { "epoch": 19.823353293413174, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1424, "step": 13242 }, { "epoch": 19.8248502994012, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1407, "step": 13243 }, { "epoch": 19.82634730538922, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1347, "step": 13244 }, { "epoch": 19.827844311377245, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 13245 }, { "epoch": 19.82934131736527, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1415, "step": 13246 }, { "epoch": 19.830838323353294, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1388, "step": 13247 }, { "epoch": 19.83233532934132, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1412, "step": 13248 }, { "epoch": 19.83383233532934, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13249 }, { "epoch": 19.835329341317365, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1319, "step": 13250 }, { "epoch": 19.83682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 13251 }, { "epoch": 19.83832335329341, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.149, "step": 13252 }, { "epoch": 19.839820359281436, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13253 }, { "epoch": 19.84131736526946, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 13254 }, { "epoch": 19.842814371257486, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 13255 }, { "epoch": 19.84431137724551, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1323, "step": 13256 }, { "epoch": 19.845808383233532, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1396, "step": 13257 }, { "epoch": 19.847305389221557, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1384, "step": 13258 }, { "epoch": 19.84880239520958, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1364, "step": 13259 }, { "epoch": 19.850299401197606, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1384, "step": 13260 }, { "epoch": 19.851796407185628, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.15, "step": 13261 }, { "epoch": 19.853293413173652, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.129, "step": 13262 }, { "epoch": 19.854790419161677, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1386, "step": 13263 }, { "epoch": 19.856287425149702, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1361, "step": 13264 }, { "epoch": 19.857784431137723, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1395, "step": 13265 }, { "epoch": 19.85928143712575, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1353, "step": 13266 }, { "epoch": 19.860778443113773, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1348, "step": 13267 }, { "epoch": 19.862275449101798, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1404, "step": 13268 }, { "epoch": 19.86377245508982, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1345, "step": 13269 }, { "epoch": 19.865269461077844, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1369, "step": 13270 }, { "epoch": 19.86676646706587, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1365, "step": 13271 }, { "epoch": 19.868263473053894, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1321, "step": 13272 }, { "epoch": 19.869760479041915, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1441, "step": 13273 }, { "epoch": 19.87125748502994, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1387, "step": 13274 }, { "epoch": 19.872754491017965, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 13275 }, { "epoch": 19.87425149700599, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13276 }, { "epoch": 19.87574850299401, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1424, "step": 13277 }, { "epoch": 19.877245508982035, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1375, "step": 13278 }, { "epoch": 19.87874251497006, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1375, "step": 13279 }, { "epoch": 19.880239520958085, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1321, "step": 13280 }, { "epoch": 19.881736526946106, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1432, "step": 13281 }, { "epoch": 19.88323353293413, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13282 }, { "epoch": 19.884730538922156, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1336, "step": 13283 }, { "epoch": 19.88622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 13284 }, { "epoch": 19.887724550898202, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.137, "step": 13285 }, { "epoch": 19.889221556886227, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1368, "step": 13286 }, { "epoch": 19.89071856287425, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1391, "step": 13287 }, { "epoch": 19.892215568862277, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13288 }, { "epoch": 19.893712574850298, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 13289 }, { "epoch": 19.895209580838323, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1357, "step": 13290 }, { "epoch": 19.896706586826348, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1327, "step": 13291 }, { "epoch": 19.898203592814372, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.128, "step": 13292 }, { "epoch": 19.899700598802394, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1305, "step": 13293 }, { "epoch": 19.90119760479042, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1411, "step": 13294 }, { "epoch": 19.902694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1384, "step": 13295 }, { "epoch": 19.904191616766468, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1366, "step": 13296 }, { "epoch": 19.90568862275449, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 13297 }, { "epoch": 19.907185628742514, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1377, "step": 13298 }, { "epoch": 19.90868263473054, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1416, "step": 13299 }, { "epoch": 19.910179640718564, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13300 }, { "epoch": 19.91167664670659, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1344, "step": 13301 }, { "epoch": 19.91317365269461, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1384, "step": 13302 }, { "epoch": 19.914670658682635, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1296, "step": 13303 }, { "epoch": 19.91616766467066, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1393, "step": 13304 }, { "epoch": 19.91766467065868, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1474, "step": 13305 }, { "epoch": 19.919161676646706, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1326, "step": 13306 }, { "epoch": 19.92065868263473, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.137, "step": 13307 }, { "epoch": 19.922155688622755, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1404, "step": 13308 }, { "epoch": 19.92365269461078, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.132, "step": 13309 }, { "epoch": 19.9251497005988, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1394, "step": 13310 }, { "epoch": 19.926646706586826, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1398, "step": 13311 }, { "epoch": 19.92814371257485, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1283, "step": 13312 }, { "epoch": 19.929640718562876, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 13313 }, { "epoch": 19.931137724550897, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1339, "step": 13314 }, { "epoch": 19.932634730538922, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.137, "step": 13315 }, { "epoch": 19.934131736526947, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1347, "step": 13316 }, { "epoch": 19.93562874251497, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1388, "step": 13317 }, { "epoch": 19.937125748502993, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1367, "step": 13318 }, { "epoch": 19.938622754491018, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1436, "step": 13319 }, { "epoch": 19.940119760479043, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1293, "step": 13320 }, { "epoch": 19.941616766467067, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1377, "step": 13321 }, { "epoch": 19.94311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.133, "step": 13322 }, { "epoch": 19.944610778443113, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1379, "step": 13323 }, { "epoch": 19.94610778443114, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1369, "step": 13324 }, { "epoch": 19.947604790419163, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1407, "step": 13325 }, { "epoch": 19.949101796407184, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1352, "step": 13326 }, { "epoch": 19.95059880239521, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13327 }, { "epoch": 19.952095808383234, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1325, "step": 13328 }, { "epoch": 19.95359281437126, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.138, "step": 13329 }, { "epoch": 19.95508982035928, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 13330 }, { "epoch": 19.956586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1379, "step": 13331 }, { "epoch": 19.95808383233533, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1393, "step": 13332 }, { "epoch": 19.959580838323355, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.139, "step": 13333 }, { "epoch": 19.961077844311376, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1379, "step": 13334 }, { "epoch": 19.9625748502994, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1342, "step": 13335 }, { "epoch": 19.964071856287426, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1349, "step": 13336 }, { "epoch": 19.96556886227545, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 13337 }, { "epoch": 19.96706586826347, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.136, "step": 13338 }, { "epoch": 19.968562874251496, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1323, "step": 13339 }, { "epoch": 19.97005988023952, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1365, "step": 13340 }, { "epoch": 19.971556886227546, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1409, "step": 13341 }, { "epoch": 19.973053892215567, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1401, "step": 13342 }, { "epoch": 19.974550898203592, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1369, "step": 13343 }, { "epoch": 19.976047904191617, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1358, "step": 13344 }, { "epoch": 19.977544910179642, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1387, "step": 13345 }, { "epoch": 19.979041916167663, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1391, "step": 13346 }, { "epoch": 19.980538922155688, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1385, "step": 13347 }, { "epoch": 19.982035928143713, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1384, "step": 13348 }, { "epoch": 19.983532934131738, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1372, "step": 13349 }, { "epoch": 19.98502994011976, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1265, "step": 13350 }, { "epoch": 19.986526946107784, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1384, "step": 13351 }, { "epoch": 19.98802395209581, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13352 }, { "epoch": 19.989520958083833, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.137, "step": 13353 }, { "epoch": 19.991017964071855, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 13354 }, { "epoch": 19.99251497005988, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1375, "step": 13355 }, { "epoch": 19.994011976047904, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1342, "step": 13356 }, { "epoch": 19.99550898203593, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1384, "step": 13357 }, { "epoch": 19.99700598802395, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.142, "step": 13358 }, { "epoch": 19.998502994011975, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1367, "step": 13359 }, { "epoch": 20.0, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1418, "step": 13360 }, { "epoch": 20.001497005988025, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1359, "step": 13361 }, { "epoch": 20.00299401197605, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13362 }, { "epoch": 20.00449101796407, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1342, "step": 13363 }, { "epoch": 20.005988023952096, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1338, "step": 13364 }, { "epoch": 20.00748502994012, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1325, "step": 13365 }, { "epoch": 20.008982035928145, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1362, "step": 13366 }, { "epoch": 20.010479041916167, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1339, "step": 13367 }, { "epoch": 20.01197604790419, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1364, "step": 13368 }, { "epoch": 20.013473053892216, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1396, "step": 13369 }, { "epoch": 20.01497005988024, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1278, "step": 13370 }, { "epoch": 20.016467065868262, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1351, "step": 13371 }, { "epoch": 20.017964071856287, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1417, "step": 13372 }, { "epoch": 20.019461077844312, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1354, "step": 13373 }, { "epoch": 20.020958083832337, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1314, "step": 13374 }, { "epoch": 20.022455089820358, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1337, "step": 13375 }, { "epoch": 20.023952095808383, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1379, "step": 13376 }, { "epoch": 20.025449101796408, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1341, "step": 13377 }, { "epoch": 20.026946107784433, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1353, "step": 13378 }, { "epoch": 20.028443113772454, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1315, "step": 13379 }, { "epoch": 20.02994011976048, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1369, "step": 13380 }, { "epoch": 20.031437125748504, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.14, "step": 13381 }, { "epoch": 20.03293413173653, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.131, "step": 13382 }, { "epoch": 20.03443113772455, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1349, "step": 13383 }, { "epoch": 20.035928143712574, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1341, "step": 13384 }, { "epoch": 20.0374251497006, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1401, "step": 13385 }, { "epoch": 20.038922155688624, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1409, "step": 13386 }, { "epoch": 20.040419161676645, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1353, "step": 13387 }, { "epoch": 20.04191616766467, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1313, "step": 13388 }, { "epoch": 20.043413173652695, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1376, "step": 13389 }, { "epoch": 20.04491017964072, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1474, "step": 13390 }, { "epoch": 20.04640718562874, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1397, "step": 13391 }, { "epoch": 20.047904191616766, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1384, "step": 13392 }, { "epoch": 20.04940119760479, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1407, "step": 13393 }, { "epoch": 20.050898203592816, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1336, "step": 13394 }, { "epoch": 20.052395209580837, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1431, "step": 13395 }, { "epoch": 20.05389221556886, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1386, "step": 13396 }, { "epoch": 20.055389221556887, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1329, "step": 13397 }, { "epoch": 20.05688622754491, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1415, "step": 13398 }, { "epoch": 20.058383233532933, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1342, "step": 13399 }, { "epoch": 20.059880239520957, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1306, "step": 13400 }, { "epoch": 20.061377245508982, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 13401 }, { "epoch": 20.062874251497007, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 13402 }, { "epoch": 20.06437125748503, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1321, "step": 13403 }, { "epoch": 20.065868263473053, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.134, "step": 13404 }, { "epoch": 20.067365269461078, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1347, "step": 13405 }, { "epoch": 20.068862275449103, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.136, "step": 13406 }, { "epoch": 20.070359281437124, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1406, "step": 13407 }, { "epoch": 20.07185628742515, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.144, "step": 13408 }, { "epoch": 20.073353293413174, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1288, "step": 13409 }, { "epoch": 20.0748502994012, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1472, "step": 13410 }, { "epoch": 20.07634730538922, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1329, "step": 13411 }, { "epoch": 20.077844311377245, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1337, "step": 13412 }, { "epoch": 20.07934131736527, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1386, "step": 13413 }, { "epoch": 20.080838323353294, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1366, "step": 13414 }, { "epoch": 20.082335329341316, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1325, "step": 13415 }, { "epoch": 20.08383233532934, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1373, "step": 13416 }, { "epoch": 20.085329341317365, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1358, "step": 13417 }, { "epoch": 20.08682634730539, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1467, "step": 13418 }, { "epoch": 20.088323353293415, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1367, "step": 13419 }, { "epoch": 20.089820359281436, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1331, "step": 13420 }, { "epoch": 20.09131736526946, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1378, "step": 13421 }, { "epoch": 20.092814371257486, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.133, "step": 13422 }, { "epoch": 20.09431137724551, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1353, "step": 13423 }, { "epoch": 20.095808383233532, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13424 }, { "epoch": 20.097305389221557, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1389, "step": 13425 }, { "epoch": 20.09880239520958, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1284, "step": 13426 }, { "epoch": 20.100299401197606, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1406, "step": 13427 }, { "epoch": 20.101796407185628, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1389, "step": 13428 }, { "epoch": 20.103293413173652, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1352, "step": 13429 }, { "epoch": 20.104790419161677, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 13430 }, { "epoch": 20.106287425149702, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1361, "step": 13431 }, { "epoch": 20.107784431137723, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1414, "step": 13432 }, { "epoch": 20.10928143712575, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1338, "step": 13433 }, { "epoch": 20.110778443113773, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1415, "step": 13434 }, { "epoch": 20.112275449101798, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1456, "step": 13435 }, { "epoch": 20.11377245508982, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1323, "step": 13436 }, { "epoch": 20.115269461077844, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1391, "step": 13437 }, { "epoch": 20.11676646706587, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.139, "step": 13438 }, { "epoch": 20.118263473053894, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1387, "step": 13439 }, { "epoch": 20.119760479041915, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1347, "step": 13440 }, { "epoch": 20.12125748502994, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1324, "step": 13441 }, { "epoch": 20.122754491017965, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1327, "step": 13442 }, { "epoch": 20.12425149700599, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1313, "step": 13443 }, { "epoch": 20.12574850299401, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1349, "step": 13444 }, { "epoch": 20.127245508982035, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1367, "step": 13445 }, { "epoch": 20.12874251497006, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1276, "step": 13446 }, { "epoch": 20.130239520958085, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.133, "step": 13447 }, { "epoch": 20.131736526946106, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.142, "step": 13448 }, { "epoch": 20.13323353293413, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1377, "step": 13449 }, { "epoch": 20.134730538922156, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1342, "step": 13450 }, { "epoch": 20.13622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1383, "step": 13451 }, { "epoch": 20.137724550898202, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 13452 }, { "epoch": 20.139221556886227, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13453 }, { "epoch": 20.14071856287425, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.138, "step": 13454 }, { "epoch": 20.142215568862277, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13455 }, { "epoch": 20.143712574850298, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1364, "step": 13456 }, { "epoch": 20.145209580838323, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1404, "step": 13457 }, { "epoch": 20.146706586826348, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.135, "step": 13458 }, { "epoch": 20.148203592814372, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 13459 }, { "epoch": 20.149700598802394, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1407, "step": 13460 }, { "epoch": 20.15119760479042, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 13461 }, { "epoch": 20.152694610778443, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1369, "step": 13462 }, { "epoch": 20.154191616766468, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1357, "step": 13463 }, { "epoch": 20.15568862275449, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1411, "step": 13464 }, { "epoch": 20.157185628742514, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13465 }, { "epoch": 20.15868263473054, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1323, "step": 13466 }, { "epoch": 20.160179640718564, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1331, "step": 13467 }, { "epoch": 20.161676646706585, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 13468 }, { "epoch": 20.16317365269461, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1406, "step": 13469 }, { "epoch": 20.164670658682635, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1359, "step": 13470 }, { "epoch": 20.16616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13471 }, { "epoch": 20.16766467065868, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1319, "step": 13472 }, { "epoch": 20.169161676646706, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13473 }, { "epoch": 20.17065868263473, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1386, "step": 13474 }, { "epoch": 20.172155688622755, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1367, "step": 13475 }, { "epoch": 20.17365269461078, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 13476 }, { "epoch": 20.1751497005988, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1301, "step": 13477 }, { "epoch": 20.176646706586826, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1391, "step": 13478 }, { "epoch": 20.17814371257485, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1299, "step": 13479 }, { "epoch": 20.179640718562876, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1387, "step": 13480 }, { "epoch": 20.181137724550897, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 13481 }, { "epoch": 20.182634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13482 }, { "epoch": 20.184131736526947, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1389, "step": 13483 }, { "epoch": 20.18562874251497, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1387, "step": 13484 }, { "epoch": 20.187125748502993, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1333, "step": 13485 }, { "epoch": 20.188622754491018, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1365, "step": 13486 }, { "epoch": 20.190119760479043, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1403, "step": 13487 }, { "epoch": 20.191616766467067, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1386, "step": 13488 }, { "epoch": 20.19311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1469, "step": 13489 }, { "epoch": 20.194610778443113, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13490 }, { "epoch": 20.19610778443114, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1381, "step": 13491 }, { "epoch": 20.197604790419163, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.137, "step": 13492 }, { "epoch": 20.199101796407184, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1362, "step": 13493 }, { "epoch": 20.20059880239521, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13494 }, { "epoch": 20.202095808383234, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1335, "step": 13495 }, { "epoch": 20.20359281437126, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1348, "step": 13496 }, { "epoch": 20.20508982035928, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.139, "step": 13497 }, { "epoch": 20.206586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 13498 }, { "epoch": 20.20808383233533, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1397, "step": 13499 }, { "epoch": 20.209580838323355, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1425, "step": 13500 }, { "epoch": 20.211077844311376, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1469, "step": 13501 }, { "epoch": 20.2125748502994, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1368, "step": 13502 }, { "epoch": 20.214071856287426, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1348, "step": 13503 }, { "epoch": 20.21556886227545, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1368, "step": 13504 }, { "epoch": 20.21706586826347, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1395, "step": 13505 }, { "epoch": 20.218562874251496, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1354, "step": 13506 }, { "epoch": 20.22005988023952, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1358, "step": 13507 }, { "epoch": 20.221556886227546, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.134, "step": 13508 }, { "epoch": 20.223053892215567, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1371, "step": 13509 }, { "epoch": 20.224550898203592, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1338, "step": 13510 }, { "epoch": 20.226047904191617, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13511 }, { "epoch": 20.227544910179642, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1312, "step": 13512 }, { "epoch": 20.229041916167663, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13513 }, { "epoch": 20.230538922155688, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 13514 }, { "epoch": 20.232035928143713, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1357, "step": 13515 }, { "epoch": 20.233532934131738, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.137, "step": 13516 }, { "epoch": 20.23502994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1322, "step": 13517 }, { "epoch": 20.236526946107784, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1381, "step": 13518 }, { "epoch": 20.23802395209581, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1422, "step": 13519 }, { "epoch": 20.239520958083833, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 13520 }, { "epoch": 20.241017964071855, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1325, "step": 13521 }, { "epoch": 20.24251497005988, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 13522 }, { "epoch": 20.244011976047904, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1361, "step": 13523 }, { "epoch": 20.24550898203593, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1312, "step": 13524 }, { "epoch": 20.24700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1336, "step": 13525 }, { "epoch": 20.248502994011975, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1277, "step": 13526 }, { "epoch": 20.25, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1312, "step": 13527 }, { "epoch": 20.251497005988025, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.133, "step": 13528 }, { "epoch": 20.25299401197605, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.138, "step": 13529 }, { "epoch": 20.25449101796407, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1379, "step": 13530 }, { "epoch": 20.255988023952096, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1368, "step": 13531 }, { "epoch": 20.25748502994012, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1333, "step": 13532 }, { "epoch": 20.258982035928145, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 13533 }, { "epoch": 20.260479041916167, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13534 }, { "epoch": 20.26197604790419, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1385, "step": 13535 }, { "epoch": 20.263473053892216, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1347, "step": 13536 }, { "epoch": 20.26497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1339, "step": 13537 }, { "epoch": 20.266467065868262, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.133, "step": 13538 }, { "epoch": 20.267964071856287, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.142, "step": 13539 }, { "epoch": 20.269461077844312, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 13540 }, { "epoch": 20.270958083832337, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1286, "step": 13541 }, { "epoch": 20.272455089820358, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1347, "step": 13542 }, { "epoch": 20.273952095808383, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13543 }, { "epoch": 20.275449101796408, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1315, "step": 13544 }, { "epoch": 20.276946107784433, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13545 }, { "epoch": 20.278443113772454, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1331, "step": 13546 }, { "epoch": 20.27994011976048, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1329, "step": 13547 }, { "epoch": 20.281437125748504, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1376, "step": 13548 }, { "epoch": 20.28293413173653, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1374, "step": 13549 }, { "epoch": 20.28443113772455, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1374, "step": 13550 }, { "epoch": 20.285928143712574, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1358, "step": 13551 }, { "epoch": 20.2874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.138, "step": 13552 }, { "epoch": 20.288922155688624, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.136, "step": 13553 }, { "epoch": 20.290419161676645, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1361, "step": 13554 }, { "epoch": 20.29191616766467, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1402, "step": 13555 }, { "epoch": 20.293413173652695, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1422, "step": 13556 }, { "epoch": 20.29491017964072, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1348, "step": 13557 }, { "epoch": 20.29640718562874, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1366, "step": 13558 }, { "epoch": 20.297904191616766, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.136, "step": 13559 }, { "epoch": 20.29940119760479, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1364, "step": 13560 }, { "epoch": 20.300898203592816, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1351, "step": 13561 }, { "epoch": 20.302395209580837, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1396, "step": 13562 }, { "epoch": 20.30389221556886, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1336, "step": 13563 }, { "epoch": 20.305389221556887, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1339, "step": 13564 }, { "epoch": 20.30688622754491, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1334, "step": 13565 }, { "epoch": 20.308383233532933, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 13566 }, { "epoch": 20.309880239520957, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1374, "step": 13567 }, { "epoch": 20.311377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.133, "step": 13568 }, { "epoch": 20.312874251497007, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1382, "step": 13569 }, { "epoch": 20.31437125748503, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 13570 }, { "epoch": 20.315868263473053, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1439, "step": 13571 }, { "epoch": 20.317365269461078, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.136, "step": 13572 }, { "epoch": 20.318862275449103, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.136, "step": 13573 }, { "epoch": 20.320359281437124, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 13574 }, { "epoch": 20.32185628742515, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1403, "step": 13575 }, { "epoch": 20.323353293413174, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1394, "step": 13576 }, { "epoch": 20.3248502994012, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1302, "step": 13577 }, { "epoch": 20.32634730538922, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13578 }, { "epoch": 20.327844311377245, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1348, "step": 13579 }, { "epoch": 20.32934131736527, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1322, "step": 13580 }, { "epoch": 20.330838323353294, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1315, "step": 13581 }, { "epoch": 20.33233532934132, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1322, "step": 13582 }, { "epoch": 20.33383233532934, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13583 }, { "epoch": 20.335329341317365, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13584 }, { "epoch": 20.33682634730539, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1364, "step": 13585 }, { "epoch": 20.338323353293415, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1342, "step": 13586 }, { "epoch": 20.339820359281436, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.143, "step": 13587 }, { "epoch": 20.34131736526946, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1283, "step": 13588 }, { "epoch": 20.342814371257486, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1438, "step": 13589 }, { "epoch": 20.34431137724551, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1293, "step": 13590 }, { "epoch": 20.345808383233532, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1297, "step": 13591 }, { "epoch": 20.347305389221557, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1426, "step": 13592 }, { "epoch": 20.34880239520958, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1378, "step": 13593 }, { "epoch": 20.350299401197606, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 13594 }, { "epoch": 20.351796407185628, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.139, "step": 13595 }, { "epoch": 20.353293413173652, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1355, "step": 13596 }, { "epoch": 20.354790419161677, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1357, "step": 13597 }, { "epoch": 20.356287425149702, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1374, "step": 13598 }, { "epoch": 20.357784431137723, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1375, "step": 13599 }, { "epoch": 20.35928143712575, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.131, "step": 13600 }, { "epoch": 20.360778443113773, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1366, "step": 13601 }, { "epoch": 20.362275449101798, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13602 }, { "epoch": 20.36377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13603 }, { "epoch": 20.365269461077844, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.142, "step": 13604 }, { "epoch": 20.36676646706587, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1369, "step": 13605 }, { "epoch": 20.368263473053894, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1385, "step": 13606 }, { "epoch": 20.369760479041915, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1359, "step": 13607 }, { "epoch": 20.37125748502994, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.136, "step": 13608 }, { "epoch": 20.372754491017965, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.138, "step": 13609 }, { "epoch": 20.37425149700599, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1338, "step": 13610 }, { "epoch": 20.37574850299401, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1416, "step": 13611 }, { "epoch": 20.377245508982035, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1373, "step": 13612 }, { "epoch": 20.37874251497006, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1337, "step": 13613 }, { "epoch": 20.380239520958085, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13614 }, { "epoch": 20.381736526946106, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 13615 }, { "epoch": 20.38323353293413, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1286, "step": 13616 }, { "epoch": 20.384730538922156, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1468, "step": 13617 }, { "epoch": 20.38622754491018, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1398, "step": 13618 }, { "epoch": 20.387724550898202, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1346, "step": 13619 }, { "epoch": 20.389221556886227, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1418, "step": 13620 }, { "epoch": 20.39071856287425, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1353, "step": 13621 }, { "epoch": 20.392215568862277, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.135, "step": 13622 }, { "epoch": 20.393712574850298, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1393, "step": 13623 }, { "epoch": 20.395209580838323, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 13624 }, { "epoch": 20.396706586826348, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1367, "step": 13625 }, { "epoch": 20.398203592814372, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1375, "step": 13626 }, { "epoch": 20.399700598802394, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1448, "step": 13627 }, { "epoch": 20.40119760479042, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1372, "step": 13628 }, { "epoch": 20.402694610778443, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1357, "step": 13629 }, { "epoch": 20.404191616766468, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13630 }, { "epoch": 20.40568862275449, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13631 }, { "epoch": 20.407185628742514, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1308, "step": 13632 }, { "epoch": 20.40868263473054, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.136, "step": 13633 }, { "epoch": 20.410179640718564, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1246, "step": 13634 }, { "epoch": 20.411676646706585, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1272, "step": 13635 }, { "epoch": 20.41317365269461, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 13636 }, { "epoch": 20.414670658682635, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1384, "step": 13637 }, { "epoch": 20.41616766467066, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1352, "step": 13638 }, { "epoch": 20.41766467065868, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1381, "step": 13639 }, { "epoch": 20.419161676646706, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1351, "step": 13640 }, { "epoch": 20.42065868263473, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1388, "step": 13641 }, { "epoch": 20.422155688622755, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.13, "step": 13642 }, { "epoch": 20.42365269461078, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 13643 }, { "epoch": 20.4251497005988, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1435, "step": 13644 }, { "epoch": 20.426646706586826, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 13645 }, { "epoch": 20.42814371257485, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1389, "step": 13646 }, { "epoch": 20.429640718562876, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1437, "step": 13647 }, { "epoch": 20.431137724550897, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13648 }, { "epoch": 20.432634730538922, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1428, "step": 13649 }, { "epoch": 20.434131736526947, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1364, "step": 13650 }, { "epoch": 20.43562874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1413, "step": 13651 }, { "epoch": 20.437125748502993, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1375, "step": 13652 }, { "epoch": 20.438622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1375, "step": 13653 }, { "epoch": 20.440119760479043, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.138, "step": 13654 }, { "epoch": 20.441616766467067, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1412, "step": 13655 }, { "epoch": 20.44311377245509, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1371, "step": 13656 }, { "epoch": 20.444610778443113, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1377, "step": 13657 }, { "epoch": 20.44610778443114, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1322, "step": 13658 }, { "epoch": 20.447604790419163, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1386, "step": 13659 }, { "epoch": 20.449101796407184, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 13660 }, { "epoch": 20.45059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1363, "step": 13661 }, { "epoch": 20.452095808383234, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1364, "step": 13662 }, { "epoch": 20.45359281437126, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1334, "step": 13663 }, { "epoch": 20.45508982035928, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1417, "step": 13664 }, { "epoch": 20.456586826347305, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1412, "step": 13665 }, { "epoch": 20.45808383233533, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1359, "step": 13666 }, { "epoch": 20.459580838323355, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1368, "step": 13667 }, { "epoch": 20.461077844311376, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 13668 }, { "epoch": 20.4625748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 13669 }, { "epoch": 20.464071856287426, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.14, "step": 13670 }, { "epoch": 20.46556886227545, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13671 }, { "epoch": 20.46706586826347, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 13672 }, { "epoch": 20.468562874251496, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1328, "step": 13673 }, { "epoch": 20.47005988023952, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1319, "step": 13674 }, { "epoch": 20.471556886227546, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1344, "step": 13675 }, { "epoch": 20.473053892215567, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1372, "step": 13676 }, { "epoch": 20.474550898203592, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1362, "step": 13677 }, { "epoch": 20.476047904191617, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.134, "step": 13678 }, { "epoch": 20.477544910179642, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1363, "step": 13679 }, { "epoch": 20.479041916167663, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1408, "step": 13680 }, { "epoch": 20.480538922155688, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1334, "step": 13681 }, { "epoch": 20.482035928143713, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1355, "step": 13682 }, { "epoch": 20.483532934131738, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1363, "step": 13683 }, { "epoch": 20.48502994011976, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 13684 }, { "epoch": 20.486526946107784, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1368, "step": 13685 }, { "epoch": 20.48802395209581, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.128, "step": 13686 }, { "epoch": 20.489520958083833, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1269, "step": 13687 }, { "epoch": 20.491017964071855, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.136, "step": 13688 }, { "epoch": 20.49251497005988, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13689 }, { "epoch": 20.494011976047904, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.1311, "step": 13690 }, { "epoch": 20.49550898203593, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.1351, "step": 13691 }, { "epoch": 20.49700598802395, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1397, "step": 13692 }, { "epoch": 20.498502994011975, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1382, "step": 13693 }, { "epoch": 20.5, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1342, "step": 13694 }, { "epoch": 20.501497005988025, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1369, "step": 13695 }, { "epoch": 20.50299401197605, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1344, "step": 13696 }, { "epoch": 20.50449101796407, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1366, "step": 13697 }, { "epoch": 20.505988023952096, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1422, "step": 13698 }, { "epoch": 20.50748502994012, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1347, "step": 13699 }, { "epoch": 20.508982035928145, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 13700 }, { "epoch": 20.510479041916167, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1406, "step": 13701 }, { "epoch": 20.51197604790419, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1372, "step": 13702 }, { "epoch": 20.513473053892216, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13703 }, { "epoch": 20.51497005988024, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1394, "step": 13704 }, { "epoch": 20.516467065868262, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13705 }, { "epoch": 20.517964071856287, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1273, "step": 13706 }, { "epoch": 20.519461077844312, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.138, "step": 13707 }, { "epoch": 20.520958083832337, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.137, "step": 13708 }, { "epoch": 20.522455089820358, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1341, "step": 13709 }, { "epoch": 20.523952095808383, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1361, "step": 13710 }, { "epoch": 20.525449101796408, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1415, "step": 13711 }, { "epoch": 20.526946107784433, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13712 }, { "epoch": 20.528443113772454, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1347, "step": 13713 }, { "epoch": 20.52994011976048, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1347, "step": 13714 }, { "epoch": 20.531437125748504, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 13715 }, { "epoch": 20.53293413173653, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1329, "step": 13716 }, { "epoch": 20.53443113772455, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1374, "step": 13717 }, { "epoch": 20.535928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.136, "step": 13718 }, { "epoch": 20.5374251497006, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 13719 }, { "epoch": 20.538922155688624, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1351, "step": 13720 }, { "epoch": 20.540419161676645, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1305, "step": 13721 }, { "epoch": 20.54191616766467, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1357, "step": 13722 }, { "epoch": 20.543413173652695, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1345, "step": 13723 }, { "epoch": 20.54491017964072, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 13724 }, { "epoch": 20.54640718562874, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1378, "step": 13725 }, { "epoch": 20.547904191616766, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1295, "step": 13726 }, { "epoch": 20.54940119760479, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1355, "step": 13727 }, { "epoch": 20.550898203592816, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1339, "step": 13728 }, { "epoch": 20.552395209580837, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1367, "step": 13729 }, { "epoch": 20.55389221556886, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1363, "step": 13730 }, { "epoch": 20.555389221556887, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1431, "step": 13731 }, { "epoch": 20.55688622754491, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13732 }, { "epoch": 20.558383233532933, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1404, "step": 13733 }, { "epoch": 20.559880239520957, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1404, "step": 13734 }, { "epoch": 20.561377245508982, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1358, "step": 13735 }, { "epoch": 20.562874251497007, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1348, "step": 13736 }, { "epoch": 20.56437125748503, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1439, "step": 13737 }, { "epoch": 20.565868263473053, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1356, "step": 13738 }, { "epoch": 20.567365269461078, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.135, "step": 13739 }, { "epoch": 20.568862275449103, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1404, "step": 13740 }, { "epoch": 20.570359281437124, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1305, "step": 13741 }, { "epoch": 20.57185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.142, "step": 13742 }, { "epoch": 20.573353293413174, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13743 }, { "epoch": 20.5748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.142, "step": 13744 }, { "epoch": 20.57634730538922, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13745 }, { "epoch": 20.577844311377245, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.136, "step": 13746 }, { "epoch": 20.57934131736527, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.135, "step": 13747 }, { "epoch": 20.580838323353294, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13748 }, { "epoch": 20.58233532934132, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1304, "step": 13749 }, { "epoch": 20.58383233532934, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1361, "step": 13750 }, { "epoch": 20.585329341317365, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1255, "step": 13751 }, { "epoch": 20.58682634730539, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1376, "step": 13752 }, { "epoch": 20.58832335329341, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1317, "step": 13753 }, { "epoch": 20.589820359281436, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1289, "step": 13754 }, { "epoch": 20.59131736526946, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 13755 }, { "epoch": 20.592814371257486, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1326, "step": 13756 }, { "epoch": 20.59431137724551, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1342, "step": 13757 }, { "epoch": 20.595808383233532, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1369, "step": 13758 }, { "epoch": 20.597305389221557, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1388, "step": 13759 }, { "epoch": 20.59880239520958, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.131, "step": 13760 }, { "epoch": 20.600299401197606, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1357, "step": 13761 }, { "epoch": 20.601796407185628, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1289, "step": 13762 }, { "epoch": 20.603293413173652, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1387, "step": 13763 }, { "epoch": 20.604790419161677, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13764 }, { "epoch": 20.606287425149702, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1393, "step": 13765 }, { "epoch": 20.607784431137723, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 13766 }, { "epoch": 20.60928143712575, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1368, "step": 13767 }, { "epoch": 20.610778443113773, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1314, "step": 13768 }, { "epoch": 20.612275449101798, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1363, "step": 13769 }, { "epoch": 20.61377245508982, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1267, "step": 13770 }, { "epoch": 20.615269461077844, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1371, "step": 13771 }, { "epoch": 20.61676646706587, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1409, "step": 13772 }, { "epoch": 20.618263473053894, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1305, "step": 13773 }, { "epoch": 20.619760479041915, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1361, "step": 13774 }, { "epoch": 20.62125748502994, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1346, "step": 13775 }, { "epoch": 20.622754491017965, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1307, "step": 13776 }, { "epoch": 20.62425149700599, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1404, "step": 13777 }, { "epoch": 20.62574850299401, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1356, "step": 13778 }, { "epoch": 20.627245508982035, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1299, "step": 13779 }, { "epoch": 20.62874251497006, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1409, "step": 13780 }, { "epoch": 20.630239520958085, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1448, "step": 13781 }, { "epoch": 20.631736526946106, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1303, "step": 13782 }, { "epoch": 20.63323353293413, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 13783 }, { "epoch": 20.634730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1326, "step": 13784 }, { "epoch": 20.63622754491018, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13785 }, { "epoch": 20.637724550898202, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1305, "step": 13786 }, { "epoch": 20.639221556886227, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1341, "step": 13787 }, { "epoch": 20.64071856287425, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.14, "step": 13788 }, { "epoch": 20.642215568862277, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1401, "step": 13789 }, { "epoch": 20.643712574850298, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1357, "step": 13790 }, { "epoch": 20.645209580838323, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1345, "step": 13791 }, { "epoch": 20.646706586826348, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1368, "step": 13792 }, { "epoch": 20.648203592814372, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1335, "step": 13793 }, { "epoch": 20.649700598802394, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1299, "step": 13794 }, { "epoch": 20.65119760479042, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1466, "step": 13795 }, { "epoch": 20.652694610778443, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1365, "step": 13796 }, { "epoch": 20.654191616766468, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1335, "step": 13797 }, { "epoch": 20.65568862275449, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13798 }, { "epoch": 20.657185628742514, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1322, "step": 13799 }, { "epoch": 20.65868263473054, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1423, "step": 13800 }, { "epoch": 20.660179640718564, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1432, "step": 13801 }, { "epoch": 20.66167664670659, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.137, "step": 13802 }, { "epoch": 20.66317365269461, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1299, "step": 13803 }, { "epoch": 20.664670658682635, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1361, "step": 13804 }, { "epoch": 20.66616766467066, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 13805 }, { "epoch": 20.66766467065868, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1372, "step": 13806 }, { "epoch": 20.669161676646706, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13807 }, { "epoch": 20.67065868263473, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1321, "step": 13808 }, { "epoch": 20.672155688622755, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13809 }, { "epoch": 20.67365269461078, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1385, "step": 13810 }, { "epoch": 20.6751497005988, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1377, "step": 13811 }, { "epoch": 20.676646706586826, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1316, "step": 13812 }, { "epoch": 20.67814371257485, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1399, "step": 13813 }, { "epoch": 20.679640718562876, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1422, "step": 13814 }, { "epoch": 20.681137724550897, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.133, "step": 13815 }, { "epoch": 20.682634730538922, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1345, "step": 13816 }, { "epoch": 20.684131736526947, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1384, "step": 13817 }, { "epoch": 20.68562874251497, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1377, "step": 13818 }, { "epoch": 20.687125748502993, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1281, "step": 13819 }, { "epoch": 20.688622754491018, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1382, "step": 13820 }, { "epoch": 20.690119760479043, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1385, "step": 13821 }, { "epoch": 20.691616766467067, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1388, "step": 13822 }, { "epoch": 20.69311377245509, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1393, "step": 13823 }, { "epoch": 20.694610778443113, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1411, "step": 13824 }, { "epoch": 20.69610778443114, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1346, "step": 13825 }, { "epoch": 20.697604790419163, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1335, "step": 13826 }, { "epoch": 20.699101796407184, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1348, "step": 13827 }, { "epoch": 20.70059880239521, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1367, "step": 13828 }, { "epoch": 20.702095808383234, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1376, "step": 13829 }, { "epoch": 20.70359281437126, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1337, "step": 13830 }, { "epoch": 20.70508982035928, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1387, "step": 13831 }, { "epoch": 20.706586826347305, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1365, "step": 13832 }, { "epoch": 20.70808383233533, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1314, "step": 13833 }, { "epoch": 20.709580838323355, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1348, "step": 13834 }, { "epoch": 20.711077844311376, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1308, "step": 13835 }, { "epoch": 20.7125748502994, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1367, "step": 13836 }, { "epoch": 20.714071856287426, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1408, "step": 13837 }, { "epoch": 20.71556886227545, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1383, "step": 13838 }, { "epoch": 20.71706586826347, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1396, "step": 13839 }, { "epoch": 20.718562874251496, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1321, "step": 13840 }, { "epoch": 20.72005988023952, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1385, "step": 13841 }, { "epoch": 20.721556886227546, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1371, "step": 13842 }, { "epoch": 20.723053892215567, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1354, "step": 13843 }, { "epoch": 20.724550898203592, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1308, "step": 13844 }, { "epoch": 20.726047904191617, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1389, "step": 13845 }, { "epoch": 20.727544910179642, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1322, "step": 13846 }, { "epoch": 20.729041916167663, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1321, "step": 13847 }, { "epoch": 20.730538922155688, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1398, "step": 13848 }, { "epoch": 20.732035928143713, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1379, "step": 13849 }, { "epoch": 20.733532934131738, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1384, "step": 13850 }, { "epoch": 20.73502994011976, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1443, "step": 13851 }, { "epoch": 20.736526946107784, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 13852 }, { "epoch": 20.73802395209581, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1317, "step": 13853 }, { "epoch": 20.739520958083833, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1406, "step": 13854 }, { "epoch": 20.741017964071855, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1247, "step": 13855 }, { "epoch": 20.74251497005988, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1406, "step": 13856 }, { "epoch": 20.744011976047904, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 13857 }, { "epoch": 20.74550898203593, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1344, "step": 13858 }, { "epoch": 20.74700598802395, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1361, "step": 13859 }, { "epoch": 20.748502994011975, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1415, "step": 13860 }, { "epoch": 20.75, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1393, "step": 13861 }, { "epoch": 20.751497005988025, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1388, "step": 13862 }, { "epoch": 20.75299401197605, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1339, "step": 13863 }, { "epoch": 20.75449101796407, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.134, "step": 13864 }, { "epoch": 20.755988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1296, "step": 13865 }, { "epoch": 20.75748502994012, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1362, "step": 13866 }, { "epoch": 20.758982035928145, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1325, "step": 13867 }, { "epoch": 20.760479041916167, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1372, "step": 13868 }, { "epoch": 20.76197604790419, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.144, "step": 13869 }, { "epoch": 20.763473053892216, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1388, "step": 13870 }, { "epoch": 20.76497005988024, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 13871 }, { "epoch": 20.766467065868262, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 13872 }, { "epoch": 20.767964071856287, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1373, "step": 13873 }, { "epoch": 20.769461077844312, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1388, "step": 13874 }, { "epoch": 20.770958083832337, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1338, "step": 13875 }, { "epoch": 20.772455089820358, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.133, "step": 13876 }, { "epoch": 20.773952095808383, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1363, "step": 13877 }, { "epoch": 20.775449101796408, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13878 }, { "epoch": 20.776946107784433, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1379, "step": 13879 }, { "epoch": 20.778443113772454, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1355, "step": 13880 }, { "epoch": 20.77994011976048, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 13881 }, { "epoch": 20.781437125748504, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1377, "step": 13882 }, { "epoch": 20.78293413173653, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1362, "step": 13883 }, { "epoch": 20.78443113772455, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 13884 }, { "epoch": 20.785928143712574, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1407, "step": 13885 }, { "epoch": 20.7874251497006, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1376, "step": 13886 }, { "epoch": 20.788922155688624, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.139, "step": 13887 }, { "epoch": 20.790419161676645, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1357, "step": 13888 }, { "epoch": 20.79191616766467, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1341, "step": 13889 }, { "epoch": 20.793413173652695, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1308, "step": 13890 }, { "epoch": 20.79491017964072, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13891 }, { "epoch": 20.79640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 13892 }, { "epoch": 20.797904191616766, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 13893 }, { "epoch": 20.79940119760479, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1371, "step": 13894 }, { "epoch": 20.800898203592816, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1327, "step": 13895 }, { "epoch": 20.802395209580837, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1429, "step": 13896 }, { "epoch": 20.80389221556886, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1416, "step": 13897 }, { "epoch": 20.805389221556887, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1319, "step": 13898 }, { "epoch": 20.80688622754491, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1479, "step": 13899 }, { "epoch": 20.808383233532933, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1385, "step": 13900 }, { "epoch": 20.809880239520957, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 13901 }, { "epoch": 20.811377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1452, "step": 13902 }, { "epoch": 20.812874251497007, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1308, "step": 13903 }, { "epoch": 20.81437125748503, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.133, "step": 13904 }, { "epoch": 20.815868263473053, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13905 }, { "epoch": 20.817365269461078, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1361, "step": 13906 }, { "epoch": 20.818862275449103, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1378, "step": 13907 }, { "epoch": 20.820359281437124, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1395, "step": 13908 }, { "epoch": 20.82185628742515, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1305, "step": 13909 }, { "epoch": 20.823353293413174, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1381, "step": 13910 }, { "epoch": 20.8248502994012, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.138, "step": 13911 }, { "epoch": 20.82634730538922, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1333, "step": 13912 }, { "epoch": 20.827844311377245, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1358, "step": 13913 }, { "epoch": 20.82934131736527, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 13914 }, { "epoch": 20.830838323353294, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1363, "step": 13915 }, { "epoch": 20.83233532934132, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1397, "step": 13916 }, { "epoch": 20.83383233532934, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1354, "step": 13917 }, { "epoch": 20.835329341317365, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1394, "step": 13918 }, { "epoch": 20.83682634730539, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1305, "step": 13919 }, { "epoch": 20.83832335329341, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1368, "step": 13920 }, { "epoch": 20.839820359281436, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13921 }, { "epoch": 20.84131736526946, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1419, "step": 13922 }, { "epoch": 20.842814371257486, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1367, "step": 13923 }, { "epoch": 20.84431137724551, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1372, "step": 13924 }, { "epoch": 20.845808383233532, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1338, "step": 13925 }, { "epoch": 20.847305389221557, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 13926 }, { "epoch": 20.84880239520958, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1412, "step": 13927 }, { "epoch": 20.850299401197606, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.131, "step": 13928 }, { "epoch": 20.851796407185628, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1384, "step": 13929 }, { "epoch": 20.853293413173652, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1318, "step": 13930 }, { "epoch": 20.854790419161677, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1368, "step": 13931 }, { "epoch": 20.856287425149702, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13932 }, { "epoch": 20.857784431137723, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1429, "step": 13933 }, { "epoch": 20.85928143712575, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1259, "step": 13934 }, { "epoch": 20.860778443113773, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13935 }, { "epoch": 20.862275449101798, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1428, "step": 13936 }, { "epoch": 20.86377245508982, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.13, "step": 13937 }, { "epoch": 20.865269461077844, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1351, "step": 13938 }, { "epoch": 20.86676646706587, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13939 }, { "epoch": 20.868263473053894, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1423, "step": 13940 }, { "epoch": 20.869760479041915, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.135, "step": 13941 }, { "epoch": 20.87125748502994, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1383, "step": 13942 }, { "epoch": 20.872754491017965, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1361, "step": 13943 }, { "epoch": 20.87425149700599, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1346, "step": 13944 }, { "epoch": 20.87574850299401, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1342, "step": 13945 }, { "epoch": 20.877245508982035, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1361, "step": 13946 }, { "epoch": 20.87874251497006, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13947 }, { "epoch": 20.880239520958085, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1438, "step": 13948 }, { "epoch": 20.881736526946106, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.138, "step": 13949 }, { "epoch": 20.88323353293413, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 13950 }, { "epoch": 20.884730538922156, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.132, "step": 13951 }, { "epoch": 20.88622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1412, "step": 13952 }, { "epoch": 20.887724550898202, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1321, "step": 13953 }, { "epoch": 20.889221556886227, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13954 }, { "epoch": 20.89071856287425, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1329, "step": 13955 }, { "epoch": 20.892215568862277, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1344, "step": 13956 }, { "epoch": 20.893712574850298, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1316, "step": 13957 }, { "epoch": 20.895209580838323, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1387, "step": 13958 }, { "epoch": 20.896706586826348, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1282, "step": 13959 }, { "epoch": 20.898203592814372, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13960 }, { "epoch": 20.899700598802394, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 13961 }, { "epoch": 20.90119760479042, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13962 }, { "epoch": 20.902694610778443, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 13963 }, { "epoch": 20.904191616766468, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1355, "step": 13964 }, { "epoch": 20.90568862275449, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.137, "step": 13965 }, { "epoch": 20.907185628742514, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1291, "step": 13966 }, { "epoch": 20.90868263473054, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1398, "step": 13967 }, { "epoch": 20.910179640718564, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1305, "step": 13968 }, { "epoch": 20.91167664670659, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1379, "step": 13969 }, { "epoch": 20.91317365269461, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1308, "step": 13970 }, { "epoch": 20.914670658682635, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1332, "step": 13971 }, { "epoch": 20.91616766467066, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1347, "step": 13972 }, { "epoch": 20.91766467065868, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1311, "step": 13973 }, { "epoch": 20.919161676646706, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1362, "step": 13974 }, { "epoch": 20.92065868263473, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1342, "step": 13975 }, { "epoch": 20.922155688622755, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.131, "step": 13976 }, { "epoch": 20.92365269461078, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.141, "step": 13977 }, { "epoch": 20.9251497005988, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1333, "step": 13978 }, { "epoch": 20.926646706586826, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 13979 }, { "epoch": 20.92814371257485, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1339, "step": 13980 }, { "epoch": 20.929640718562876, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1409, "step": 13981 }, { "epoch": 20.931137724550897, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13982 }, { "epoch": 20.932634730538922, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1368, "step": 13983 }, { "epoch": 20.934131736526947, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.135, "step": 13984 }, { "epoch": 20.93562874251497, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1395, "step": 13985 }, { "epoch": 20.937125748502993, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1389, "step": 13986 }, { "epoch": 20.938622754491018, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1321, "step": 13987 }, { "epoch": 20.940119760479043, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1397, "step": 13988 }, { "epoch": 20.941616766467067, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1353, "step": 13989 }, { "epoch": 20.94311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1352, "step": 13990 }, { "epoch": 20.944610778443113, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1435, "step": 13991 }, { "epoch": 20.94610778443114, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1348, "step": 13992 }, { "epoch": 20.947604790419163, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1332, "step": 13993 }, { "epoch": 20.949101796407184, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1334, "step": 13994 }, { "epoch": 20.95059880239521, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1354, "step": 13995 }, { "epoch": 20.952095808383234, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.137, "step": 13996 }, { "epoch": 20.95359281437126, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1361, "step": 13997 }, { "epoch": 20.95508982035928, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1334, "step": 13998 }, { "epoch": 20.956586826347305, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1323, "step": 13999 }, { "epoch": 20.95808383233533, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1366, "step": 14000 }, { "epoch": 20.959580838323355, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1351, "step": 14001 }, { "epoch": 20.961077844311376, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.135, "step": 14002 }, { "epoch": 20.9625748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1436, "step": 14003 }, { "epoch": 20.964071856287426, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1404, "step": 14004 }, { "epoch": 20.96556886227545, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1407, "step": 14005 }, { "epoch": 20.96706586826347, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1418, "step": 14006 }, { "epoch": 20.968562874251496, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1322, "step": 14007 }, { "epoch": 20.97005988023952, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1395, "step": 14008 }, { "epoch": 20.971556886227546, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1378, "step": 14009 }, { "epoch": 20.973053892215567, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14010 }, { "epoch": 20.974550898203592, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.141, "step": 14011 }, { "epoch": 20.976047904191617, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1386, "step": 14012 }, { "epoch": 20.977544910179642, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1393, "step": 14013 }, { "epoch": 20.979041916167663, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1339, "step": 14014 }, { "epoch": 20.980538922155688, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1372, "step": 14015 }, { "epoch": 20.982035928143713, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1341, "step": 14016 }, { "epoch": 20.983532934131738, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 14017 }, { "epoch": 20.98502994011976, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1377, "step": 14018 }, { "epoch": 20.986526946107784, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1292, "step": 14019 }, { "epoch": 20.98802395209581, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1332, "step": 14020 }, { "epoch": 20.989520958083833, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14021 }, { "epoch": 20.991017964071855, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1379, "step": 14022 }, { "epoch": 20.99251497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1379, "step": 14023 }, { "epoch": 20.994011976047904, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1392, "step": 14024 }, { "epoch": 20.99550898203593, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1344, "step": 14025 }, { "epoch": 20.99700598802395, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1295, "step": 14026 }, { "epoch": 20.998502994011975, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.139, "step": 14027 }, { "epoch": 21.0, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1373, "step": 14028 }, { "epoch": 21.001497005988025, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1394, "step": 14029 }, { "epoch": 21.00299401197605, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 14030 }, { "epoch": 21.00449101796407, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1362, "step": 14031 }, { "epoch": 21.005988023952096, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1422, "step": 14032 }, { "epoch": 21.00748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1379, "step": 14033 }, { "epoch": 21.008982035928145, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1311, "step": 14034 }, { "epoch": 21.010479041916167, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1383, "step": 14035 }, { "epoch": 21.01197604790419, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.138, "step": 14036 }, { "epoch": 21.013473053892216, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1312, "step": 14037 }, { "epoch": 21.01497005988024, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1308, "step": 14038 }, { "epoch": 21.016467065868262, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1337, "step": 14039 }, { "epoch": 21.017964071856287, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1326, "step": 14040 }, { "epoch": 21.019461077844312, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1348, "step": 14041 }, { "epoch": 21.020958083832337, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.136, "step": 14042 }, { "epoch": 21.022455089820358, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1426, "step": 14043 }, { "epoch": 21.023952095808383, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1281, "step": 14044 }, { "epoch": 21.025449101796408, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1342, "step": 14045 }, { "epoch": 21.026946107784433, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1471, "step": 14046 }, { "epoch": 21.028443113772454, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1416, "step": 14047 }, { "epoch": 21.02994011976048, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1393, "step": 14048 }, { "epoch": 21.031437125748504, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1327, "step": 14049 }, { "epoch": 21.03293413173653, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1324, "step": 14050 }, { "epoch": 21.03443113772455, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.129, "step": 14051 }, { "epoch": 21.035928143712574, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1377, "step": 14052 }, { "epoch": 21.0374251497006, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1367, "step": 14053 }, { "epoch": 21.038922155688624, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1346, "step": 14054 }, { "epoch": 21.040419161676645, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1321, "step": 14055 }, { "epoch": 21.04191616766467, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.131, "step": 14056 }, { "epoch": 21.043413173652695, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1384, "step": 14057 }, { "epoch": 21.04491017964072, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1358, "step": 14058 }, { "epoch": 21.04640718562874, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1378, "step": 14059 }, { "epoch": 21.047904191616766, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1352, "step": 14060 }, { "epoch": 21.04940119760479, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1408, "step": 14061 }, { "epoch": 21.050898203592816, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1421, "step": 14062 }, { "epoch": 21.052395209580837, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1374, "step": 14063 }, { "epoch": 21.05389221556886, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1375, "step": 14064 }, { "epoch": 21.055389221556887, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.141, "step": 14065 }, { "epoch": 21.05688622754491, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 14066 }, { "epoch": 21.058383233532933, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1298, "step": 14067 }, { "epoch": 21.059880239520957, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1389, "step": 14068 }, { "epoch": 21.061377245508982, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1365, "step": 14069 }, { "epoch": 21.062874251497007, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1358, "step": 14070 }, { "epoch": 21.06437125748503, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 14071 }, { "epoch": 21.065868263473053, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.132, "step": 14072 }, { "epoch": 21.067365269461078, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 14073 }, { "epoch": 21.068862275449103, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1346, "step": 14074 }, { "epoch": 21.070359281437124, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14075 }, { "epoch": 21.07185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1319, "step": 14076 }, { "epoch": 21.073353293413174, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.139, "step": 14077 }, { "epoch": 21.0748502994012, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.136, "step": 14078 }, { "epoch": 21.07634730538922, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 14079 }, { "epoch": 21.077844311377245, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14080 }, { "epoch": 21.07934131736527, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1375, "step": 14081 }, { "epoch": 21.080838323353294, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 14082 }, { "epoch": 21.082335329341316, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1388, "step": 14083 }, { "epoch": 21.08383233532934, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 14084 }, { "epoch": 21.085329341317365, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.135, "step": 14085 }, { "epoch": 21.08682634730539, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1317, "step": 14086 }, { "epoch": 21.088323353293415, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1384, "step": 14087 }, { "epoch": 21.089820359281436, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1349, "step": 14088 }, { "epoch": 21.09131736526946, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1318, "step": 14089 }, { "epoch": 21.092814371257486, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.134, "step": 14090 }, { "epoch": 21.09431137724551, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1274, "step": 14091 }, { "epoch": 21.095808383233532, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1364, "step": 14092 }, { "epoch": 21.097305389221557, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14093 }, { "epoch": 21.09880239520958, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1344, "step": 14094 }, { "epoch": 21.100299401197606, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1291, "step": 14095 }, { "epoch": 21.101796407185628, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1323, "step": 14096 }, { "epoch": 21.103293413173652, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1371, "step": 14097 }, { "epoch": 21.104790419161677, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1323, "step": 14098 }, { "epoch": 21.106287425149702, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 14099 }, { "epoch": 21.107784431137723, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 14100 }, { "epoch": 21.10928143712575, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.134, "step": 14101 }, { "epoch": 21.110778443113773, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1385, "step": 14102 }, { "epoch": 21.112275449101798, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1316, "step": 14103 }, { "epoch": 21.11377245508982, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1377, "step": 14104 }, { "epoch": 21.115269461077844, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1329, "step": 14105 }, { "epoch": 21.11676646706587, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1377, "step": 14106 }, { "epoch": 21.118263473053894, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.136, "step": 14107 }, { "epoch": 21.119760479041915, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1388, "step": 14108 }, { "epoch": 21.12125748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.127, "step": 14109 }, { "epoch": 21.122754491017965, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1349, "step": 14110 }, { "epoch": 21.12425149700599, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1396, "step": 14111 }, { "epoch": 21.12574850299401, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 14112 }, { "epoch": 21.127245508982035, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1399, "step": 14113 }, { "epoch": 21.12874251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1369, "step": 14114 }, { "epoch": 21.130239520958085, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1417, "step": 14115 }, { "epoch": 21.131736526946106, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 14116 }, { "epoch": 21.13323353293413, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14117 }, { "epoch": 21.134730538922156, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 14118 }, { "epoch": 21.13622754491018, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1379, "step": 14119 }, { "epoch": 21.137724550898202, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1376, "step": 14120 }, { "epoch": 21.139221556886227, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1315, "step": 14121 }, { "epoch": 21.14071856287425, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1326, "step": 14122 }, { "epoch": 21.142215568862277, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1346, "step": 14123 }, { "epoch": 21.143712574850298, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.138, "step": 14124 }, { "epoch": 21.145209580838323, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1384, "step": 14125 }, { "epoch": 21.146706586826348, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1293, "step": 14126 }, { "epoch": 21.148203592814372, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.135, "step": 14127 }, { "epoch": 21.149700598802394, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 14128 }, { "epoch": 21.15119760479042, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1391, "step": 14129 }, { "epoch": 21.152694610778443, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1354, "step": 14130 }, { "epoch": 21.154191616766468, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1372, "step": 14131 }, { "epoch": 21.15568862275449, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1337, "step": 14132 }, { "epoch": 21.157185628742514, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1323, "step": 14133 }, { "epoch": 21.15868263473054, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1344, "step": 14134 }, { "epoch": 21.160179640718564, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1235, "step": 14135 }, { "epoch": 21.161676646706585, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.141, "step": 14136 }, { "epoch": 21.16317365269461, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1365, "step": 14137 }, { "epoch": 21.164670658682635, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1392, "step": 14138 }, { "epoch": 21.16616766467066, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1359, "step": 14139 }, { "epoch": 21.16766467065868, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1325, "step": 14140 }, { "epoch": 21.169161676646706, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1324, "step": 14141 }, { "epoch": 21.17065868263473, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1416, "step": 14142 }, { "epoch": 21.172155688622755, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1274, "step": 14143 }, { "epoch": 21.17365269461078, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1357, "step": 14144 }, { "epoch": 21.1751497005988, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.133, "step": 14145 }, { "epoch": 21.176646706586826, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1361, "step": 14146 }, { "epoch": 21.17814371257485, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 14147 }, { "epoch": 21.179640718562876, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1312, "step": 14148 }, { "epoch": 21.181137724550897, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.136, "step": 14149 }, { "epoch": 21.182634730538922, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1365, "step": 14150 }, { "epoch": 21.184131736526947, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 14151 }, { "epoch": 21.18562874251497, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1341, "step": 14152 }, { "epoch": 21.187125748502993, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1367, "step": 14153 }, { "epoch": 21.188622754491018, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1383, "step": 14154 }, { "epoch": 21.190119760479043, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1482, "step": 14155 }, { "epoch": 21.191616766467067, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1347, "step": 14156 }, { "epoch": 21.19311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1423, "step": 14157 }, { "epoch": 21.194610778443113, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1343, "step": 14158 }, { "epoch": 21.19610778443114, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1318, "step": 14159 }, { "epoch": 21.197604790419163, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1301, "step": 14160 }, { "epoch": 21.199101796407184, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1301, "step": 14161 }, { "epoch": 21.20059880239521, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1386, "step": 14162 }, { "epoch": 21.202095808383234, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1361, "step": 14163 }, { "epoch": 21.20359281437126, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.128, "step": 14164 }, { "epoch": 21.20508982035928, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14165 }, { "epoch": 21.206586826347305, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1387, "step": 14166 }, { "epoch": 21.20808383233533, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1328, "step": 14167 }, { "epoch": 21.209580838323355, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1341, "step": 14168 }, { "epoch": 21.211077844311376, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1314, "step": 14169 }, { "epoch": 21.2125748502994, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14170 }, { "epoch": 21.214071856287426, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1384, "step": 14171 }, { "epoch": 21.21556886227545, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1379, "step": 14172 }, { "epoch": 21.21706586826347, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1386, "step": 14173 }, { "epoch": 21.218562874251496, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14174 }, { "epoch": 21.22005988023952, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 14175 }, { "epoch": 21.221556886227546, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1372, "step": 14176 }, { "epoch": 21.223053892215567, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1411, "step": 14177 }, { "epoch": 21.224550898203592, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1347, "step": 14178 }, { "epoch": 21.226047904191617, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1332, "step": 14179 }, { "epoch": 21.227544910179642, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1425, "step": 14180 }, { "epoch": 21.229041916167663, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1351, "step": 14181 }, { "epoch": 21.230538922155688, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1337, "step": 14182 }, { "epoch": 21.232035928143713, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 14183 }, { "epoch": 21.233532934131738, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1398, "step": 14184 }, { "epoch": 21.23502994011976, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1313, "step": 14185 }, { "epoch": 21.236526946107784, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 14186 }, { "epoch": 21.23802395209581, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14187 }, { "epoch": 21.239520958083833, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1355, "step": 14188 }, { "epoch": 21.241017964071855, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1401, "step": 14189 }, { "epoch": 21.24251497005988, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 14190 }, { "epoch": 21.244011976047904, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.14, "step": 14191 }, { "epoch": 21.24550898203593, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1329, "step": 14192 }, { "epoch": 21.24700598802395, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1261, "step": 14193 }, { "epoch": 21.248502994011975, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1311, "step": 14194 }, { "epoch": 21.25, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1349, "step": 14195 }, { "epoch": 21.251497005988025, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14196 }, { "epoch": 21.25299401197605, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1386, "step": 14197 }, { "epoch": 21.25449101796407, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1375, "step": 14198 }, { "epoch": 21.255988023952096, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1367, "step": 14199 }, { "epoch": 21.25748502994012, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.14, "step": 14200 }, { "epoch": 21.258982035928145, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.141, "step": 14201 }, { "epoch": 21.260479041916167, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1401, "step": 14202 }, { "epoch": 21.26197604790419, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1276, "step": 14203 }, { "epoch": 21.263473053892216, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1337, "step": 14204 }, { "epoch": 21.26497005988024, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1362, "step": 14205 }, { "epoch": 21.266467065868262, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14206 }, { "epoch": 21.267964071856287, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 14207 }, { "epoch": 21.269461077844312, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1317, "step": 14208 }, { "epoch": 21.270958083832337, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1351, "step": 14209 }, { "epoch": 21.272455089820358, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1319, "step": 14210 }, { "epoch": 21.273952095808383, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.14, "step": 14211 }, { "epoch": 21.275449101796408, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1407, "step": 14212 }, { "epoch": 21.276946107784433, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1318, "step": 14213 }, { "epoch": 21.278443113772454, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1306, "step": 14214 }, { "epoch": 21.27994011976048, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1315, "step": 14215 }, { "epoch": 21.281437125748504, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 14216 }, { "epoch": 21.28293413173653, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 14217 }, { "epoch": 21.28443113772455, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1382, "step": 14218 }, { "epoch": 21.285928143712574, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1373, "step": 14219 }, { "epoch": 21.2874251497006, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 14220 }, { "epoch": 21.288922155688624, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.138, "step": 14221 }, { "epoch": 21.290419161676645, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1415, "step": 14222 }, { "epoch": 21.29191616766467, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1326, "step": 14223 }, { "epoch": 21.293413173652695, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1386, "step": 14224 }, { "epoch": 21.29491017964072, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1423, "step": 14225 }, { "epoch": 21.29640718562874, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1367, "step": 14226 }, { "epoch": 21.297904191616766, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1347, "step": 14227 }, { "epoch": 21.29940119760479, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1341, "step": 14228 }, { "epoch": 21.300898203592816, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1302, "step": 14229 }, { "epoch": 21.302395209580837, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1311, "step": 14230 }, { "epoch": 21.30389221556886, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 14231 }, { "epoch": 21.305389221556887, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1409, "step": 14232 }, { "epoch": 21.30688622754491, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.134, "step": 14233 }, { "epoch": 21.308383233532933, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1357, "step": 14234 }, { "epoch": 21.309880239520957, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1376, "step": 14235 }, { "epoch": 21.311377245508982, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1354, "step": 14236 }, { "epoch": 21.312874251497007, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14237 }, { "epoch": 21.31437125748503, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1388, "step": 14238 }, { "epoch": 21.315868263473053, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 14239 }, { "epoch": 21.317365269461078, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 14240 }, { "epoch": 21.318862275449103, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14241 }, { "epoch": 21.320359281437124, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1329, "step": 14242 }, { "epoch": 21.32185628742515, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1374, "step": 14243 }, { "epoch": 21.323353293413174, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.139, "step": 14244 }, { "epoch": 21.3248502994012, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1315, "step": 14245 }, { "epoch": 21.32634730538922, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1322, "step": 14246 }, { "epoch": 21.327844311377245, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1332, "step": 14247 }, { "epoch": 21.32934131736527, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1303, "step": 14248 }, { "epoch": 21.330838323353294, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14249 }, { "epoch": 21.33233532934132, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1362, "step": 14250 }, { "epoch": 21.33383233532934, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1381, "step": 14251 }, { "epoch": 21.335329341317365, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1371, "step": 14252 }, { "epoch": 21.33682634730539, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1373, "step": 14253 }, { "epoch": 21.338323353293415, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1387, "step": 14254 }, { "epoch": 21.339820359281436, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1423, "step": 14255 }, { "epoch": 21.34131736526946, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1372, "step": 14256 }, { "epoch": 21.342814371257486, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1401, "step": 14257 }, { "epoch": 21.34431137724551, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1413, "step": 14258 }, { "epoch": 21.345808383233532, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1379, "step": 14259 }, { "epoch": 21.347305389221557, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1382, "step": 14260 }, { "epoch": 21.34880239520958, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1306, "step": 14261 }, { "epoch": 21.350299401197606, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1367, "step": 14262 }, { "epoch": 21.351796407185628, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1312, "step": 14263 }, { "epoch": 21.353293413173652, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1406, "step": 14264 }, { "epoch": 21.354790419161677, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 14265 }, { "epoch": 21.356287425149702, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1343, "step": 14266 }, { "epoch": 21.357784431137723, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1344, "step": 14267 }, { "epoch": 21.35928143712575, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1338, "step": 14268 }, { "epoch": 21.360778443113773, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1333, "step": 14269 }, { "epoch": 21.362275449101798, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1383, "step": 14270 }, { "epoch": 21.36377245508982, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1341, "step": 14271 }, { "epoch": 21.365269461077844, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.138, "step": 14272 }, { "epoch": 21.36676646706587, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1434, "step": 14273 }, { "epoch": 21.368263473053894, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1322, "step": 14274 }, { "epoch": 21.369760479041915, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1348, "step": 14275 }, { "epoch": 21.37125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.134, "step": 14276 }, { "epoch": 21.372754491017965, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1313, "step": 14277 }, { "epoch": 21.37425149700599, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.142, "step": 14278 }, { "epoch": 21.37574850299401, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1398, "step": 14279 }, { "epoch": 21.377245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1431, "step": 14280 }, { "epoch": 21.37874251497006, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1346, "step": 14281 }, { "epoch": 21.380239520958085, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14282 }, { "epoch": 21.381736526946106, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1345, "step": 14283 }, { "epoch": 21.38323353293413, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 14284 }, { "epoch": 21.384730538922156, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 14285 }, { "epoch": 21.38622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 14286 }, { "epoch": 21.387724550898202, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1394, "step": 14287 }, { "epoch": 21.389221556886227, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1306, "step": 14288 }, { "epoch": 21.39071856287425, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1304, "step": 14289 }, { "epoch": 21.392215568862277, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1342, "step": 14290 }, { "epoch": 21.393712574850298, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.139, "step": 14291 }, { "epoch": 21.395209580838323, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 14292 }, { "epoch": 21.396706586826348, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 14293 }, { "epoch": 21.398203592814372, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1383, "step": 14294 }, { "epoch": 21.399700598802394, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1366, "step": 14295 }, { "epoch": 21.40119760479042, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1337, "step": 14296 }, { "epoch": 21.402694610778443, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1408, "step": 14297 }, { "epoch": 21.404191616766468, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14298 }, { "epoch": 21.40568862275449, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1333, "step": 14299 }, { "epoch": 21.407185628742514, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1323, "step": 14300 }, { "epoch": 21.40868263473054, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1349, "step": 14301 }, { "epoch": 21.410179640718564, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1338, "step": 14302 }, { "epoch": 21.411676646706585, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1407, "step": 14303 }, { "epoch": 21.41317365269461, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1364, "step": 14304 }, { "epoch": 21.414670658682635, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1398, "step": 14305 }, { "epoch": 21.41616766467066, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1354, "step": 14306 }, { "epoch": 21.41766467065868, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1373, "step": 14307 }, { "epoch": 21.419161676646706, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1353, "step": 14308 }, { "epoch": 21.42065868263473, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.129, "step": 14309 }, { "epoch": 21.422155688622755, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1368, "step": 14310 }, { "epoch": 21.42365269461078, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1312, "step": 14311 }, { "epoch": 21.4251497005988, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1356, "step": 14312 }, { "epoch": 21.426646706586826, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1329, "step": 14313 }, { "epoch": 21.42814371257485, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1336, "step": 14314 }, { "epoch": 21.429640718562876, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.134, "step": 14315 }, { "epoch": 21.431137724550897, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1354, "step": 14316 }, { "epoch": 21.432634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1398, "step": 14317 }, { "epoch": 21.434131736526947, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1323, "step": 14318 }, { "epoch": 21.43562874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1284, "step": 14319 }, { "epoch": 21.437125748502993, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1362, "step": 14320 }, { "epoch": 21.438622754491018, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1374, "step": 14321 }, { "epoch": 21.440119760479043, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1323, "step": 14322 }, { "epoch": 21.441616766467067, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1342, "step": 14323 }, { "epoch": 21.44311377245509, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1308, "step": 14324 }, { "epoch": 21.444610778443113, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14325 }, { "epoch": 21.44610778443114, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.136, "step": 14326 }, { "epoch": 21.447604790419163, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 14327 }, { "epoch": 21.449101796407184, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 14328 }, { "epoch": 21.45059880239521, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 14329 }, { "epoch": 21.452095808383234, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1361, "step": 14330 }, { "epoch": 21.45359281437126, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1324, "step": 14331 }, { "epoch": 21.45508982035928, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1376, "step": 14332 }, { "epoch": 21.456586826347305, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1285, "step": 14333 }, { "epoch": 21.45808383233533, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1332, "step": 14334 }, { "epoch": 21.459580838323355, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1394, "step": 14335 }, { "epoch": 21.461077844311376, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1385, "step": 14336 }, { "epoch": 21.4625748502994, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1378, "step": 14337 }, { "epoch": 21.464071856287426, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1292, "step": 14338 }, { "epoch": 21.46556886227545, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 14339 }, { "epoch": 21.46706586826347, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14340 }, { "epoch": 21.468562874251496, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 14341 }, { "epoch": 21.47005988023952, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1409, "step": 14342 }, { "epoch": 21.471556886227546, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1353, "step": 14343 }, { "epoch": 21.473053892215567, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1362, "step": 14344 }, { "epoch": 21.474550898203592, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 14345 }, { "epoch": 21.476047904191617, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 14346 }, { "epoch": 21.477544910179642, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1423, "step": 14347 }, { "epoch": 21.479041916167663, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1356, "step": 14348 }, { "epoch": 21.480538922155688, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 14349 }, { "epoch": 21.482035928143713, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1364, "step": 14350 }, { "epoch": 21.483532934131738, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1296, "step": 14351 }, { "epoch": 21.48502994011976, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1286, "step": 14352 }, { "epoch": 21.486526946107784, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1315, "step": 14353 }, { "epoch": 21.48802395209581, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1273, "step": 14354 }, { "epoch": 21.489520958083833, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1374, "step": 14355 }, { "epoch": 21.491017964071855, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 14356 }, { "epoch": 21.49251497005988, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1312, "step": 14357 }, { "epoch": 21.494011976047904, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1365, "step": 14358 }, { "epoch": 21.49550898203593, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14359 }, { "epoch": 21.49700598802395, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 14360 }, { "epoch": 21.498502994011975, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1359, "step": 14361 }, { "epoch": 21.5, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1374, "step": 14362 }, { "epoch": 21.501497005988025, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1349, "step": 14363 }, { "epoch": 21.50299401197605, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 14364 }, { "epoch": 21.50449101796407, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1392, "step": 14365 }, { "epoch": 21.505988023952096, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.143, "step": 14366 }, { "epoch": 21.50748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1382, "step": 14367 }, { "epoch": 21.508982035928145, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.137, "step": 14368 }, { "epoch": 21.510479041916167, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1366, "step": 14369 }, { "epoch": 21.51197604790419, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.127, "step": 14370 }, { "epoch": 21.513473053892216, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1368, "step": 14371 }, { "epoch": 21.51497005988024, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 14372 }, { "epoch": 21.516467065868262, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14373 }, { "epoch": 21.517964071856287, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1345, "step": 14374 }, { "epoch": 21.519461077844312, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.13, "step": 14375 }, { "epoch": 21.520958083832337, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1387, "step": 14376 }, { "epoch": 21.522455089820358, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1315, "step": 14377 }, { "epoch": 21.523952095808383, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1317, "step": 14378 }, { "epoch": 21.525449101796408, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1302, "step": 14379 }, { "epoch": 21.526946107784433, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1363, "step": 14380 }, { "epoch": 21.528443113772454, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14381 }, { "epoch": 21.52994011976048, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1365, "step": 14382 }, { "epoch": 21.531437125748504, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1356, "step": 14383 }, { "epoch": 21.53293413173653, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.13, "step": 14384 }, { "epoch": 21.53443113772455, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1344, "step": 14385 }, { "epoch": 21.535928143712574, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1337, "step": 14386 }, { "epoch": 21.5374251497006, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1349, "step": 14387 }, { "epoch": 21.538922155688624, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1283, "step": 14388 }, { "epoch": 21.540419161676645, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1376, "step": 14389 }, { "epoch": 21.54191616766467, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1365, "step": 14390 }, { "epoch": 21.543413173652695, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1299, "step": 14391 }, { "epoch": 21.54491017964072, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1394, "step": 14392 }, { "epoch": 21.54640718562874, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 14393 }, { "epoch": 21.547904191616766, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1315, "step": 14394 }, { "epoch": 21.54940119760479, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1311, "step": 14395 }, { "epoch": 21.550898203592816, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1259, "step": 14396 }, { "epoch": 21.552395209580837, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1353, "step": 14397 }, { "epoch": 21.55389221556886, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1263, "step": 14398 }, { "epoch": 21.555389221556887, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1383, "step": 14399 }, { "epoch": 21.55688622754491, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1373, "step": 14400 }, { "epoch": 21.558383233532933, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1309, "step": 14401 }, { "epoch": 21.559880239520957, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1351, "step": 14402 }, { "epoch": 21.561377245508982, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1329, "step": 14403 }, { "epoch": 21.562874251497007, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1315, "step": 14404 }, { "epoch": 21.56437125748503, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1388, "step": 14405 }, { "epoch": 21.565868263473053, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1384, "step": 14406 }, { "epoch": 21.567365269461078, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1283, "step": 14407 }, { "epoch": 21.568862275449103, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1307, "step": 14408 }, { "epoch": 21.570359281437124, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1397, "step": 14409 }, { "epoch": 21.57185628742515, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1384, "step": 14410 }, { "epoch": 21.573353293413174, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1338, "step": 14411 }, { "epoch": 21.5748502994012, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1357, "step": 14412 }, { "epoch": 21.57634730538922, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 14413 }, { "epoch": 21.577844311377245, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1378, "step": 14414 }, { "epoch": 21.57934131736527, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1332, "step": 14415 }, { "epoch": 21.580838323353294, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 14416 }, { "epoch": 21.58233532934132, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1401, "step": 14417 }, { "epoch": 21.58383233532934, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1363, "step": 14418 }, { "epoch": 21.585329341317365, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1284, "step": 14419 }, { "epoch": 21.58682634730539, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.136, "step": 14420 }, { "epoch": 21.58832335329341, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 14421 }, { "epoch": 21.589820359281436, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1294, "step": 14422 }, { "epoch": 21.59131736526946, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1376, "step": 14423 }, { "epoch": 21.592814371257486, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1319, "step": 14424 }, { "epoch": 21.59431137724551, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1338, "step": 14425 }, { "epoch": 21.595808383233532, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1389, "step": 14426 }, { "epoch": 21.597305389221557, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1392, "step": 14427 }, { "epoch": 21.59880239520958, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1379, "step": 14428 }, { "epoch": 21.600299401197606, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1361, "step": 14429 }, { "epoch": 21.601796407185628, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.136, "step": 14430 }, { "epoch": 21.603293413173652, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1406, "step": 14431 }, { "epoch": 21.604790419161677, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.131, "step": 14432 }, { "epoch": 21.606287425149702, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1303, "step": 14433 }, { "epoch": 21.607784431137723, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1402, "step": 14434 }, { "epoch": 21.60928143712575, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 14435 }, { "epoch": 21.610778443113773, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1351, "step": 14436 }, { "epoch": 21.612275449101798, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1268, "step": 14437 }, { "epoch": 21.61377245508982, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1359, "step": 14438 }, { "epoch": 21.615269461077844, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1398, "step": 14439 }, { "epoch": 21.61676646706587, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1351, "step": 14440 }, { "epoch": 21.618263473053894, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 14441 }, { "epoch": 21.619760479041915, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1339, "step": 14442 }, { "epoch": 21.62125748502994, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1338, "step": 14443 }, { "epoch": 21.622754491017965, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1367, "step": 14444 }, { "epoch": 21.62425149700599, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1381, "step": 14445 }, { "epoch": 21.62574850299401, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1335, "step": 14446 }, { "epoch": 21.627245508982035, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.134, "step": 14447 }, { "epoch": 21.62874251497006, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1328, "step": 14448 }, { "epoch": 21.630239520958085, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1354, "step": 14449 }, { "epoch": 21.631736526946106, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1363, "step": 14450 }, { "epoch": 21.63323353293413, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1336, "step": 14451 }, { "epoch": 21.634730538922156, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.129, "step": 14452 }, { "epoch": 21.63622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14453 }, { "epoch": 21.637724550898202, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1338, "step": 14454 }, { "epoch": 21.639221556886227, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1329, "step": 14455 }, { "epoch": 21.64071856287425, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1317, "step": 14456 }, { "epoch": 21.642215568862277, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1272, "step": 14457 }, { "epoch": 21.643712574850298, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1305, "step": 14458 }, { "epoch": 21.645209580838323, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1353, "step": 14459 }, { "epoch": 21.646706586826348, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1354, "step": 14460 }, { "epoch": 21.648203592814372, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.137, "step": 14461 }, { "epoch": 21.649700598802394, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1361, "step": 14462 }, { "epoch": 21.65119760479042, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1337, "step": 14463 }, { "epoch": 21.652694610778443, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1419, "step": 14464 }, { "epoch": 21.654191616766468, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1333, "step": 14465 }, { "epoch": 21.65568862275449, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1322, "step": 14466 }, { "epoch": 21.657185628742514, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1352, "step": 14467 }, { "epoch": 21.65868263473054, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 14468 }, { "epoch": 21.660179640718564, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1388, "step": 14469 }, { "epoch": 21.66167664670659, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1368, "step": 14470 }, { "epoch": 21.66317365269461, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.1296, "step": 14471 }, { "epoch": 21.664670658682635, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.128, "step": 14472 }, { "epoch": 21.66616766467066, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1354, "step": 14473 }, { "epoch": 21.66766467065868, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1361, "step": 14474 }, { "epoch": 21.669161676646706, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1376, "step": 14475 }, { "epoch": 21.67065868263473, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.129, "step": 14476 }, { "epoch": 21.672155688622755, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1415, "step": 14477 }, { "epoch": 21.67365269461078, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.133, "step": 14478 }, { "epoch": 21.6751497005988, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14479 }, { "epoch": 21.676646706586826, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1322, "step": 14480 }, { "epoch": 21.67814371257485, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1383, "step": 14481 }, { "epoch": 21.679640718562876, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 14482 }, { "epoch": 21.681137724550897, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1306, "step": 14483 }, { "epoch": 21.682634730538922, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1321, "step": 14484 }, { "epoch": 21.684131736526947, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14485 }, { "epoch": 21.68562874251497, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1372, "step": 14486 }, { "epoch": 21.687125748502993, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1255, "step": 14487 }, { "epoch": 21.688622754491018, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1344, "step": 14488 }, { "epoch": 21.690119760479043, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1297, "step": 14489 }, { "epoch": 21.691616766467067, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1279, "step": 14490 }, { "epoch": 21.69311377245509, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1378, "step": 14491 }, { "epoch": 21.694610778443113, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 14492 }, { "epoch": 21.69610778443114, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1337, "step": 14493 }, { "epoch": 21.697604790419163, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1359, "step": 14494 }, { "epoch": 21.699101796407184, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1344, "step": 14495 }, { "epoch": 21.70059880239521, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.133, "step": 14496 }, { "epoch": 21.702095808383234, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1375, "step": 14497 }, { "epoch": 21.70359281437126, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1297, "step": 14498 }, { "epoch": 21.70508982035928, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1338, "step": 14499 }, { "epoch": 21.706586826347305, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 14500 }, { "epoch": 21.70808383233533, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1324, "step": 14501 }, { "epoch": 21.709580838323355, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1376, "step": 14502 }, { "epoch": 21.711077844311376, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1316, "step": 14503 }, { "epoch": 21.7125748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1284, "step": 14504 }, { "epoch": 21.714071856287426, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1402, "step": 14505 }, { "epoch": 21.71556886227545, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1328, "step": 14506 }, { "epoch": 21.71706586826347, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.131, "step": 14507 }, { "epoch": 21.718562874251496, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1343, "step": 14508 }, { "epoch": 21.72005988023952, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.135, "step": 14509 }, { "epoch": 21.721556886227546, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1272, "step": 14510 }, { "epoch": 21.723053892215567, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1382, "step": 14511 }, { "epoch": 21.724550898203592, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.135, "step": 14512 }, { "epoch": 21.726047904191617, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1413, "step": 14513 }, { "epoch": 21.727544910179642, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1352, "step": 14514 }, { "epoch": 21.729041916167663, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1375, "step": 14515 }, { "epoch": 21.730538922155688, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.133, "step": 14516 }, { "epoch": 21.732035928143713, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1341, "step": 14517 }, { "epoch": 21.733532934131738, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1345, "step": 14518 }, { "epoch": 21.73502994011976, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.13, "step": 14519 }, { "epoch": 21.736526946107784, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1357, "step": 14520 }, { "epoch": 21.73802395209581, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1289, "step": 14521 }, { "epoch": 21.739520958083833, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1395, "step": 14522 }, { "epoch": 21.741017964071855, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1389, "step": 14523 }, { "epoch": 21.74251497005988, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1404, "step": 14524 }, { "epoch": 21.744011976047904, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1357, "step": 14525 }, { "epoch": 21.74550898203593, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.131, "step": 14526 }, { "epoch": 21.74700598802395, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.1349, "step": 14527 }, { "epoch": 21.748502994011975, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1339, "step": 14528 }, { "epoch": 21.75, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.1315, "step": 14529 }, { "epoch": 21.751497005988025, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1337, "step": 14530 }, { "epoch": 21.75299401197605, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1335, "step": 14531 }, { "epoch": 21.75449101796407, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1313, "step": 14532 }, { "epoch": 21.755988023952096, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1319, "step": 14533 }, { "epoch": 21.75748502994012, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14534 }, { "epoch": 21.758982035928145, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.134, "step": 14535 }, { "epoch": 21.760479041916167, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1378, "step": 14536 }, { "epoch": 21.76197604790419, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1358, "step": 14537 }, { "epoch": 21.763473053892216, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.132, "step": 14538 }, { "epoch": 21.76497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1345, "step": 14539 }, { "epoch": 21.766467065868262, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 14540 }, { "epoch": 21.767964071856287, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1354, "step": 14541 }, { "epoch": 21.769461077844312, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1389, "step": 14542 }, { "epoch": 21.770958083832337, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1262, "step": 14543 }, { "epoch": 21.772455089820358, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1306, "step": 14544 }, { "epoch": 21.773952095808383, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1316, "step": 14545 }, { "epoch": 21.775449101796408, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14546 }, { "epoch": 21.776946107784433, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 14547 }, { "epoch": 21.778443113772454, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 14548 }, { "epoch": 21.77994011976048, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1284, "step": 14549 }, { "epoch": 21.781437125748504, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1354, "step": 14550 }, { "epoch": 21.78293413173653, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1355, "step": 14551 }, { "epoch": 21.78443113772455, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1275, "step": 14552 }, { "epoch": 21.785928143712574, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1339, "step": 14553 }, { "epoch": 21.7874251497006, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1286, "step": 14554 }, { "epoch": 21.788922155688624, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 14555 }, { "epoch": 21.790419161676645, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.128, "step": 14556 }, { "epoch": 21.79191616766467, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1319, "step": 14557 }, { "epoch": 21.793413173652695, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1362, "step": 14558 }, { "epoch": 21.79491017964072, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1349, "step": 14559 }, { "epoch": 21.79640718562874, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1349, "step": 14560 }, { "epoch": 21.797904191616766, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1347, "step": 14561 }, { "epoch": 21.79940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1346, "step": 14562 }, { "epoch": 21.800898203592816, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1358, "step": 14563 }, { "epoch": 21.802395209580837, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1348, "step": 14564 }, { "epoch": 21.80389221556886, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.135, "step": 14565 }, { "epoch": 21.805389221556887, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1388, "step": 14566 }, { "epoch": 21.80688622754491, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1388, "step": 14567 }, { "epoch": 21.808383233532933, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14568 }, { "epoch": 21.809880239520957, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1347, "step": 14569 }, { "epoch": 21.811377245508982, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1404, "step": 14570 }, { "epoch": 21.812874251497007, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 14571 }, { "epoch": 21.81437125748503, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 14572 }, { "epoch": 21.815868263473053, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.14, "step": 14573 }, { "epoch": 21.817365269461078, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1372, "step": 14574 }, { "epoch": 21.818862275449103, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.13, "step": 14575 }, { "epoch": 21.820359281437124, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.141, "step": 14576 }, { "epoch": 21.82185628742515, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 14577 }, { "epoch": 21.823353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.134, "step": 14578 }, { "epoch": 21.8248502994012, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 14579 }, { "epoch": 21.82634730538922, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1368, "step": 14580 }, { "epoch": 21.827844311377245, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1365, "step": 14581 }, { "epoch": 21.82934131736527, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 14582 }, { "epoch": 21.830838323353294, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1357, "step": 14583 }, { "epoch": 21.83233532934132, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 14584 }, { "epoch": 21.83383233532934, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1277, "step": 14585 }, { "epoch": 21.835329341317365, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1407, "step": 14586 }, { "epoch": 21.83682634730539, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1402, "step": 14587 }, { "epoch": 21.83832335329341, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.139, "step": 14588 }, { "epoch": 21.839820359281436, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1278, "step": 14589 }, { "epoch": 21.84131736526946, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1354, "step": 14590 }, { "epoch": 21.842814371257486, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1465, "step": 14591 }, { "epoch": 21.84431137724551, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1364, "step": 14592 }, { "epoch": 21.845808383233532, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1333, "step": 14593 }, { "epoch": 21.847305389221557, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14594 }, { "epoch": 21.84880239520958, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.127, "step": 14595 }, { "epoch": 21.850299401197606, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1351, "step": 14596 }, { "epoch": 21.851796407185628, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1378, "step": 14597 }, { "epoch": 21.853293413173652, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1366, "step": 14598 }, { "epoch": 21.854790419161677, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.139, "step": 14599 }, { "epoch": 21.856287425149702, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1366, "step": 14600 }, { "epoch": 21.857784431137723, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1308, "step": 14601 }, { "epoch": 21.85928143712575, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14602 }, { "epoch": 21.860778443113773, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14603 }, { "epoch": 21.862275449101798, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1356, "step": 14604 }, { "epoch": 21.86377245508982, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1332, "step": 14605 }, { "epoch": 21.865269461077844, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1426, "step": 14606 }, { "epoch": 21.86676646706587, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1293, "step": 14607 }, { "epoch": 21.868263473053894, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1397, "step": 14608 }, { "epoch": 21.869760479041915, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1298, "step": 14609 }, { "epoch": 21.87125748502994, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1336, "step": 14610 }, { "epoch": 21.872754491017965, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1307, "step": 14611 }, { "epoch": 21.87425149700599, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14612 }, { "epoch": 21.87574850299401, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 14613 }, { "epoch": 21.877245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1356, "step": 14614 }, { "epoch": 21.87874251497006, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1358, "step": 14615 }, { "epoch": 21.880239520958085, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1323, "step": 14616 }, { "epoch": 21.881736526946106, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 14617 }, { "epoch": 21.88323353293413, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1369, "step": 14618 }, { "epoch": 21.884730538922156, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1366, "step": 14619 }, { "epoch": 21.88622754491018, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1371, "step": 14620 }, { "epoch": 21.887724550898202, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1233, "step": 14621 }, { "epoch": 21.889221556886227, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1315, "step": 14622 }, { "epoch": 21.89071856287425, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1276, "step": 14623 }, { "epoch": 21.892215568862277, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1401, "step": 14624 }, { "epoch": 21.893712574850298, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1349, "step": 14625 }, { "epoch": 21.895209580838323, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14626 }, { "epoch": 21.896706586826348, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14627 }, { "epoch": 21.898203592814372, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1266, "step": 14628 }, { "epoch": 21.899700598802394, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1324, "step": 14629 }, { "epoch": 21.90119760479042, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1376, "step": 14630 }, { "epoch": 21.902694610778443, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1338, "step": 14631 }, { "epoch": 21.904191616766468, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1342, "step": 14632 }, { "epoch": 21.90568862275449, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1352, "step": 14633 }, { "epoch": 21.907185628742514, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1338, "step": 14634 }, { "epoch": 21.90868263473054, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1363, "step": 14635 }, { "epoch": 21.910179640718564, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1287, "step": 14636 }, { "epoch": 21.91167664670659, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1347, "step": 14637 }, { "epoch": 21.91317365269461, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1353, "step": 14638 }, { "epoch": 21.914670658682635, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1348, "step": 14639 }, { "epoch": 21.91616766467066, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 14640 }, { "epoch": 21.91766467065868, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.133, "step": 14641 }, { "epoch": 21.919161676646706, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.138, "step": 14642 }, { "epoch": 21.92065868263473, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1336, "step": 14643 }, { "epoch": 21.922155688622755, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 14644 }, { "epoch": 21.92365269461078, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1291, "step": 14645 }, { "epoch": 21.9251497005988, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1312, "step": 14646 }, { "epoch": 21.926646706586826, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1341, "step": 14647 }, { "epoch": 21.92814371257485, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1385, "step": 14648 }, { "epoch": 21.929640718562876, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1343, "step": 14649 }, { "epoch": 21.931137724550897, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1343, "step": 14650 }, { "epoch": 21.932634730538922, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1416, "step": 14651 }, { "epoch": 21.934131736526947, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 14652 }, { "epoch": 21.93562874251497, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1334, "step": 14653 }, { "epoch": 21.937125748502993, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1369, "step": 14654 }, { "epoch": 21.938622754491018, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1319, "step": 14655 }, { "epoch": 21.940119760479043, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1308, "step": 14656 }, { "epoch": 21.941616766467067, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1277, "step": 14657 }, { "epoch": 21.94311377245509, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1428, "step": 14658 }, { "epoch": 21.944610778443113, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1409, "step": 14659 }, { "epoch": 21.94610778443114, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14660 }, { "epoch": 21.947604790419163, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.135, "step": 14661 }, { "epoch": 21.949101796407184, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1445, "step": 14662 }, { "epoch": 21.95059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1323, "step": 14663 }, { "epoch": 21.952095808383234, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1341, "step": 14664 }, { "epoch": 21.95359281437126, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1294, "step": 14665 }, { "epoch": 21.95508982035928, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1353, "step": 14666 }, { "epoch": 21.956586826347305, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1378, "step": 14667 }, { "epoch": 21.95808383233533, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1304, "step": 14668 }, { "epoch": 21.959580838323355, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1281, "step": 14669 }, { "epoch": 21.961077844311376, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1252, "step": 14670 }, { "epoch": 21.9625748502994, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1377, "step": 14671 }, { "epoch": 21.964071856287426, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1305, "step": 14672 }, { "epoch": 21.96556886227545, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1285, "step": 14673 }, { "epoch": 21.96706586826347, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1287, "step": 14674 }, { "epoch": 21.968562874251496, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1179, "step": 14675 }, { "epoch": 21.97005988023952, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1339, "step": 14676 }, { "epoch": 21.971556886227546, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1376, "step": 14677 }, { "epoch": 21.973053892215567, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 14678 }, { "epoch": 21.974550898203592, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1356, "step": 14679 }, { "epoch": 21.976047904191617, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1382, "step": 14680 }, { "epoch": 21.977544910179642, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1334, "step": 14681 }, { "epoch": 21.979041916167663, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1293, "step": 14682 }, { "epoch": 21.980538922155688, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1388, "step": 14683 }, { "epoch": 21.982035928143713, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1317, "step": 14684 }, { "epoch": 21.983532934131738, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1269, "step": 14685 }, { "epoch": 21.98502994011976, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 14686 }, { "epoch": 21.986526946107784, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1351, "step": 14687 }, { "epoch": 21.98802395209581, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1326, "step": 14688 }, { "epoch": 21.989520958083833, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1295, "step": 14689 }, { "epoch": 21.991017964071855, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1406, "step": 14690 }, { "epoch": 21.99251497005988, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1346, "step": 14691 }, { "epoch": 21.994011976047904, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1265, "step": 14692 }, { "epoch": 21.99550898203593, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1339, "step": 14693 }, { "epoch": 21.99700598802395, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1311, "step": 14694 }, { "epoch": 21.998502994011975, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.134, "step": 14695 }, { "epoch": 22.0, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.131, "step": 14696 }, { "epoch": 22.001497005988025, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1381, "step": 14697 }, { "epoch": 22.00299401197605, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1292, "step": 14698 }, { "epoch": 22.00449101796407, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1362, "step": 14699 }, { "epoch": 22.005988023952096, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1319, "step": 14700 }, { "epoch": 22.00748502994012, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1281, "step": 14701 }, { "epoch": 22.008982035928145, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.135, "step": 14702 }, { "epoch": 22.010479041916167, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1384, "step": 14703 }, { "epoch": 22.01197604790419, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.136, "step": 14704 }, { "epoch": 22.013473053892216, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1327, "step": 14705 }, { "epoch": 22.01497005988024, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1327, "step": 14706 }, { "epoch": 22.016467065868262, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1374, "step": 14707 }, { "epoch": 22.017964071856287, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1373, "step": 14708 }, { "epoch": 22.019461077844312, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 14709 }, { "epoch": 22.020958083832337, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1339, "step": 14710 }, { "epoch": 22.022455089820358, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1294, "step": 14711 }, { "epoch": 22.023952095808383, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1338, "step": 14712 }, { "epoch": 22.025449101796408, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 14713 }, { "epoch": 22.026946107784433, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1372, "step": 14714 }, { "epoch": 22.028443113772454, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1329, "step": 14715 }, { "epoch": 22.02994011976048, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1281, "step": 14716 }, { "epoch": 22.031437125748504, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1354, "step": 14717 }, { "epoch": 22.03293413173653, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1321, "step": 14718 }, { "epoch": 22.03443113772455, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1333, "step": 14719 }, { "epoch": 22.035928143712574, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1346, "step": 14720 }, { "epoch": 22.0374251497006, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1379, "step": 14721 }, { "epoch": 22.038922155688624, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1374, "step": 14722 }, { "epoch": 22.040419161676645, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.1362, "step": 14723 }, { "epoch": 22.04191616766467, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1407, "step": 14724 }, { "epoch": 22.043413173652695, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1317, "step": 14725 }, { "epoch": 22.04491017964072, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1349, "step": 14726 }, { "epoch": 22.04640718562874, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1325, "step": 14727 }, { "epoch": 22.047904191616766, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 14728 }, { "epoch": 22.04940119760479, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1345, "step": 14729 }, { "epoch": 22.050898203592816, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.13, "step": 14730 }, { "epoch": 22.052395209580837, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1297, "step": 14731 }, { "epoch": 22.05389221556886, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.137, "step": 14732 }, { "epoch": 22.055389221556887, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1276, "step": 14733 }, { "epoch": 22.05688622754491, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1305, "step": 14734 }, { "epoch": 22.058383233532933, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1372, "step": 14735 }, { "epoch": 22.059880239520957, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.133, "step": 14736 }, { "epoch": 22.061377245508982, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1303, "step": 14737 }, { "epoch": 22.062874251497007, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1334, "step": 14738 }, { "epoch": 22.06437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 14739 }, { "epoch": 22.065868263473053, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1319, "step": 14740 }, { "epoch": 22.067365269461078, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1332, "step": 14741 }, { "epoch": 22.068862275449103, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1372, "step": 14742 }, { "epoch": 22.070359281437124, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 14743 }, { "epoch": 22.07185628742515, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1268, "step": 14744 }, { "epoch": 22.073353293413174, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1401, "step": 14745 }, { "epoch": 22.0748502994012, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14746 }, { "epoch": 22.07634730538922, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1331, "step": 14747 }, { "epoch": 22.077844311377245, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14748 }, { "epoch": 22.07934131736527, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1337, "step": 14749 }, { "epoch": 22.080838323353294, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1357, "step": 14750 }, { "epoch": 22.082335329341316, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1292, "step": 14751 }, { "epoch": 22.08383233532934, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 14752 }, { "epoch": 22.085329341317365, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1316, "step": 14753 }, { "epoch": 22.08682634730539, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14754 }, { "epoch": 22.088323353293415, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1406, "step": 14755 }, { "epoch": 22.089820359281436, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1357, "step": 14756 }, { "epoch": 22.09131736526946, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1341, "step": 14757 }, { "epoch": 22.092814371257486, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14758 }, { "epoch": 22.09431137724551, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1342, "step": 14759 }, { "epoch": 22.095808383233532, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1336, "step": 14760 }, { "epoch": 22.097305389221557, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1298, "step": 14761 }, { "epoch": 22.09880239520958, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.13, "step": 14762 }, { "epoch": 22.100299401197606, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1267, "step": 14763 }, { "epoch": 22.101796407185628, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1359, "step": 14764 }, { "epoch": 22.103293413173652, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.135, "step": 14765 }, { "epoch": 22.104790419161677, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1337, "step": 14766 }, { "epoch": 22.106287425149702, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1364, "step": 14767 }, { "epoch": 22.107784431137723, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.136, "step": 14768 }, { "epoch": 22.10928143712575, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1334, "step": 14769 }, { "epoch": 22.110778443113773, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 14770 }, { "epoch": 22.112275449101798, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1346, "step": 14771 }, { "epoch": 22.11377245508982, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1386, "step": 14772 }, { "epoch": 22.115269461077844, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 14773 }, { "epoch": 22.11676646706587, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1419, "step": 14774 }, { "epoch": 22.118263473053894, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1324, "step": 14775 }, { "epoch": 22.119760479041915, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.137, "step": 14776 }, { "epoch": 22.12125748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1376, "step": 14777 }, { "epoch": 22.122754491017965, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 14778 }, { "epoch": 22.12425149700599, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1356, "step": 14779 }, { "epoch": 22.12574850299401, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1317, "step": 14780 }, { "epoch": 22.127245508982035, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1345, "step": 14781 }, { "epoch": 22.12874251497006, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1234, "step": 14782 }, { "epoch": 22.130239520958085, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14783 }, { "epoch": 22.131736526946106, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 14784 }, { "epoch": 22.13323353293413, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1376, "step": 14785 }, { "epoch": 22.134730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1341, "step": 14786 }, { "epoch": 22.13622754491018, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1337, "step": 14787 }, { "epoch": 22.137724550898202, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1371, "step": 14788 }, { "epoch": 22.139221556886227, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 14789 }, { "epoch": 22.14071856287425, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1354, "step": 14790 }, { "epoch": 22.142215568862277, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.134, "step": 14791 }, { "epoch": 22.143712574850298, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1307, "step": 14792 }, { "epoch": 22.145209580838323, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1326, "step": 14793 }, { "epoch": 22.146706586826348, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1342, "step": 14794 }, { "epoch": 22.148203592814372, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1354, "step": 14795 }, { "epoch": 22.149700598802394, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.137, "step": 14796 }, { "epoch": 22.15119760479042, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.133, "step": 14797 }, { "epoch": 22.152694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 14798 }, { "epoch": 22.154191616766468, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.133, "step": 14799 }, { "epoch": 22.15568862275449, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.13, "step": 14800 }, { "epoch": 22.157185628742514, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.134, "step": 14801 }, { "epoch": 22.15868263473054, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1322, "step": 14802 }, { "epoch": 22.160179640718564, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1324, "step": 14803 }, { "epoch": 22.161676646706585, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1382, "step": 14804 }, { "epoch": 22.16317365269461, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1348, "step": 14805 }, { "epoch": 22.164670658682635, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14806 }, { "epoch": 22.16616766467066, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1315, "step": 14807 }, { "epoch": 22.16766467065868, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14808 }, { "epoch": 22.169161676646706, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1335, "step": 14809 }, { "epoch": 22.17065868263473, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 14810 }, { "epoch": 22.172155688622755, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1331, "step": 14811 }, { "epoch": 22.17365269461078, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1392, "step": 14812 }, { "epoch": 22.1751497005988, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1318, "step": 14813 }, { "epoch": 22.176646706586826, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.132, "step": 14814 }, { "epoch": 22.17814371257485, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1358, "step": 14815 }, { "epoch": 22.179640718562876, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1357, "step": 14816 }, { "epoch": 22.181137724550897, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1431, "step": 14817 }, { "epoch": 22.182634730538922, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1319, "step": 14818 }, { "epoch": 22.184131736526947, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1376, "step": 14819 }, { "epoch": 22.18562874251497, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1358, "step": 14820 }, { "epoch": 22.187125748502993, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1345, "step": 14821 }, { "epoch": 22.188622754491018, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1351, "step": 14822 }, { "epoch": 22.190119760479043, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1357, "step": 14823 }, { "epoch": 22.191616766467067, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 14824 }, { "epoch": 22.19311377245509, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1328, "step": 14825 }, { "epoch": 22.194610778443113, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1384, "step": 14826 }, { "epoch": 22.19610778443114, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1311, "step": 14827 }, { "epoch": 22.197604790419163, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1288, "step": 14828 }, { "epoch": 22.199101796407184, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 14829 }, { "epoch": 22.20059880239521, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1354, "step": 14830 }, { "epoch": 22.202095808383234, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1257, "step": 14831 }, { "epoch": 22.20359281437126, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.14, "step": 14832 }, { "epoch": 22.20508982035928, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1334, "step": 14833 }, { "epoch": 22.206586826347305, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1257, "step": 14834 }, { "epoch": 22.20808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1382, "step": 14835 }, { "epoch": 22.209580838323355, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.131, "step": 14836 }, { "epoch": 22.211077844311376, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1276, "step": 14837 }, { "epoch": 22.2125748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 14838 }, { "epoch": 22.214071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1333, "step": 14839 }, { "epoch": 22.21556886227545, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1373, "step": 14840 }, { "epoch": 22.21706586826347, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1373, "step": 14841 }, { "epoch": 22.218562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1308, "step": 14842 }, { "epoch": 22.22005988023952, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1432, "step": 14843 }, { "epoch": 22.221556886227546, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.131, "step": 14844 }, { "epoch": 22.223053892215567, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1313, "step": 14845 }, { "epoch": 22.224550898203592, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1339, "step": 14846 }, { "epoch": 22.226047904191617, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1393, "step": 14847 }, { "epoch": 22.227544910179642, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.132, "step": 14848 }, { "epoch": 22.229041916167663, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1287, "step": 14849 }, { "epoch": 22.230538922155688, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.135, "step": 14850 }, { "epoch": 22.232035928143713, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 14851 }, { "epoch": 22.233532934131738, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14852 }, { "epoch": 22.23502994011976, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1336, "step": 14853 }, { "epoch": 22.236526946107784, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1389, "step": 14854 }, { "epoch": 22.23802395209581, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1339, "step": 14855 }, { "epoch": 22.239520958083833, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1227, "step": 14856 }, { "epoch": 22.241017964071855, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 14857 }, { "epoch": 22.24251497005988, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1288, "step": 14858 }, { "epoch": 22.244011976047904, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1385, "step": 14859 }, { "epoch": 22.24550898203593, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1312, "step": 14860 }, { "epoch": 22.24700598802395, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1311, "step": 14861 }, { "epoch": 22.248502994011975, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1314, "step": 14862 }, { "epoch": 22.25, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14863 }, { "epoch": 22.251497005988025, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.129, "step": 14864 }, { "epoch": 22.25299401197605, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1339, "step": 14865 }, { "epoch": 22.25449101796407, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14866 }, { "epoch": 22.255988023952096, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14867 }, { "epoch": 22.25748502994012, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1315, "step": 14868 }, { "epoch": 22.258982035928145, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1254, "step": 14869 }, { "epoch": 22.260479041916167, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1289, "step": 14870 }, { "epoch": 22.26197604790419, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1331, "step": 14871 }, { "epoch": 22.263473053892216, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1414, "step": 14872 }, { "epoch": 22.26497005988024, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1309, "step": 14873 }, { "epoch": 22.266467065868262, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1294, "step": 14874 }, { "epoch": 22.267964071856287, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 14875 }, { "epoch": 22.269461077844312, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1288, "step": 14876 }, { "epoch": 22.270958083832337, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 14877 }, { "epoch": 22.272455089820358, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.141, "step": 14878 }, { "epoch": 22.273952095808383, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1333, "step": 14879 }, { "epoch": 22.275449101796408, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1336, "step": 14880 }, { "epoch": 22.276946107784433, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1353, "step": 14881 }, { "epoch": 22.278443113772454, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1316, "step": 14882 }, { "epoch": 22.27994011976048, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.138, "step": 14883 }, { "epoch": 22.281437125748504, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14884 }, { "epoch": 22.28293413173653, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1353, "step": 14885 }, { "epoch": 22.28443113772455, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1296, "step": 14886 }, { "epoch": 22.285928143712574, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.131, "step": 14887 }, { "epoch": 22.2874251497006, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 14888 }, { "epoch": 22.288922155688624, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1303, "step": 14889 }, { "epoch": 22.290419161676645, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1346, "step": 14890 }, { "epoch": 22.29191616766467, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1322, "step": 14891 }, { "epoch": 22.293413173652695, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1333, "step": 14892 }, { "epoch": 22.29491017964072, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1396, "step": 14893 }, { "epoch": 22.29640718562874, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.137, "step": 14894 }, { "epoch": 22.297904191616766, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1353, "step": 14895 }, { "epoch": 22.29940119760479, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1344, "step": 14896 }, { "epoch": 22.300898203592816, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1386, "step": 14897 }, { "epoch": 22.302395209580837, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14898 }, { "epoch": 22.30389221556886, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1266, "step": 14899 }, { "epoch": 22.305389221556887, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.133, "step": 14900 }, { "epoch": 22.30688622754491, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1362, "step": 14901 }, { "epoch": 22.308383233532933, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1353, "step": 14902 }, { "epoch": 22.309880239520957, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1341, "step": 14903 }, { "epoch": 22.311377245508982, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.143, "step": 14904 }, { "epoch": 22.312874251497007, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 14905 }, { "epoch": 22.31437125748503, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1407, "step": 14906 }, { "epoch": 22.315868263473053, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1279, "step": 14907 }, { "epoch": 22.317365269461078, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1299, "step": 14908 }, { "epoch": 22.318862275449103, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1282, "step": 14909 }, { "epoch": 22.320359281437124, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.134, "step": 14910 }, { "epoch": 22.32185628742515, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1299, "step": 14911 }, { "epoch": 22.323353293413174, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1371, "step": 14912 }, { "epoch": 22.3248502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.132, "step": 14913 }, { "epoch": 22.32634730538922, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1345, "step": 14914 }, { "epoch": 22.327844311377245, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1284, "step": 14915 }, { "epoch": 22.32934131736527, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1321, "step": 14916 }, { "epoch": 22.330838323353294, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1369, "step": 14917 }, { "epoch": 22.33233532934132, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 14918 }, { "epoch": 22.33383233532934, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1353, "step": 14919 }, { "epoch": 22.335329341317365, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1292, "step": 14920 }, { "epoch": 22.33682634730539, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.136, "step": 14921 }, { "epoch": 22.338323353293415, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1377, "step": 14922 }, { "epoch": 22.339820359281436, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1284, "step": 14923 }, { "epoch": 22.34131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1272, "step": 14924 }, { "epoch": 22.342814371257486, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1268, "step": 14925 }, { "epoch": 22.34431137724551, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1334, "step": 14926 }, { "epoch": 22.345808383233532, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1423, "step": 14927 }, { "epoch": 22.347305389221557, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1349, "step": 14928 }, { "epoch": 22.34880239520958, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1323, "step": 14929 }, { "epoch": 22.350299401197606, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1289, "step": 14930 }, { "epoch": 22.351796407185628, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1347, "step": 14931 }, { "epoch": 22.353293413173652, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1303, "step": 14932 }, { "epoch": 22.354790419161677, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1316, "step": 14933 }, { "epoch": 22.356287425149702, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1371, "step": 14934 }, { "epoch": 22.357784431137723, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1305, "step": 14935 }, { "epoch": 22.35928143712575, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1354, "step": 14936 }, { "epoch": 22.360778443113773, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14937 }, { "epoch": 22.362275449101798, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1303, "step": 14938 }, { "epoch": 22.36377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1335, "step": 14939 }, { "epoch": 22.365269461077844, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1316, "step": 14940 }, { "epoch": 22.36676646706587, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1389, "step": 14941 }, { "epoch": 22.368263473053894, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1366, "step": 14942 }, { "epoch": 22.369760479041915, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1315, "step": 14943 }, { "epoch": 22.37125748502994, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.143, "step": 14944 }, { "epoch": 22.372754491017965, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14945 }, { "epoch": 22.37425149700599, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1385, "step": 14946 }, { "epoch": 22.37574850299401, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1361, "step": 14947 }, { "epoch": 22.377245508982035, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1333, "step": 14948 }, { "epoch": 22.37874251497006, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.13, "step": 14949 }, { "epoch": 22.380239520958085, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1313, "step": 14950 }, { "epoch": 22.381736526946106, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1345, "step": 14951 }, { "epoch": 22.38323353293413, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1382, "step": 14952 }, { "epoch": 22.384730538922156, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1256, "step": 14953 }, { "epoch": 22.38622754491018, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1338, "step": 14954 }, { "epoch": 22.387724550898202, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1395, "step": 14955 }, { "epoch": 22.389221556886227, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1312, "step": 14956 }, { "epoch": 22.39071856287425, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1311, "step": 14957 }, { "epoch": 22.392215568862277, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1347, "step": 14958 }, { "epoch": 22.393712574850298, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1306, "step": 14959 }, { "epoch": 22.395209580838323, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1338, "step": 14960 }, { "epoch": 22.396706586826348, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1336, "step": 14961 }, { "epoch": 22.398203592814372, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1315, "step": 14962 }, { "epoch": 22.399700598802394, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14963 }, { "epoch": 22.40119760479042, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 14964 }, { "epoch": 22.402694610778443, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1349, "step": 14965 }, { "epoch": 22.404191616766468, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1368, "step": 14966 }, { "epoch": 22.40568862275449, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1294, "step": 14967 }, { "epoch": 22.407185628742514, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1299, "step": 14968 }, { "epoch": 22.40868263473054, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1369, "step": 14969 }, { "epoch": 22.410179640718564, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1342, "step": 14970 }, { "epoch": 22.411676646706585, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1343, "step": 14971 }, { "epoch": 22.41317365269461, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1352, "step": 14972 }, { "epoch": 22.414670658682635, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 14973 }, { "epoch": 22.41616766467066, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.132, "step": 14974 }, { "epoch": 22.41766467065868, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1312, "step": 14975 }, { "epoch": 22.419161676646706, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1257, "step": 14976 }, { "epoch": 22.42065868263473, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.135, "step": 14977 }, { "epoch": 22.422155688622755, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1213, "step": 14978 }, { "epoch": 22.42365269461078, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 14979 }, { "epoch": 22.4251497005988, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1331, "step": 14980 }, { "epoch": 22.426646706586826, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.136, "step": 14981 }, { "epoch": 22.42814371257485, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1358, "step": 14982 }, { "epoch": 22.429640718562876, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1344, "step": 14983 }, { "epoch": 22.431137724550897, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1277, "step": 14984 }, { "epoch": 22.432634730538922, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1361, "step": 14985 }, { "epoch": 22.434131736526947, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1282, "step": 14986 }, { "epoch": 22.43562874251497, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1309, "step": 14987 }, { "epoch": 22.437125748502993, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1381, "step": 14988 }, { "epoch": 22.438622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1325, "step": 14989 }, { "epoch": 22.440119760479043, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1343, "step": 14990 }, { "epoch": 22.441616766467067, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1356, "step": 14991 }, { "epoch": 22.44311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 14992 }, { "epoch": 22.444610778443113, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1349, "step": 14993 }, { "epoch": 22.44610778443114, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1276, "step": 14994 }, { "epoch": 22.447604790419163, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1305, "step": 14995 }, { "epoch": 22.449101796407184, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1375, "step": 14996 }, { "epoch": 22.45059880239521, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.131, "step": 14997 }, { "epoch": 22.452095808383234, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1289, "step": 14998 }, { "epoch": 22.45359281437126, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1377, "step": 14999 }, { "epoch": 22.45508982035928, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1386, "step": 15000 }, { "epoch": 22.456586826347305, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1304, "step": 15001 }, { "epoch": 22.45808383233533, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.132, "step": 15002 }, { "epoch": 22.459580838323355, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1414, "step": 15003 }, { "epoch": 22.461077844311376, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1383, "step": 15004 }, { "epoch": 22.4625748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15005 }, { "epoch": 22.464071856287426, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1351, "step": 15006 }, { "epoch": 22.46556886227545, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1382, "step": 15007 }, { "epoch": 22.46706586826347, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1358, "step": 15008 }, { "epoch": 22.468562874251496, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1333, "step": 15009 }, { "epoch": 22.47005988023952, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1362, "step": 15010 }, { "epoch": 22.471556886227546, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1388, "step": 15011 }, { "epoch": 22.473053892215567, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15012 }, { "epoch": 22.474550898203592, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1287, "step": 15013 }, { "epoch": 22.476047904191617, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1321, "step": 15014 }, { "epoch": 22.477544910179642, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1344, "step": 15015 }, { "epoch": 22.479041916167663, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1379, "step": 15016 }, { "epoch": 22.480538922155688, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.133, "step": 15017 }, { "epoch": 22.482035928143713, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1342, "step": 15018 }, { "epoch": 22.483532934131738, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1268, "step": 15019 }, { "epoch": 22.48502994011976, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1276, "step": 15020 }, { "epoch": 22.486526946107784, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1295, "step": 15021 }, { "epoch": 22.48802395209581, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1324, "step": 15022 }, { "epoch": 22.489520958083833, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1313, "step": 15023 }, { "epoch": 22.491017964071855, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1324, "step": 15024 }, { "epoch": 22.49251497005988, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1261, "step": 15025 }, { "epoch": 22.494011976047904, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1343, "step": 15026 }, { "epoch": 22.49550898203593, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.135, "step": 15027 }, { "epoch": 22.49700598802395, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1371, "step": 15028 }, { "epoch": 22.498502994011975, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1335, "step": 15029 }, { "epoch": 22.5, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1396, "step": 15030 }, { "epoch": 22.501497005988025, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1307, "step": 15031 }, { "epoch": 22.50299401197605, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15032 }, { "epoch": 22.50449101796407, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1267, "step": 15033 }, { "epoch": 22.505988023952096, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1371, "step": 15034 }, { "epoch": 22.50748502994012, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1319, "step": 15035 }, { "epoch": 22.508982035928145, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1303, "step": 15036 }, { "epoch": 22.510479041916167, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1333, "step": 15037 }, { "epoch": 22.51197604790419, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1319, "step": 15038 }, { "epoch": 22.513473053892216, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1301, "step": 15039 }, { "epoch": 22.51497005988024, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1304, "step": 15040 }, { "epoch": 22.516467065868262, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1317, "step": 15041 }, { "epoch": 22.517964071856287, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1315, "step": 15042 }, { "epoch": 22.519461077844312, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1301, "step": 15043 }, { "epoch": 22.520958083832337, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.129, "step": 15044 }, { "epoch": 22.522455089820358, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1308, "step": 15045 }, { "epoch": 22.523952095808383, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 15046 }, { "epoch": 22.525449101796408, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1284, "step": 15047 }, { "epoch": 22.526946107784433, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1301, "step": 15048 }, { "epoch": 22.528443113772454, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1295, "step": 15049 }, { "epoch": 22.52994011976048, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1306, "step": 15050 }, { "epoch": 22.531437125748504, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1297, "step": 15051 }, { "epoch": 22.53293413173653, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 15052 }, { "epoch": 22.53443113772455, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1396, "step": 15053 }, { "epoch": 22.535928143712574, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1393, "step": 15054 }, { "epoch": 22.5374251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1286, "step": 15055 }, { "epoch": 22.538922155688624, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1337, "step": 15056 }, { "epoch": 22.540419161676645, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1321, "step": 15057 }, { "epoch": 22.54191616766467, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1376, "step": 15058 }, { "epoch": 22.543413173652695, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1301, "step": 15059 }, { "epoch": 22.54491017964072, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1344, "step": 15060 }, { "epoch": 22.54640718562874, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1278, "step": 15061 }, { "epoch": 22.547904191616766, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1399, "step": 15062 }, { "epoch": 22.54940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 15063 }, { "epoch": 22.550898203592816, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1341, "step": 15064 }, { "epoch": 22.552395209580837, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1372, "step": 15065 }, { "epoch": 22.55389221556886, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1346, "step": 15066 }, { "epoch": 22.555389221556887, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1345, "step": 15067 }, { "epoch": 22.55688622754491, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1335, "step": 15068 }, { "epoch": 22.558383233532933, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1287, "step": 15069 }, { "epoch": 22.559880239520957, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.131, "step": 15070 }, { "epoch": 22.561377245508982, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.133, "step": 15071 }, { "epoch": 22.562874251497007, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1308, "step": 15072 }, { "epoch": 22.56437125748503, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1352, "step": 15073 }, { "epoch": 22.565868263473053, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1331, "step": 15074 }, { "epoch": 22.567365269461078, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1362, "step": 15075 }, { "epoch": 22.568862275449103, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1297, "step": 15076 }, { "epoch": 22.570359281437124, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1358, "step": 15077 }, { "epoch": 22.57185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15078 }, { "epoch": 22.573353293413174, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1312, "step": 15079 }, { "epoch": 22.5748502994012, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.132, "step": 15080 }, { "epoch": 22.57634730538922, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1371, "step": 15081 }, { "epoch": 22.577844311377245, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1391, "step": 15082 }, { "epoch": 22.57934131736527, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.133, "step": 15083 }, { "epoch": 22.580838323353294, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 15084 }, { "epoch": 22.58233532934132, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 15085 }, { "epoch": 22.58383233532934, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1297, "step": 15086 }, { "epoch": 22.585329341317365, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1308, "step": 15087 }, { "epoch": 22.58682634730539, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1337, "step": 15088 }, { "epoch": 22.58832335329341, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1356, "step": 15089 }, { "epoch": 22.589820359281436, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.131, "step": 15090 }, { "epoch": 22.59131736526946, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1357, "step": 15091 }, { "epoch": 22.592814371257486, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1388, "step": 15092 }, { "epoch": 22.59431137724551, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1339, "step": 15093 }, { "epoch": 22.595808383233532, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1239, "step": 15094 }, { "epoch": 22.597305389221557, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1298, "step": 15095 }, { "epoch": 22.59880239520958, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.13, "step": 15096 }, { "epoch": 22.600299401197606, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 15097 }, { "epoch": 22.601796407185628, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1314, "step": 15098 }, { "epoch": 22.603293413173652, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.137, "step": 15099 }, { "epoch": 22.604790419161677, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1411, "step": 15100 }, { "epoch": 22.606287425149702, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1265, "step": 15101 }, { "epoch": 22.607784431137723, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1304, "step": 15102 }, { "epoch": 22.60928143712575, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1364, "step": 15103 }, { "epoch": 22.610778443113773, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 15104 }, { "epoch": 22.612275449101798, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1289, "step": 15105 }, { "epoch": 22.61377245508982, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1304, "step": 15106 }, { "epoch": 22.615269461077844, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1347, "step": 15107 }, { "epoch": 22.61676646706587, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1295, "step": 15108 }, { "epoch": 22.618263473053894, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.1372, "step": 15109 }, { "epoch": 22.619760479041915, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1321, "step": 15110 }, { "epoch": 22.62125748502994, "grad_norm": 0.20703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 15111 }, { "epoch": 22.622754491017965, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1314, "step": 15112 }, { "epoch": 22.62425149700599, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 15113 }, { "epoch": 22.62574850299401, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1398, "step": 15114 }, { "epoch": 22.627245508982035, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1292, "step": 15115 }, { "epoch": 22.62874251497006, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 15116 }, { "epoch": 22.630239520958085, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1358, "step": 15117 }, { "epoch": 22.631736526946106, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 15118 }, { "epoch": 22.63323353293413, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1365, "step": 15119 }, { "epoch": 22.634730538922156, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1304, "step": 15120 }, { "epoch": 22.63622754491018, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1322, "step": 15121 }, { "epoch": 22.637724550898202, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1316, "step": 15122 }, { "epoch": 22.639221556886227, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.128, "step": 15123 }, { "epoch": 22.64071856287425, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1298, "step": 15124 }, { "epoch": 22.642215568862277, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.135, "step": 15125 }, { "epoch": 22.643712574850298, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1294, "step": 15126 }, { "epoch": 22.645209580838323, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.134, "step": 15127 }, { "epoch": 22.646706586826348, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1294, "step": 15128 }, { "epoch": 22.648203592814372, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1362, "step": 15129 }, { "epoch": 22.649700598802394, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1249, "step": 15130 }, { "epoch": 22.65119760479042, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1338, "step": 15131 }, { "epoch": 22.652694610778443, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1371, "step": 15132 }, { "epoch": 22.654191616766468, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1307, "step": 15133 }, { "epoch": 22.65568862275449, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1312, "step": 15134 }, { "epoch": 22.657185628742514, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15135 }, { "epoch": 22.65868263473054, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1349, "step": 15136 }, { "epoch": 22.660179640718564, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1373, "step": 15137 }, { "epoch": 22.66167664670659, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1367, "step": 15138 }, { "epoch": 22.66317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1316, "step": 15139 }, { "epoch": 22.664670658682635, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1355, "step": 15140 }, { "epoch": 22.66616766467066, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1353, "step": 15141 }, { "epoch": 22.66766467065868, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1303, "step": 15142 }, { "epoch": 22.669161676646706, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1324, "step": 15143 }, { "epoch": 22.67065868263473, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1345, "step": 15144 }, { "epoch": 22.672155688622755, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1373, "step": 15145 }, { "epoch": 22.67365269461078, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1373, "step": 15146 }, { "epoch": 22.6751497005988, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1265, "step": 15147 }, { "epoch": 22.676646706586826, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 15148 }, { "epoch": 22.67814371257485, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15149 }, { "epoch": 22.679640718562876, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1347, "step": 15150 }, { "epoch": 22.681137724550897, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1272, "step": 15151 }, { "epoch": 22.682634730538922, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1367, "step": 15152 }, { "epoch": 22.684131736526947, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1349, "step": 15153 }, { "epoch": 22.68562874251497, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1339, "step": 15154 }, { "epoch": 22.687125748502993, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15155 }, { "epoch": 22.688622754491018, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1295, "step": 15156 }, { "epoch": 22.690119760479043, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1369, "step": 15157 }, { "epoch": 22.691616766467067, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1335, "step": 15158 }, { "epoch": 22.69311377245509, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1356, "step": 15159 }, { "epoch": 22.694610778443113, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1323, "step": 15160 }, { "epoch": 22.69610778443114, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 15161 }, { "epoch": 22.697604790419163, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 15162 }, { "epoch": 22.699101796407184, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1306, "step": 15163 }, { "epoch": 22.70059880239521, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1345, "step": 15164 }, { "epoch": 22.702095808383234, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1307, "step": 15165 }, { "epoch": 22.70359281437126, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1383, "step": 15166 }, { "epoch": 22.70508982035928, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1319, "step": 15167 }, { "epoch": 22.706586826347305, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1225, "step": 15168 }, { "epoch": 22.70808383233533, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1372, "step": 15169 }, { "epoch": 22.709580838323355, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1326, "step": 15170 }, { "epoch": 22.711077844311376, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1356, "step": 15171 }, { "epoch": 22.7125748502994, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1375, "step": 15172 }, { "epoch": 22.714071856287426, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1329, "step": 15173 }, { "epoch": 22.71556886227545, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 15174 }, { "epoch": 22.71706586826347, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 15175 }, { "epoch": 22.718562874251496, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 15176 }, { "epoch": 22.72005988023952, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1336, "step": 15177 }, { "epoch": 22.721556886227546, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1331, "step": 15178 }, { "epoch": 22.723053892215567, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1303, "step": 15179 }, { "epoch": 22.724550898203592, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1353, "step": 15180 }, { "epoch": 22.726047904191617, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1279, "step": 15181 }, { "epoch": 22.727544910179642, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1366, "step": 15182 }, { "epoch": 22.729041916167663, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 15183 }, { "epoch": 22.730538922155688, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1352, "step": 15184 }, { "epoch": 22.732035928143713, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.13, "step": 15185 }, { "epoch": 22.733532934131738, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 15186 }, { "epoch": 22.73502994011976, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1379, "step": 15187 }, { "epoch": 22.736526946107784, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 15188 }, { "epoch": 22.73802395209581, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 15189 }, { "epoch": 22.739520958083833, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1419, "step": 15190 }, { "epoch": 22.741017964071855, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 15191 }, { "epoch": 22.74251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15192 }, { "epoch": 22.744011976047904, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1265, "step": 15193 }, { "epoch": 22.74550898203593, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1332, "step": 15194 }, { "epoch": 22.74700598802395, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 15195 }, { "epoch": 22.748502994011975, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 15196 }, { "epoch": 22.75, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1366, "step": 15197 }, { "epoch": 22.751497005988025, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1204, "step": 15198 }, { "epoch": 22.75299401197605, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1288, "step": 15199 }, { "epoch": 22.75449101796407, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1344, "step": 15200 }, { "epoch": 22.755988023952096, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1341, "step": 15201 }, { "epoch": 22.75748502994012, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.131, "step": 15202 }, { "epoch": 22.758982035928145, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1325, "step": 15203 }, { "epoch": 22.760479041916167, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1377, "step": 15204 }, { "epoch": 22.76197604790419, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 15205 }, { "epoch": 22.763473053892216, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 15206 }, { "epoch": 22.76497005988024, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1393, "step": 15207 }, { "epoch": 22.766467065868262, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1355, "step": 15208 }, { "epoch": 22.767964071856287, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 15209 }, { "epoch": 22.769461077844312, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1337, "step": 15210 }, { "epoch": 22.770958083832337, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.13, "step": 15211 }, { "epoch": 22.772455089820358, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1268, "step": 15212 }, { "epoch": 22.773952095808383, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.135, "step": 15213 }, { "epoch": 22.775449101796408, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1414, "step": 15214 }, { "epoch": 22.776946107784433, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1282, "step": 15215 }, { "epoch": 22.778443113772454, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1328, "step": 15216 }, { "epoch": 22.77994011976048, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 15217 }, { "epoch": 22.781437125748504, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 15218 }, { "epoch": 22.78293413173653, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 15219 }, { "epoch": 22.78443113772455, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.131, "step": 15220 }, { "epoch": 22.785928143712574, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1283, "step": 15221 }, { "epoch": 22.7874251497006, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 15222 }, { "epoch": 22.788922155688624, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1325, "step": 15223 }, { "epoch": 22.790419161676645, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1283, "step": 15224 }, { "epoch": 22.79191616766467, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1356, "step": 15225 }, { "epoch": 22.793413173652695, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1294, "step": 15226 }, { "epoch": 22.79491017964072, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1373, "step": 15227 }, { "epoch": 22.79640718562874, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1363, "step": 15228 }, { "epoch": 22.797904191616766, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1316, "step": 15229 }, { "epoch": 22.79940119760479, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1385, "step": 15230 }, { "epoch": 22.800898203592816, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.138, "step": 15231 }, { "epoch": 22.802395209580837, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1359, "step": 15232 }, { "epoch": 22.80389221556886, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 15233 }, { "epoch": 22.805389221556887, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1328, "step": 15234 }, { "epoch": 22.80688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 15235 }, { "epoch": 22.808383233532933, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1355, "step": 15236 }, { "epoch": 22.809880239520957, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1333, "step": 15237 }, { "epoch": 22.811377245508982, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1242, "step": 15238 }, { "epoch": 22.812874251497007, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1356, "step": 15239 }, { "epoch": 22.81437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1332, "step": 15240 }, { "epoch": 22.815868263473053, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 15241 }, { "epoch": 22.817365269461078, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1333, "step": 15242 }, { "epoch": 22.818862275449103, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1316, "step": 15243 }, { "epoch": 22.820359281437124, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1326, "step": 15244 }, { "epoch": 22.82185628742515, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1346, "step": 15245 }, { "epoch": 22.823353293413174, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1316, "step": 15246 }, { "epoch": 22.8248502994012, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1382, "step": 15247 }, { "epoch": 22.82634730538922, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 15248 }, { "epoch": 22.827844311377245, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1283, "step": 15249 }, { "epoch": 22.82934131736527, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 15250 }, { "epoch": 22.830838323353294, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1351, "step": 15251 }, { "epoch": 22.83233532934132, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 15252 }, { "epoch": 22.83383233532934, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1382, "step": 15253 }, { "epoch": 22.835329341317365, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1364, "step": 15254 }, { "epoch": 22.83682634730539, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1277, "step": 15255 }, { "epoch": 22.83832335329341, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1369, "step": 15256 }, { "epoch": 22.839820359281436, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1416, "step": 15257 }, { "epoch": 22.84131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1338, "step": 15258 }, { "epoch": 22.842814371257486, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1281, "step": 15259 }, { "epoch": 22.84431137724551, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1328, "step": 15260 }, { "epoch": 22.845808383233532, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1324, "step": 15261 }, { "epoch": 22.847305389221557, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1318, "step": 15262 }, { "epoch": 22.84880239520958, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1262, "step": 15263 }, { "epoch": 22.850299401197606, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1317, "step": 15264 }, { "epoch": 22.851796407185628, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1283, "step": 15265 }, { "epoch": 22.853293413173652, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1346, "step": 15266 }, { "epoch": 22.854790419161677, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1366, "step": 15267 }, { "epoch": 22.856287425149702, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 15268 }, { "epoch": 22.857784431137723, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 15269 }, { "epoch": 22.85928143712575, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1362, "step": 15270 }, { "epoch": 22.860778443113773, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1384, "step": 15271 }, { "epoch": 22.862275449101798, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1374, "step": 15272 }, { "epoch": 22.86377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1349, "step": 15273 }, { "epoch": 22.865269461077844, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.131, "step": 15274 }, { "epoch": 22.86676646706587, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.13, "step": 15275 }, { "epoch": 22.868263473053894, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1335, "step": 15276 }, { "epoch": 22.869760479041915, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1325, "step": 15277 }, { "epoch": 22.87125748502994, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1293, "step": 15278 }, { "epoch": 22.872754491017965, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1383, "step": 15279 }, { "epoch": 22.87425149700599, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1269, "step": 15280 }, { "epoch": 22.87574850299401, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15281 }, { "epoch": 22.877245508982035, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1354, "step": 15282 }, { "epoch": 22.87874251497006, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.133, "step": 15283 }, { "epoch": 22.880239520958085, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1376, "step": 15284 }, { "epoch": 22.881736526946106, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1333, "step": 15285 }, { "epoch": 22.88323353293413, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1351, "step": 15286 }, { "epoch": 22.884730538922156, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1336, "step": 15287 }, { "epoch": 22.88622754491018, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1314, "step": 15288 }, { "epoch": 22.887724550898202, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 15289 }, { "epoch": 22.889221556886227, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1338, "step": 15290 }, { "epoch": 22.89071856287425, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1329, "step": 15291 }, { "epoch": 22.892215568862277, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 15292 }, { "epoch": 22.893712574850298, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1297, "step": 15293 }, { "epoch": 22.895209580838323, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1329, "step": 15294 }, { "epoch": 22.896706586826348, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1289, "step": 15295 }, { "epoch": 22.898203592814372, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1276, "step": 15296 }, { "epoch": 22.899700598802394, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1294, "step": 15297 }, { "epoch": 22.90119760479042, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 15298 }, { "epoch": 22.902694610778443, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1356, "step": 15299 }, { "epoch": 22.904191616766468, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1353, "step": 15300 }, { "epoch": 22.90568862275449, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1327, "step": 15301 }, { "epoch": 22.907185628742514, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1277, "step": 15302 }, { "epoch": 22.90868263473054, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1415, "step": 15303 }, { "epoch": 22.910179640718564, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 15304 }, { "epoch": 22.91167664670659, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1367, "step": 15305 }, { "epoch": 22.91317365269461, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1346, "step": 15306 }, { "epoch": 22.914670658682635, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1301, "step": 15307 }, { "epoch": 22.91616766467066, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1288, "step": 15308 }, { "epoch": 22.91766467065868, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1281, "step": 15309 }, { "epoch": 22.919161676646706, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1338, "step": 15310 }, { "epoch": 22.92065868263473, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1341, "step": 15311 }, { "epoch": 22.922155688622755, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1317, "step": 15312 }, { "epoch": 22.92365269461078, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1327, "step": 15313 }, { "epoch": 22.9251497005988, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1323, "step": 15314 }, { "epoch": 22.926646706586826, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 15315 }, { "epoch": 22.92814371257485, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1381, "step": 15316 }, { "epoch": 22.929640718562876, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1321, "step": 15317 }, { "epoch": 22.931137724550897, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1391, "step": 15318 }, { "epoch": 22.932634730538922, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.142, "step": 15319 }, { "epoch": 22.934131736526947, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.137, "step": 15320 }, { "epoch": 22.93562874251497, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1316, "step": 15321 }, { "epoch": 22.937125748502993, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1342, "step": 15322 }, { "epoch": 22.938622754491018, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.132, "step": 15323 }, { "epoch": 22.940119760479043, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1395, "step": 15324 }, { "epoch": 22.941616766467067, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1393, "step": 15325 }, { "epoch": 22.94311377245509, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1398, "step": 15326 }, { "epoch": 22.944610778443113, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1396, "step": 15327 }, { "epoch": 22.94610778443114, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1272, "step": 15328 }, { "epoch": 22.947604790419163, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1295, "step": 15329 }, { "epoch": 22.949101796407184, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1345, "step": 15330 }, { "epoch": 22.95059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 15331 }, { "epoch": 22.952095808383234, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 15332 }, { "epoch": 22.95359281437126, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 15333 }, { "epoch": 22.95508982035928, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 15334 }, { "epoch": 22.956586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1376, "step": 15335 }, { "epoch": 22.95808383233533, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.13, "step": 15336 }, { "epoch": 22.959580838323355, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.133, "step": 15337 }, { "epoch": 22.961077844311376, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1389, "step": 15338 }, { "epoch": 22.9625748502994, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1371, "step": 15339 }, { "epoch": 22.964071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1326, "step": 15340 }, { "epoch": 22.96556886227545, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 15341 }, { "epoch": 22.96706586826347, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1276, "step": 15342 }, { "epoch": 22.968562874251496, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1317, "step": 15343 }, { "epoch": 22.97005988023952, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1329, "step": 15344 }, { "epoch": 22.971556886227546, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1384, "step": 15345 }, { "epoch": 22.973053892215567, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.128, "step": 15346 }, { "epoch": 22.974550898203592, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1363, "step": 15347 }, { "epoch": 22.976047904191617, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1343, "step": 15348 }, { "epoch": 22.977544910179642, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1287, "step": 15349 }, { "epoch": 22.979041916167663, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1346, "step": 15350 }, { "epoch": 22.980538922155688, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.128, "step": 15351 }, { "epoch": 22.982035928143713, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1337, "step": 15352 }, { "epoch": 22.983532934131738, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1339, "step": 15353 }, { "epoch": 22.98502994011976, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1337, "step": 15354 }, { "epoch": 22.986526946107784, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 15355 }, { "epoch": 22.98802395209581, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1365, "step": 15356 }, { "epoch": 22.989520958083833, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1344, "step": 15357 }, { "epoch": 22.991017964071855, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1312, "step": 15358 }, { "epoch": 22.99251497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1278, "step": 15359 }, { "epoch": 22.994011976047904, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1393, "step": 15360 }, { "epoch": 22.99550898203593, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1367, "step": 15361 }, { "epoch": 22.99700598802395, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 15362 }, { "epoch": 22.998502994011975, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1262, "step": 15363 }, { "epoch": 23.0, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 15364 }, { "epoch": 23.001497005988025, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1261, "step": 15365 }, { "epoch": 23.00299401197605, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1319, "step": 15366 }, { "epoch": 23.00449101796407, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.124, "step": 15367 }, { "epoch": 23.005988023952096, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1335, "step": 15368 }, { "epoch": 23.00748502994012, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.132, "step": 15369 }, { "epoch": 23.008982035928145, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1353, "step": 15370 }, { "epoch": 23.010479041916167, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1326, "step": 15371 }, { "epoch": 23.01197604790419, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1359, "step": 15372 }, { "epoch": 23.013473053892216, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1352, "step": 15373 }, { "epoch": 23.01497005988024, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1383, "step": 15374 }, { "epoch": 23.016467065868262, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 15375 }, { "epoch": 23.017964071856287, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1362, "step": 15376 }, { "epoch": 23.019461077844312, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1261, "step": 15377 }, { "epoch": 23.020958083832337, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1355, "step": 15378 }, { "epoch": 23.022455089820358, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1298, "step": 15379 }, { "epoch": 23.023952095808383, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1367, "step": 15380 }, { "epoch": 23.025449101796408, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1372, "step": 15381 }, { "epoch": 23.026946107784433, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1341, "step": 15382 }, { "epoch": 23.028443113772454, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1326, "step": 15383 }, { "epoch": 23.02994011976048, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1284, "step": 15384 }, { "epoch": 23.031437125748504, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 15385 }, { "epoch": 23.03293413173653, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1319, "step": 15386 }, { "epoch": 23.03443113772455, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 15387 }, { "epoch": 23.035928143712574, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.138, "step": 15388 }, { "epoch": 23.0374251497006, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1354, "step": 15389 }, { "epoch": 23.038922155688624, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1345, "step": 15390 }, { "epoch": 23.040419161676645, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1275, "step": 15391 }, { "epoch": 23.04191616766467, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15392 }, { "epoch": 23.043413173652695, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 15393 }, { "epoch": 23.04491017964072, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1274, "step": 15394 }, { "epoch": 23.04640718562874, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1287, "step": 15395 }, { "epoch": 23.047904191616766, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.136, "step": 15396 }, { "epoch": 23.04940119760479, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1347, "step": 15397 }, { "epoch": 23.050898203592816, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1332, "step": 15398 }, { "epoch": 23.052395209580837, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1389, "step": 15399 }, { "epoch": 23.05389221556886, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1302, "step": 15400 }, { "epoch": 23.055389221556887, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1312, "step": 15401 }, { "epoch": 23.05688622754491, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1363, "step": 15402 }, { "epoch": 23.058383233532933, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1324, "step": 15403 }, { "epoch": 23.059880239520957, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 15404 }, { "epoch": 23.061377245508982, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1292, "step": 15405 }, { "epoch": 23.062874251497007, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1382, "step": 15406 }, { "epoch": 23.06437125748503, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1303, "step": 15407 }, { "epoch": 23.065868263473053, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1401, "step": 15408 }, { "epoch": 23.067365269461078, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1338, "step": 15409 }, { "epoch": 23.068862275449103, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1317, "step": 15410 }, { "epoch": 23.070359281437124, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1279, "step": 15411 }, { "epoch": 23.07185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1312, "step": 15412 }, { "epoch": 23.073353293413174, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1322, "step": 15413 }, { "epoch": 23.0748502994012, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15414 }, { "epoch": 23.07634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15415 }, { "epoch": 23.077844311377245, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1365, "step": 15416 }, { "epoch": 23.07934131736527, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1342, "step": 15417 }, { "epoch": 23.080838323353294, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1319, "step": 15418 }, { "epoch": 23.082335329341316, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 15419 }, { "epoch": 23.08383233532934, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1336, "step": 15420 }, { "epoch": 23.085329341317365, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1314, "step": 15421 }, { "epoch": 23.08682634730539, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1313, "step": 15422 }, { "epoch": 23.088323353293415, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1352, "step": 15423 }, { "epoch": 23.089820359281436, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 15424 }, { "epoch": 23.09131736526946, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1313, "step": 15425 }, { "epoch": 23.092814371257486, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1295, "step": 15426 }, { "epoch": 23.09431137724551, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1314, "step": 15427 }, { "epoch": 23.095808383233532, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1369, "step": 15428 }, { "epoch": 23.097305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1335, "step": 15429 }, { "epoch": 23.09880239520958, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1287, "step": 15430 }, { "epoch": 23.100299401197606, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1347, "step": 15431 }, { "epoch": 23.101796407185628, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1352, "step": 15432 }, { "epoch": 23.103293413173652, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1325, "step": 15433 }, { "epoch": 23.104790419161677, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1299, "step": 15434 }, { "epoch": 23.106287425149702, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1274, "step": 15435 }, { "epoch": 23.107784431137723, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 15436 }, { "epoch": 23.10928143712575, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 15437 }, { "epoch": 23.110778443113773, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1275, "step": 15438 }, { "epoch": 23.112275449101798, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 15439 }, { "epoch": 23.11377245508982, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1274, "step": 15440 }, { "epoch": 23.115269461077844, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.131, "step": 15441 }, { "epoch": 23.11676646706587, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1302, "step": 15442 }, { "epoch": 23.118263473053894, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1338, "step": 15443 }, { "epoch": 23.119760479041915, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1343, "step": 15444 }, { "epoch": 23.12125748502994, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1315, "step": 15445 }, { "epoch": 23.122754491017965, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.13, "step": 15446 }, { "epoch": 23.12425149700599, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1315, "step": 15447 }, { "epoch": 23.12574850299401, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1278, "step": 15448 }, { "epoch": 23.127245508982035, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 15449 }, { "epoch": 23.12874251497006, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1252, "step": 15450 }, { "epoch": 23.130239520958085, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.133, "step": 15451 }, { "epoch": 23.131736526946106, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15452 }, { "epoch": 23.13323353293413, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1335, "step": 15453 }, { "epoch": 23.134730538922156, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1274, "step": 15454 }, { "epoch": 23.13622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1331, "step": 15455 }, { "epoch": 23.137724550898202, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1339, "step": 15456 }, { "epoch": 23.139221556886227, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15457 }, { "epoch": 23.14071856287425, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.127, "step": 15458 }, { "epoch": 23.142215568862277, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 15459 }, { "epoch": 23.143712574850298, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1345, "step": 15460 }, { "epoch": 23.145209580838323, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 15461 }, { "epoch": 23.146706586826348, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1331, "step": 15462 }, { "epoch": 23.148203592814372, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1321, "step": 15463 }, { "epoch": 23.149700598802394, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1367, "step": 15464 }, { "epoch": 23.15119760479042, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1285, "step": 15465 }, { "epoch": 23.152694610778443, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1324, "step": 15466 }, { "epoch": 23.154191616766468, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 15467 }, { "epoch": 23.15568862275449, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.133, "step": 15468 }, { "epoch": 23.157185628742514, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1352, "step": 15469 }, { "epoch": 23.15868263473054, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1317, "step": 15470 }, { "epoch": 23.160179640718564, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1399, "step": 15471 }, { "epoch": 23.161676646706585, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 15472 }, { "epoch": 23.16317365269461, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15473 }, { "epoch": 23.164670658682635, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15474 }, { "epoch": 23.16616766467066, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1291, "step": 15475 }, { "epoch": 23.16766467065868, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1331, "step": 15476 }, { "epoch": 23.169161676646706, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1309, "step": 15477 }, { "epoch": 23.17065868263473, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1347, "step": 15478 }, { "epoch": 23.172155688622755, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 15479 }, { "epoch": 23.17365269461078, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1366, "step": 15480 }, { "epoch": 23.1751497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15481 }, { "epoch": 23.176646706586826, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1223, "step": 15482 }, { "epoch": 23.17814371257485, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1333, "step": 15483 }, { "epoch": 23.179640718562876, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1343, "step": 15484 }, { "epoch": 23.181137724550897, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1345, "step": 15485 }, { "epoch": 23.182634730538922, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1356, "step": 15486 }, { "epoch": 23.184131736526947, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1365, "step": 15487 }, { "epoch": 23.18562874251497, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1415, "step": 15488 }, { "epoch": 23.187125748502993, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 15489 }, { "epoch": 23.188622754491018, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1289, "step": 15490 }, { "epoch": 23.190119760479043, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.14, "step": 15491 }, { "epoch": 23.191616766467067, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1387, "step": 15492 }, { "epoch": 23.19311377245509, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1281, "step": 15493 }, { "epoch": 23.194610778443113, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1343, "step": 15494 }, { "epoch": 23.19610778443114, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1311, "step": 15495 }, { "epoch": 23.197604790419163, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1329, "step": 15496 }, { "epoch": 23.199101796407184, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1286, "step": 15497 }, { "epoch": 23.20059880239521, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1361, "step": 15498 }, { "epoch": 23.202095808383234, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.135, "step": 15499 }, { "epoch": 23.20359281437126, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1255, "step": 15500 }, { "epoch": 23.20508982035928, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1337, "step": 15501 }, { "epoch": 23.206586826347305, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1337, "step": 15502 }, { "epoch": 23.20808383233533, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1358, "step": 15503 }, { "epoch": 23.209580838323355, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1342, "step": 15504 }, { "epoch": 23.211077844311376, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1351, "step": 15505 }, { "epoch": 23.2125748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1329, "step": 15506 }, { "epoch": 23.214071856287426, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1293, "step": 15507 }, { "epoch": 23.21556886227545, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1268, "step": 15508 }, { "epoch": 23.21706586826347, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1384, "step": 15509 }, { "epoch": 23.218562874251496, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 15510 }, { "epoch": 23.22005988023952, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1353, "step": 15511 }, { "epoch": 23.221556886227546, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1316, "step": 15512 }, { "epoch": 23.223053892215567, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1378, "step": 15513 }, { "epoch": 23.224550898203592, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1263, "step": 15514 }, { "epoch": 23.226047904191617, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1328, "step": 15515 }, { "epoch": 23.227544910179642, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1291, "step": 15516 }, { "epoch": 23.229041916167663, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15517 }, { "epoch": 23.230538922155688, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1292, "step": 15518 }, { "epoch": 23.232035928143713, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1306, "step": 15519 }, { "epoch": 23.233532934131738, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1298, "step": 15520 }, { "epoch": 23.23502994011976, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.135, "step": 15521 }, { "epoch": 23.236526946107784, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1333, "step": 15522 }, { "epoch": 23.23802395209581, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.134, "step": 15523 }, { "epoch": 23.239520958083833, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.129, "step": 15524 }, { "epoch": 23.241017964071855, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1319, "step": 15525 }, { "epoch": 23.24251497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1392, "step": 15526 }, { "epoch": 23.244011976047904, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1304, "step": 15527 }, { "epoch": 23.24550898203593, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1388, "step": 15528 }, { "epoch": 23.24700598802395, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1343, "step": 15529 }, { "epoch": 23.248502994011975, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 15530 }, { "epoch": 23.25, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1338, "step": 15531 }, { "epoch": 23.251497005988025, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1322, "step": 15532 }, { "epoch": 23.25299401197605, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1287, "step": 15533 }, { "epoch": 23.25449101796407, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1365, "step": 15534 }, { "epoch": 23.255988023952096, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1318, "step": 15535 }, { "epoch": 23.25748502994012, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.132, "step": 15536 }, { "epoch": 23.258982035928145, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1346, "step": 15537 }, { "epoch": 23.260479041916167, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1318, "step": 15538 }, { "epoch": 23.26197604790419, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1341, "step": 15539 }, { "epoch": 23.263473053892216, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1291, "step": 15540 }, { "epoch": 23.26497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15541 }, { "epoch": 23.266467065868262, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1313, "step": 15542 }, { "epoch": 23.267964071856287, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 15543 }, { "epoch": 23.269461077844312, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1338, "step": 15544 }, { "epoch": 23.270958083832337, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1318, "step": 15545 }, { "epoch": 23.272455089820358, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.135, "step": 15546 }, { "epoch": 23.273952095808383, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1324, "step": 15547 }, { "epoch": 23.275449101796408, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 15548 }, { "epoch": 23.276946107784433, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1398, "step": 15549 }, { "epoch": 23.278443113772454, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1304, "step": 15550 }, { "epoch": 23.27994011976048, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1331, "step": 15551 }, { "epoch": 23.281437125748504, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1316, "step": 15552 }, { "epoch": 23.28293413173653, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1361, "step": 15553 }, { "epoch": 23.28443113772455, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 15554 }, { "epoch": 23.285928143712574, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1289, "step": 15555 }, { "epoch": 23.2874251497006, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1264, "step": 15556 }, { "epoch": 23.288922155688624, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1364, "step": 15557 }, { "epoch": 23.290419161676645, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.123, "step": 15558 }, { "epoch": 23.29191616766467, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1358, "step": 15559 }, { "epoch": 23.293413173652695, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15560 }, { "epoch": 23.29491017964072, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1245, "step": 15561 }, { "epoch": 23.29640718562874, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1285, "step": 15562 }, { "epoch": 23.297904191616766, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1358, "step": 15563 }, { "epoch": 23.29940119760479, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1385, "step": 15564 }, { "epoch": 23.300898203592816, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1333, "step": 15565 }, { "epoch": 23.302395209580837, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1333, "step": 15566 }, { "epoch": 23.30389221556886, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1358, "step": 15567 }, { "epoch": 23.305389221556887, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1343, "step": 15568 }, { "epoch": 23.30688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 15569 }, { "epoch": 23.308383233532933, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 15570 }, { "epoch": 23.309880239520957, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1261, "step": 15571 }, { "epoch": 23.311377245508982, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1294, "step": 15572 }, { "epoch": 23.312874251497007, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1335, "step": 15573 }, { "epoch": 23.31437125748503, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1319, "step": 15574 }, { "epoch": 23.315868263473053, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.135, "step": 15575 }, { "epoch": 23.317365269461078, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1359, "step": 15576 }, { "epoch": 23.318862275449103, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1386, "step": 15577 }, { "epoch": 23.320359281437124, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1261, "step": 15578 }, { "epoch": 23.32185628742515, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.125, "step": 15579 }, { "epoch": 23.323353293413174, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1281, "step": 15580 }, { "epoch": 23.3248502994012, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1362, "step": 15581 }, { "epoch": 23.32634730538922, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1299, "step": 15582 }, { "epoch": 23.327844311377245, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1334, "step": 15583 }, { "epoch": 23.32934131736527, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1325, "step": 15584 }, { "epoch": 23.330838323353294, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1277, "step": 15585 }, { "epoch": 23.33233532934132, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 15586 }, { "epoch": 23.33383233532934, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1281, "step": 15587 }, { "epoch": 23.335329341317365, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1399, "step": 15588 }, { "epoch": 23.33682634730539, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 15589 }, { "epoch": 23.338323353293415, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1282, "step": 15590 }, { "epoch": 23.339820359281436, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1339, "step": 15591 }, { "epoch": 23.34131736526946, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1308, "step": 15592 }, { "epoch": 23.342814371257486, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1358, "step": 15593 }, { "epoch": 23.34431137724551, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1289, "step": 15594 }, { "epoch": 23.345808383233532, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1274, "step": 15595 }, { "epoch": 23.347305389221557, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1314, "step": 15596 }, { "epoch": 23.34880239520958, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1277, "step": 15597 }, { "epoch": 23.350299401197606, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1386, "step": 15598 }, { "epoch": 23.351796407185628, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1366, "step": 15599 }, { "epoch": 23.353293413173652, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1317, "step": 15600 }, { "epoch": 23.354790419161677, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 15601 }, { "epoch": 23.356287425149702, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.127, "step": 15602 }, { "epoch": 23.357784431137723, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1338, "step": 15603 }, { "epoch": 23.35928143712575, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1329, "step": 15604 }, { "epoch": 23.360778443113773, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1277, "step": 15605 }, { "epoch": 23.362275449101798, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15606 }, { "epoch": 23.36377245508982, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1327, "step": 15607 }, { "epoch": 23.365269461077844, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1323, "step": 15608 }, { "epoch": 23.36676646706587, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1315, "step": 15609 }, { "epoch": 23.368263473053894, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1398, "step": 15610 }, { "epoch": 23.369760479041915, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1298, "step": 15611 }, { "epoch": 23.37125748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1311, "step": 15612 }, { "epoch": 23.372754491017965, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1295, "step": 15613 }, { "epoch": 23.37425149700599, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15614 }, { "epoch": 23.37574850299401, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1349, "step": 15615 }, { "epoch": 23.377245508982035, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1326, "step": 15616 }, { "epoch": 23.37874251497006, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15617 }, { "epoch": 23.380239520958085, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1323, "step": 15618 }, { "epoch": 23.381736526946106, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 15619 }, { "epoch": 23.38323353293413, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 15620 }, { "epoch": 23.384730538922156, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1343, "step": 15621 }, { "epoch": 23.38622754491018, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1329, "step": 15622 }, { "epoch": 23.387724550898202, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1308, "step": 15623 }, { "epoch": 23.389221556886227, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1371, "step": 15624 }, { "epoch": 23.39071856287425, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1302, "step": 15625 }, { "epoch": 23.392215568862277, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1363, "step": 15626 }, { "epoch": 23.393712574850298, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.137, "step": 15627 }, { "epoch": 23.395209580838323, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1337, "step": 15628 }, { "epoch": 23.396706586826348, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1389, "step": 15629 }, { "epoch": 23.398203592814372, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1357, "step": 15630 }, { "epoch": 23.399700598802394, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1298, "step": 15631 }, { "epoch": 23.40119760479042, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1276, "step": 15632 }, { "epoch": 23.402694610778443, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1377, "step": 15633 }, { "epoch": 23.404191616766468, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1353, "step": 15634 }, { "epoch": 23.40568862275449, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1318, "step": 15635 }, { "epoch": 23.407185628742514, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1316, "step": 15636 }, { "epoch": 23.40868263473054, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1369, "step": 15637 }, { "epoch": 23.410179640718564, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1335, "step": 15638 }, { "epoch": 23.411676646706585, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1413, "step": 15639 }, { "epoch": 23.41317365269461, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1293, "step": 15640 }, { "epoch": 23.414670658682635, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1313, "step": 15641 }, { "epoch": 23.41616766467066, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1239, "step": 15642 }, { "epoch": 23.41766467065868, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1396, "step": 15643 }, { "epoch": 23.419161676646706, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1355, "step": 15644 }, { "epoch": 23.42065868263473, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 15645 }, { "epoch": 23.422155688622755, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1312, "step": 15646 }, { "epoch": 23.42365269461078, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15647 }, { "epoch": 23.4251497005988, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1317, "step": 15648 }, { "epoch": 23.426646706586826, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1359, "step": 15649 }, { "epoch": 23.42814371257485, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1241, "step": 15650 }, { "epoch": 23.429640718562876, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1286, "step": 15651 }, { "epoch": 23.431137724550897, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1341, "step": 15652 }, { "epoch": 23.432634730538922, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1299, "step": 15653 }, { "epoch": 23.434131736526947, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1308, "step": 15654 }, { "epoch": 23.43562874251497, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1333, "step": 15655 }, { "epoch": 23.437125748502993, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1347, "step": 15656 }, { "epoch": 23.438622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1351, "step": 15657 }, { "epoch": 23.440119760479043, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 15658 }, { "epoch": 23.441616766467067, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1326, "step": 15659 }, { "epoch": 23.44311377245509, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1302, "step": 15660 }, { "epoch": 23.444610778443113, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1273, "step": 15661 }, { "epoch": 23.44610778443114, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1342, "step": 15662 }, { "epoch": 23.447604790419163, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1327, "step": 15663 }, { "epoch": 23.449101796407184, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1358, "step": 15664 }, { "epoch": 23.45059880239521, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1307, "step": 15665 }, { "epoch": 23.452095808383234, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.136, "step": 15666 }, { "epoch": 23.45359281437126, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1361, "step": 15667 }, { "epoch": 23.45508982035928, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.131, "step": 15668 }, { "epoch": 23.456586826347305, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1298, "step": 15669 }, { "epoch": 23.45808383233533, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.127, "step": 15670 }, { "epoch": 23.459580838323355, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1307, "step": 15671 }, { "epoch": 23.461077844311376, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1337, "step": 15672 }, { "epoch": 23.4625748502994, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1326, "step": 15673 }, { "epoch": 23.464071856287426, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1348, "step": 15674 }, { "epoch": 23.46556886227545, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1306, "step": 15675 }, { "epoch": 23.46706586826347, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1291, "step": 15676 }, { "epoch": 23.468562874251496, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1343, "step": 15677 }, { "epoch": 23.47005988023952, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1291, "step": 15678 }, { "epoch": 23.471556886227546, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 15679 }, { "epoch": 23.473053892215567, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1321, "step": 15680 }, { "epoch": 23.474550898203592, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1308, "step": 15681 }, { "epoch": 23.476047904191617, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1339, "step": 15682 }, { "epoch": 23.477544910179642, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.134, "step": 15683 }, { "epoch": 23.479041916167663, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1299, "step": 15684 }, { "epoch": 23.480538922155688, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1355, "step": 15685 }, { "epoch": 23.482035928143713, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1375, "step": 15686 }, { "epoch": 23.483532934131738, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1274, "step": 15687 }, { "epoch": 23.48502994011976, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1279, "step": 15688 }, { "epoch": 23.486526946107784, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1293, "step": 15689 }, { "epoch": 23.48802395209581, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.132, "step": 15690 }, { "epoch": 23.489520958083833, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 15691 }, { "epoch": 23.491017964071855, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 15692 }, { "epoch": 23.49251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1259, "step": 15693 }, { "epoch": 23.494011976047904, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1244, "step": 15694 }, { "epoch": 23.49550898203593, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15695 }, { "epoch": 23.49700598802395, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.136, "step": 15696 }, { "epoch": 23.498502994011975, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1285, "step": 15697 }, { "epoch": 23.5, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1413, "step": 15698 }, { "epoch": 23.501497005988025, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1337, "step": 15699 }, { "epoch": 23.50299401197605, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1341, "step": 15700 }, { "epoch": 23.50449101796407, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 15701 }, { "epoch": 23.505988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 15702 }, { "epoch": 23.50748502994012, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 15703 }, { "epoch": 23.508982035928145, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1331, "step": 15704 }, { "epoch": 23.510479041916167, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 15705 }, { "epoch": 23.51197604790419, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1273, "step": 15706 }, { "epoch": 23.513473053892216, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1328, "step": 15707 }, { "epoch": 23.51497005988024, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1363, "step": 15708 }, { "epoch": 23.516467065868262, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15709 }, { "epoch": 23.517964071856287, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 15710 }, { "epoch": 23.519461077844312, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 15711 }, { "epoch": 23.520958083832337, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1328, "step": 15712 }, { "epoch": 23.522455089820358, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1362, "step": 15713 }, { "epoch": 23.523952095808383, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1314, "step": 15714 }, { "epoch": 23.525449101796408, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1262, "step": 15715 }, { "epoch": 23.526946107784433, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.136, "step": 15716 }, { "epoch": 23.528443113772454, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1344, "step": 15717 }, { "epoch": 23.52994011976048, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1349, "step": 15718 }, { "epoch": 23.531437125748504, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1311, "step": 15719 }, { "epoch": 23.53293413173653, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1333, "step": 15720 }, { "epoch": 23.53443113772455, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1274, "step": 15721 }, { "epoch": 23.535928143712574, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1317, "step": 15722 }, { "epoch": 23.5374251497006, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1277, "step": 15723 }, { "epoch": 23.538922155688624, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1313, "step": 15724 }, { "epoch": 23.540419161676645, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1385, "step": 15725 }, { "epoch": 23.54191616766467, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1287, "step": 15726 }, { "epoch": 23.543413173652695, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1271, "step": 15727 }, { "epoch": 23.54491017964072, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1347, "step": 15728 }, { "epoch": 23.54640718562874, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1363, "step": 15729 }, { "epoch": 23.547904191616766, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1342, "step": 15730 }, { "epoch": 23.54940119760479, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1385, "step": 15731 }, { "epoch": 23.550898203592816, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1357, "step": 15732 }, { "epoch": 23.552395209580837, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.134, "step": 15733 }, { "epoch": 23.55389221556886, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1328, "step": 15734 }, { "epoch": 23.555389221556887, "grad_norm": 0.05029296875, "learning_rate": 0.0008, "loss": 1.1303, "step": 15735 }, { "epoch": 23.55688622754491, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1342, "step": 15736 }, { "epoch": 23.558383233532933, "grad_norm": 0.0498046875, "learning_rate": 0.0008, "loss": 1.1365, "step": 15737 }, { "epoch": 23.559880239520957, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1319, "step": 15738 }, { "epoch": 23.561377245508982, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1309, "step": 15739 }, { "epoch": 23.562874251497007, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1312, "step": 15740 }, { "epoch": 23.56437125748503, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1347, "step": 15741 }, { "epoch": 23.565868263473053, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1345, "step": 15742 }, { "epoch": 23.567365269461078, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 15743 }, { "epoch": 23.568862275449103, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1236, "step": 15744 }, { "epoch": 23.570359281437124, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1256, "step": 15745 }, { "epoch": 23.57185628742515, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1369, "step": 15746 }, { "epoch": 23.573353293413174, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1337, "step": 15747 }, { "epoch": 23.5748502994012, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1292, "step": 15748 }, { "epoch": 23.57634730538922, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1398, "step": 15749 }, { "epoch": 23.577844311377245, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1321, "step": 15750 }, { "epoch": 23.57934131736527, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1296, "step": 15751 }, { "epoch": 23.580838323353294, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1376, "step": 15752 }, { "epoch": 23.58233532934132, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1323, "step": 15753 }, { "epoch": 23.58383233532934, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.134, "step": 15754 }, { "epoch": 23.585329341317365, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.127, "step": 15755 }, { "epoch": 23.58682634730539, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1326, "step": 15756 }, { "epoch": 23.58832335329341, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1395, "step": 15757 }, { "epoch": 23.589820359281436, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1308, "step": 15758 }, { "epoch": 23.59131736526946, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1302, "step": 15759 }, { "epoch": 23.592814371257486, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.134, "step": 15760 }, { "epoch": 23.59431137724551, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1342, "step": 15761 }, { "epoch": 23.595808383233532, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1322, "step": 15762 }, { "epoch": 23.597305389221557, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1276, "step": 15763 }, { "epoch": 23.59880239520958, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.131, "step": 15764 }, { "epoch": 23.600299401197606, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.126, "step": 15765 }, { "epoch": 23.601796407185628, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1279, "step": 15766 }, { "epoch": 23.603293413173652, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1353, "step": 15767 }, { "epoch": 23.604790419161677, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1281, "step": 15768 }, { "epoch": 23.606287425149702, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1372, "step": 15769 }, { "epoch": 23.607784431137723, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1336, "step": 15770 }, { "epoch": 23.60928143712575, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 15771 }, { "epoch": 23.610778443113773, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1276, "step": 15772 }, { "epoch": 23.612275449101798, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1352, "step": 15773 }, { "epoch": 23.61377245508982, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1317, "step": 15774 }, { "epoch": 23.615269461077844, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1284, "step": 15775 }, { "epoch": 23.61676646706587, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1327, "step": 15776 }, { "epoch": 23.618263473053894, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1307, "step": 15777 }, { "epoch": 23.619760479041915, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1337, "step": 15778 }, { "epoch": 23.62125748502994, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1372, "step": 15779 }, { "epoch": 23.622754491017965, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1265, "step": 15780 }, { "epoch": 23.62425149700599, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1289, "step": 15781 }, { "epoch": 23.62574850299401, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1368, "step": 15782 }, { "epoch": 23.627245508982035, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1395, "step": 15783 }, { "epoch": 23.62874251497006, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.1331, "step": 15784 }, { "epoch": 23.630239520958085, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.1357, "step": 15785 }, { "epoch": 23.631736526946106, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1308, "step": 15786 }, { "epoch": 23.63323353293413, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1364, "step": 15787 }, { "epoch": 23.634730538922156, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15788 }, { "epoch": 23.63622754491018, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.127, "step": 15789 }, { "epoch": 23.637724550898202, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1243, "step": 15790 }, { "epoch": 23.639221556886227, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1314, "step": 15791 }, { "epoch": 23.64071856287425, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1366, "step": 15792 }, { "epoch": 23.642215568862277, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1288, "step": 15793 }, { "epoch": 23.643712574850298, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1387, "step": 15794 }, { "epoch": 23.645209580838323, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1368, "step": 15795 }, { "epoch": 23.646706586826348, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.134, "step": 15796 }, { "epoch": 23.648203592814372, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.14, "step": 15797 }, { "epoch": 23.649700598802394, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1385, "step": 15798 }, { "epoch": 23.65119760479042, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1263, "step": 15799 }, { "epoch": 23.652694610778443, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1407, "step": 15800 }, { "epoch": 23.654191616766468, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1337, "step": 15801 }, { "epoch": 23.65568862275449, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1291, "step": 15802 }, { "epoch": 23.657185628742514, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1332, "step": 15803 }, { "epoch": 23.65868263473054, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.131, "step": 15804 }, { "epoch": 23.660179640718564, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1324, "step": 15805 }, { "epoch": 23.66167664670659, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1364, "step": 15806 }, { "epoch": 23.66317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1282, "step": 15807 }, { "epoch": 23.664670658682635, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1288, "step": 15808 }, { "epoch": 23.66616766467066, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 15809 }, { "epoch": 23.66766467065868, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1344, "step": 15810 }, { "epoch": 23.669161676646706, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.133, "step": 15811 }, { "epoch": 23.67065868263473, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1335, "step": 15812 }, { "epoch": 23.672155688622755, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1297, "step": 15813 }, { "epoch": 23.67365269461078, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1305, "step": 15814 }, { "epoch": 23.6751497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1261, "step": 15815 }, { "epoch": 23.676646706586826, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1229, "step": 15816 }, { "epoch": 23.67814371257485, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1305, "step": 15817 }, { "epoch": 23.679640718562876, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.135, "step": 15818 }, { "epoch": 23.681137724550897, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1292, "step": 15819 }, { "epoch": 23.682634730538922, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1328, "step": 15820 }, { "epoch": 23.684131736526947, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.135, "step": 15821 }, { "epoch": 23.68562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1337, "step": 15822 }, { "epoch": 23.687125748502993, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1286, "step": 15823 }, { "epoch": 23.688622754491018, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1308, "step": 15824 }, { "epoch": 23.690119760479043, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1314, "step": 15825 }, { "epoch": 23.691616766467067, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 15826 }, { "epoch": 23.69311377245509, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.13, "step": 15827 }, { "epoch": 23.694610778443113, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1305, "step": 15828 }, { "epoch": 23.69610778443114, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1343, "step": 15829 }, { "epoch": 23.697604790419163, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.13, "step": 15830 }, { "epoch": 23.699101796407184, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1256, "step": 15831 }, { "epoch": 23.70059880239521, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1304, "step": 15832 }, { "epoch": 23.702095808383234, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1308, "step": 15833 }, { "epoch": 23.70359281437126, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1326, "step": 15834 }, { "epoch": 23.70508982035928, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.129, "step": 15835 }, { "epoch": 23.706586826347305, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.13, "step": 15836 }, { "epoch": 23.70808383233533, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1276, "step": 15837 }, { "epoch": 23.709580838323355, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1329, "step": 15838 }, { "epoch": 23.711077844311376, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1364, "step": 15839 }, { "epoch": 23.7125748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15840 }, { "epoch": 23.714071856287426, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 15841 }, { "epoch": 23.71556886227545, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1222, "step": 15842 }, { "epoch": 23.71706586826347, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 15843 }, { "epoch": 23.718562874251496, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1308, "step": 15844 }, { "epoch": 23.72005988023952, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1358, "step": 15845 }, { "epoch": 23.721556886227546, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.13, "step": 15846 }, { "epoch": 23.723053892215567, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1387, "step": 15847 }, { "epoch": 23.724550898203592, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1354, "step": 15848 }, { "epoch": 23.726047904191617, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1294, "step": 15849 }, { "epoch": 23.727544910179642, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.128, "step": 15850 }, { "epoch": 23.729041916167663, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1334, "step": 15851 }, { "epoch": 23.730538922155688, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 15852 }, { "epoch": 23.732035928143713, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1293, "step": 15853 }, { "epoch": 23.733532934131738, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1358, "step": 15854 }, { "epoch": 23.73502994011976, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1297, "step": 15855 }, { "epoch": 23.736526946107784, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1291, "step": 15856 }, { "epoch": 23.73802395209581, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1295, "step": 15857 }, { "epoch": 23.739520958083833, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1281, "step": 15858 }, { "epoch": 23.741017964071855, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1415, "step": 15859 }, { "epoch": 23.74251497005988, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1318, "step": 15860 }, { "epoch": 23.744011976047904, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.129, "step": 15861 }, { "epoch": 23.74550898203593, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1375, "step": 15862 }, { "epoch": 23.74700598802395, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.139, "step": 15863 }, { "epoch": 23.748502994011975, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1339, "step": 15864 }, { "epoch": 23.75, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1331, "step": 15865 }, { "epoch": 23.751497005988025, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 15866 }, { "epoch": 23.75299401197605, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 15867 }, { "epoch": 23.75449101796407, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1267, "step": 15868 }, { "epoch": 23.755988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1287, "step": 15869 }, { "epoch": 23.75748502994012, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1378, "step": 15870 }, { "epoch": 23.758982035928145, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1358, "step": 15871 }, { "epoch": 23.760479041916167, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1371, "step": 15872 }, { "epoch": 23.76197604790419, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1247, "step": 15873 }, { "epoch": 23.763473053892216, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1412, "step": 15874 }, { "epoch": 23.76497005988024, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.137, "step": 15875 }, { "epoch": 23.766467065868262, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1281, "step": 15876 }, { "epoch": 23.767964071856287, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1264, "step": 15877 }, { "epoch": 23.769461077844312, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 15878 }, { "epoch": 23.770958083832337, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1343, "step": 15879 }, { "epoch": 23.772455089820358, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15880 }, { "epoch": 23.773952095808383, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1335, "step": 15881 }, { "epoch": 23.775449101796408, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1335, "step": 15882 }, { "epoch": 23.776946107784433, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1264, "step": 15883 }, { "epoch": 23.778443113772454, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1305, "step": 15884 }, { "epoch": 23.77994011976048, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.13, "step": 15885 }, { "epoch": 23.781437125748504, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1359, "step": 15886 }, { "epoch": 23.78293413173653, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1405, "step": 15887 }, { "epoch": 23.78443113772455, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1354, "step": 15888 }, { "epoch": 23.785928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1341, "step": 15889 }, { "epoch": 23.7874251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1279, "step": 15890 }, { "epoch": 23.788922155688624, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1332, "step": 15891 }, { "epoch": 23.790419161676645, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1342, "step": 15892 }, { "epoch": 23.79191616766467, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1268, "step": 15893 }, { "epoch": 23.793413173652695, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1262, "step": 15894 }, { "epoch": 23.79491017964072, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1314, "step": 15895 }, { "epoch": 23.79640718562874, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1365, "step": 15896 }, { "epoch": 23.797904191616766, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1329, "step": 15897 }, { "epoch": 23.79940119760479, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1235, "step": 15898 }, { "epoch": 23.800898203592816, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1358, "step": 15899 }, { "epoch": 23.802395209580837, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1337, "step": 15900 }, { "epoch": 23.80389221556886, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1337, "step": 15901 }, { "epoch": 23.805389221556887, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 15902 }, { "epoch": 23.80688622754491, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1253, "step": 15903 }, { "epoch": 23.808383233532933, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1287, "step": 15904 }, { "epoch": 23.809880239520957, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1264, "step": 15905 }, { "epoch": 23.811377245508982, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1397, "step": 15906 }, { "epoch": 23.812874251497007, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 15907 }, { "epoch": 23.81437125748503, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1361, "step": 15908 }, { "epoch": 23.815868263473053, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.127, "step": 15909 }, { "epoch": 23.817365269461078, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1324, "step": 15910 }, { "epoch": 23.818862275449103, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.132, "step": 15911 }, { "epoch": 23.820359281437124, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1311, "step": 15912 }, { "epoch": 23.82185628742515, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.126, "step": 15913 }, { "epoch": 23.823353293413174, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15914 }, { "epoch": 23.8248502994012, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1332, "step": 15915 }, { "epoch": 23.82634730538922, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 15916 }, { "epoch": 23.827844311377245, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1328, "step": 15917 }, { "epoch": 23.82934131736527, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 15918 }, { "epoch": 23.830838323353294, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1347, "step": 15919 }, { "epoch": 23.83233532934132, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1354, "step": 15920 }, { "epoch": 23.83383233532934, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.133, "step": 15921 }, { "epoch": 23.835329341317365, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15922 }, { "epoch": 23.83682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.133, "step": 15923 }, { "epoch": 23.83832335329341, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15924 }, { "epoch": 23.839820359281436, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.131, "step": 15925 }, { "epoch": 23.84131736526946, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1379, "step": 15926 }, { "epoch": 23.842814371257486, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1344, "step": 15927 }, { "epoch": 23.84431137724551, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1349, "step": 15928 }, { "epoch": 23.845808383233532, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.135, "step": 15929 }, { "epoch": 23.847305389221557, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1366, "step": 15930 }, { "epoch": 23.84880239520958, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1304, "step": 15931 }, { "epoch": 23.850299401197606, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1381, "step": 15932 }, { "epoch": 23.851796407185628, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1262, "step": 15933 }, { "epoch": 23.853293413173652, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1305, "step": 15934 }, { "epoch": 23.854790419161677, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1273, "step": 15935 }, { "epoch": 23.856287425149702, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1268, "step": 15936 }, { "epoch": 23.857784431137723, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1262, "step": 15937 }, { "epoch": 23.85928143712575, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1326, "step": 15938 }, { "epoch": 23.860778443113773, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1332, "step": 15939 }, { "epoch": 23.862275449101798, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1373, "step": 15940 }, { "epoch": 23.86377245508982, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1237, "step": 15941 }, { "epoch": 23.865269461077844, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1375, "step": 15942 }, { "epoch": 23.86676646706587, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1289, "step": 15943 }, { "epoch": 23.868263473053894, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1279, "step": 15944 }, { "epoch": 23.869760479041915, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1294, "step": 15945 }, { "epoch": 23.87125748502994, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1322, "step": 15946 }, { "epoch": 23.872754491017965, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.137, "step": 15947 }, { "epoch": 23.87425149700599, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1259, "step": 15948 }, { "epoch": 23.87574850299401, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.123, "step": 15949 }, { "epoch": 23.877245508982035, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1319, "step": 15950 }, { "epoch": 23.87874251497006, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1405, "step": 15951 }, { "epoch": 23.880239520958085, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1396, "step": 15952 }, { "epoch": 23.881736526946106, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1295, "step": 15953 }, { "epoch": 23.88323353293413, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1323, "step": 15954 }, { "epoch": 23.884730538922156, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.132, "step": 15955 }, { "epoch": 23.88622754491018, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 15956 }, { "epoch": 23.887724550898202, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.131, "step": 15957 }, { "epoch": 23.889221556886227, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1299, "step": 15958 }, { "epoch": 23.89071856287425, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1371, "step": 15959 }, { "epoch": 23.892215568862277, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1347, "step": 15960 }, { "epoch": 23.893712574850298, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 15961 }, { "epoch": 23.895209580838323, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.132, "step": 15962 }, { "epoch": 23.896706586826348, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1299, "step": 15963 }, { "epoch": 23.898203592814372, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1363, "step": 15964 }, { "epoch": 23.899700598802394, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1309, "step": 15965 }, { "epoch": 23.90119760479042, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1333, "step": 15966 }, { "epoch": 23.902694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.137, "step": 15967 }, { "epoch": 23.904191616766468, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1368, "step": 15968 }, { "epoch": 23.90568862275449, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.127, "step": 15969 }, { "epoch": 23.907185628742514, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1339, "step": 15970 }, { "epoch": 23.90868263473054, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1361, "step": 15971 }, { "epoch": 23.910179640718564, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1294, "step": 15972 }, { "epoch": 23.91167664670659, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1412, "step": 15973 }, { "epoch": 23.91317365269461, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 15974 }, { "epoch": 23.914670658682635, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1276, "step": 15975 }, { "epoch": 23.91616766467066, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.13, "step": 15976 }, { "epoch": 23.91766467065868, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15977 }, { "epoch": 23.919161676646706, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1349, "step": 15978 }, { "epoch": 23.92065868263473, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1307, "step": 15979 }, { "epoch": 23.922155688622755, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1305, "step": 15980 }, { "epoch": 23.92365269461078, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1373, "step": 15981 }, { "epoch": 23.9251497005988, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 15982 }, { "epoch": 23.926646706586826, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1263, "step": 15983 }, { "epoch": 23.92814371257485, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 15984 }, { "epoch": 23.929640718562876, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1308, "step": 15985 }, { "epoch": 23.931137724550897, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1348, "step": 15986 }, { "epoch": 23.932634730538922, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1282, "step": 15987 }, { "epoch": 23.934131736526947, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1235, "step": 15988 }, { "epoch": 23.93562874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1257, "step": 15989 }, { "epoch": 23.937125748502993, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1328, "step": 15990 }, { "epoch": 23.938622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 15991 }, { "epoch": 23.940119760479043, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1329, "step": 15992 }, { "epoch": 23.941616766467067, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.139, "step": 15993 }, { "epoch": 23.94311377245509, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1277, "step": 15994 }, { "epoch": 23.944610778443113, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1347, "step": 15995 }, { "epoch": 23.94610778443114, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1335, "step": 15996 }, { "epoch": 23.947604790419163, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1299, "step": 15997 }, { "epoch": 23.949101796407184, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1386, "step": 15998 }, { "epoch": 23.95059880239521, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1352, "step": 15999 }, { "epoch": 23.952095808383234, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1335, "step": 16000 }, { "epoch": 23.95359281437126, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16001 }, { "epoch": 23.95508982035928, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 16002 }, { "epoch": 23.956586826347305, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1353, "step": 16003 }, { "epoch": 23.95808383233533, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16004 }, { "epoch": 23.959580838323355, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1313, "step": 16005 }, { "epoch": 23.961077844311376, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.127, "step": 16006 }, { "epoch": 23.9625748502994, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1309, "step": 16007 }, { "epoch": 23.964071856287426, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1395, "step": 16008 }, { "epoch": 23.96556886227545, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.133, "step": 16009 }, { "epoch": 23.96706586826347, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1318, "step": 16010 }, { "epoch": 23.968562874251496, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16011 }, { "epoch": 23.97005988023952, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1349, "step": 16012 }, { "epoch": 23.971556886227546, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1332, "step": 16013 }, { "epoch": 23.973053892215567, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1319, "step": 16014 }, { "epoch": 23.974550898203592, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1332, "step": 16015 }, { "epoch": 23.976047904191617, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16016 }, { "epoch": 23.977544910179642, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16017 }, { "epoch": 23.979041916167663, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16018 }, { "epoch": 23.980538922155688, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1335, "step": 16019 }, { "epoch": 23.982035928143713, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 16020 }, { "epoch": 23.983532934131738, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.135, "step": 16021 }, { "epoch": 23.98502994011976, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1331, "step": 16022 }, { "epoch": 23.986526946107784, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1294, "step": 16023 }, { "epoch": 23.98802395209581, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 16024 }, { "epoch": 23.989520958083833, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1397, "step": 16025 }, { "epoch": 23.991017964071855, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1309, "step": 16026 }, { "epoch": 23.99251497005988, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1322, "step": 16027 }, { "epoch": 23.994011976047904, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 16028 }, { "epoch": 23.99550898203593, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1297, "step": 16029 }, { "epoch": 23.99700598802395, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1345, "step": 16030 }, { "epoch": 23.998502994011975, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1332, "step": 16031 }, { "epoch": 24.0, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 16032 }, { "epoch": 24.001497005988025, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16033 }, { "epoch": 24.00299401197605, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1313, "step": 16034 }, { "epoch": 24.00449101796407, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1307, "step": 16035 }, { "epoch": 24.005988023952096, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1313, "step": 16036 }, { "epoch": 24.00748502994012, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 16037 }, { "epoch": 24.008982035928145, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.133, "step": 16038 }, { "epoch": 24.010479041916167, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1292, "step": 16039 }, { "epoch": 24.01197604790419, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16040 }, { "epoch": 24.013473053892216, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1363, "step": 16041 }, { "epoch": 24.01497005988024, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1277, "step": 16042 }, { "epoch": 24.016467065868262, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1386, "step": 16043 }, { "epoch": 24.017964071856287, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1264, "step": 16044 }, { "epoch": 24.019461077844312, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1337, "step": 16045 }, { "epoch": 24.020958083832337, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1296, "step": 16046 }, { "epoch": 24.022455089820358, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1293, "step": 16047 }, { "epoch": 24.023952095808383, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1292, "step": 16048 }, { "epoch": 24.025449101796408, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1327, "step": 16049 }, { "epoch": 24.026946107784433, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1361, "step": 16050 }, { "epoch": 24.028443113772454, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1337, "step": 16051 }, { "epoch": 24.02994011976048, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1347, "step": 16052 }, { "epoch": 24.031437125748504, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1338, "step": 16053 }, { "epoch": 24.03293413173653, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1266, "step": 16054 }, { "epoch": 24.03443113772455, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1275, "step": 16055 }, { "epoch": 24.035928143712574, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1211, "step": 16056 }, { "epoch": 24.0374251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.128, "step": 16057 }, { "epoch": 24.038922155688624, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1314, "step": 16058 }, { "epoch": 24.040419161676645, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16059 }, { "epoch": 24.04191616766467, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1302, "step": 16060 }, { "epoch": 24.043413173652695, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 16061 }, { "epoch": 24.04491017964072, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1315, "step": 16062 }, { "epoch": 24.04640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1304, "step": 16063 }, { "epoch": 24.047904191616766, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1308, "step": 16064 }, { "epoch": 24.04940119760479, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1292, "step": 16065 }, { "epoch": 24.050898203592816, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 16066 }, { "epoch": 24.052395209580837, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1324, "step": 16067 }, { "epoch": 24.05389221556886, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1284, "step": 16068 }, { "epoch": 24.055389221556887, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1272, "step": 16069 }, { "epoch": 24.05688622754491, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1301, "step": 16070 }, { "epoch": 24.058383233532933, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1317, "step": 16071 }, { "epoch": 24.059880239520957, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1287, "step": 16072 }, { "epoch": 24.061377245508982, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1329, "step": 16073 }, { "epoch": 24.062874251497007, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1284, "step": 16074 }, { "epoch": 24.06437125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.138, "step": 16075 }, { "epoch": 24.065868263473053, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1264, "step": 16076 }, { "epoch": 24.067365269461078, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.133, "step": 16077 }, { "epoch": 24.068862275449103, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1319, "step": 16078 }, { "epoch": 24.070359281437124, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1284, "step": 16079 }, { "epoch": 24.07185628742515, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1311, "step": 16080 }, { "epoch": 24.073353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1364, "step": 16081 }, { "epoch": 24.0748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1285, "step": 16082 }, { "epoch": 24.07634730538922, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1318, "step": 16083 }, { "epoch": 24.077844311377245, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1342, "step": 16084 }, { "epoch": 24.07934131736527, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1281, "step": 16085 }, { "epoch": 24.080838323353294, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1331, "step": 16086 }, { "epoch": 24.082335329341316, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 16087 }, { "epoch": 24.08383233532934, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1288, "step": 16088 }, { "epoch": 24.085329341317365, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1361, "step": 16089 }, { "epoch": 24.08682634730539, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 16090 }, { "epoch": 24.088323353293415, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.133, "step": 16091 }, { "epoch": 24.089820359281436, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1385, "step": 16092 }, { "epoch": 24.09131736526946, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1304, "step": 16093 }, { "epoch": 24.092814371257486, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1321, "step": 16094 }, { "epoch": 24.09431137724551, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1241, "step": 16095 }, { "epoch": 24.095808383233532, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16096 }, { "epoch": 24.097305389221557, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1354, "step": 16097 }, { "epoch": 24.09880239520958, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1261, "step": 16098 }, { "epoch": 24.100299401197606, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1304, "step": 16099 }, { "epoch": 24.101796407185628, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1353, "step": 16100 }, { "epoch": 24.103293413173652, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 16101 }, { "epoch": 24.104790419161677, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.133, "step": 16102 }, { "epoch": 24.106287425149702, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1252, "step": 16103 }, { "epoch": 24.107784431137723, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1292, "step": 16104 }, { "epoch": 24.10928143712575, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1333, "step": 16105 }, { "epoch": 24.110778443113773, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 16106 }, { "epoch": 24.112275449101798, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1262, "step": 16107 }, { "epoch": 24.11377245508982, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1325, "step": 16108 }, { "epoch": 24.115269461077844, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1289, "step": 16109 }, { "epoch": 24.11676646706587, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1277, "step": 16110 }, { "epoch": 24.118263473053894, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1289, "step": 16111 }, { "epoch": 24.119760479041915, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1334, "step": 16112 }, { "epoch": 24.12125748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 16113 }, { "epoch": 24.122754491017965, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1288, "step": 16114 }, { "epoch": 24.12425149700599, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16115 }, { "epoch": 24.12574850299401, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 16116 }, { "epoch": 24.127245508982035, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 16117 }, { "epoch": 24.12874251497006, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1328, "step": 16118 }, { "epoch": 24.130239520958085, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1328, "step": 16119 }, { "epoch": 24.131736526946106, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16120 }, { "epoch": 24.13323353293413, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1273, "step": 16121 }, { "epoch": 24.134730538922156, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1298, "step": 16122 }, { "epoch": 24.13622754491018, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1292, "step": 16123 }, { "epoch": 24.137724550898202, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1303, "step": 16124 }, { "epoch": 24.139221556886227, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1377, "step": 16125 }, { "epoch": 24.14071856287425, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1367, "step": 16126 }, { "epoch": 24.142215568862277, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.138, "step": 16127 }, { "epoch": 24.143712574850298, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1353, "step": 16128 }, { "epoch": 24.145209580838323, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1336, "step": 16129 }, { "epoch": 24.146706586826348, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 16130 }, { "epoch": 24.148203592814372, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1269, "step": 16131 }, { "epoch": 24.149700598802394, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1288, "step": 16132 }, { "epoch": 24.15119760479042, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 16133 }, { "epoch": 24.152694610778443, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1301, "step": 16134 }, { "epoch": 24.154191616766468, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1302, "step": 16135 }, { "epoch": 24.15568862275449, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1347, "step": 16136 }, { "epoch": 24.157185628742514, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1326, "step": 16137 }, { "epoch": 24.15868263473054, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1352, "step": 16138 }, { "epoch": 24.160179640718564, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1256, "step": 16139 }, { "epoch": 24.161676646706585, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1271, "step": 16140 }, { "epoch": 24.16317365269461, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1293, "step": 16141 }, { "epoch": 24.164670658682635, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 16142 }, { "epoch": 24.16616766467066, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1242, "step": 16143 }, { "epoch": 24.16766467065868, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1328, "step": 16144 }, { "epoch": 24.169161676646706, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1295, "step": 16145 }, { "epoch": 24.17065868263473, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16146 }, { "epoch": 24.172155688622755, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1254, "step": 16147 }, { "epoch": 24.17365269461078, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 16148 }, { "epoch": 24.1751497005988, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1276, "step": 16149 }, { "epoch": 24.176646706586826, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1334, "step": 16150 }, { "epoch": 24.17814371257485, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1311, "step": 16151 }, { "epoch": 24.179640718562876, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.14, "step": 16152 }, { "epoch": 24.181137724550897, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1308, "step": 16153 }, { "epoch": 24.182634730538922, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1281, "step": 16154 }, { "epoch": 24.184131736526947, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 16155 }, { "epoch": 24.18562874251497, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 16156 }, { "epoch": 24.187125748502993, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1324, "step": 16157 }, { "epoch": 24.188622754491018, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1277, "step": 16158 }, { "epoch": 24.190119760479043, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1328, "step": 16159 }, { "epoch": 24.191616766467067, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1342, "step": 16160 }, { "epoch": 24.19311377245509, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 16161 }, { "epoch": 24.194610778443113, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1346, "step": 16162 }, { "epoch": 24.19610778443114, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.132, "step": 16163 }, { "epoch": 24.197604790419163, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1297, "step": 16164 }, { "epoch": 24.199101796407184, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16165 }, { "epoch": 24.20059880239521, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1273, "step": 16166 }, { "epoch": 24.202095808383234, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.134, "step": 16167 }, { "epoch": 24.20359281437126, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16168 }, { "epoch": 24.20508982035928, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.13, "step": 16169 }, { "epoch": 24.206586826347305, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1343, "step": 16170 }, { "epoch": 24.20808383233533, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.135, "step": 16171 }, { "epoch": 24.209580838323355, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1294, "step": 16172 }, { "epoch": 24.211077844311376, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1355, "step": 16173 }, { "epoch": 24.2125748502994, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1301, "step": 16174 }, { "epoch": 24.214071856287426, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1366, "step": 16175 }, { "epoch": 24.21556886227545, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.135, "step": 16176 }, { "epoch": 24.21706586826347, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1312, "step": 16177 }, { "epoch": 24.218562874251496, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1298, "step": 16178 }, { "epoch": 24.22005988023952, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.132, "step": 16179 }, { "epoch": 24.221556886227546, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1338, "step": 16180 }, { "epoch": 24.223053892215567, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1297, "step": 16181 }, { "epoch": 24.224550898203592, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.125, "step": 16182 }, { "epoch": 24.226047904191617, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1333, "step": 16183 }, { "epoch": 24.227544910179642, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1332, "step": 16184 }, { "epoch": 24.229041916167663, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1321, "step": 16185 }, { "epoch": 24.230538922155688, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1337, "step": 16186 }, { "epoch": 24.232035928143713, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.132, "step": 16187 }, { "epoch": 24.233532934131738, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1326, "step": 16188 }, { "epoch": 24.23502994011976, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 16189 }, { "epoch": 24.236526946107784, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 16190 }, { "epoch": 24.23802395209581, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1382, "step": 16191 }, { "epoch": 24.239520958083833, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1286, "step": 16192 }, { "epoch": 24.241017964071855, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1317, "step": 16193 }, { "epoch": 24.24251497005988, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1293, "step": 16194 }, { "epoch": 24.244011976047904, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1377, "step": 16195 }, { "epoch": 24.24550898203593, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1315, "step": 16196 }, { "epoch": 24.24700598802395, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1366, "step": 16197 }, { "epoch": 24.248502994011975, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1284, "step": 16198 }, { "epoch": 24.25, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 16199 }, { "epoch": 24.251497005988025, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1313, "step": 16200 }, { "epoch": 24.25299401197605, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1304, "step": 16201 }, { "epoch": 24.25449101796407, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1308, "step": 16202 }, { "epoch": 24.255988023952096, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16203 }, { "epoch": 24.25748502994012, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1257, "step": 16204 }, { "epoch": 24.258982035928145, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1299, "step": 16205 }, { "epoch": 24.260479041916167, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1387, "step": 16206 }, { "epoch": 24.26197604790419, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1308, "step": 16207 }, { "epoch": 24.263473053892216, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1345, "step": 16208 }, { "epoch": 24.26497005988024, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.131, "step": 16209 }, { "epoch": 24.266467065868262, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16210 }, { "epoch": 24.267964071856287, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1288, "step": 16211 }, { "epoch": 24.269461077844312, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1329, "step": 16212 }, { "epoch": 24.270958083832337, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1296, "step": 16213 }, { "epoch": 24.272455089820358, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1262, "step": 16214 }, { "epoch": 24.273952095808383, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1285, "step": 16215 }, { "epoch": 24.275449101796408, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1234, "step": 16216 }, { "epoch": 24.276946107784433, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1302, "step": 16217 }, { "epoch": 24.278443113772454, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1378, "step": 16218 }, { "epoch": 24.27994011976048, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1323, "step": 16219 }, { "epoch": 24.281437125748504, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.132, "step": 16220 }, { "epoch": 24.28293413173653, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1327, "step": 16221 }, { "epoch": 24.28443113772455, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1266, "step": 16222 }, { "epoch": 24.285928143712574, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1325, "step": 16223 }, { "epoch": 24.2874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1324, "step": 16224 }, { "epoch": 24.288922155688624, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 16225 }, { "epoch": 24.290419161676645, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1306, "step": 16226 }, { "epoch": 24.29191616766467, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1364, "step": 16227 }, { "epoch": 24.293413173652695, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1408, "step": 16228 }, { "epoch": 24.29491017964072, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1328, "step": 16229 }, { "epoch": 24.29640718562874, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1278, "step": 16230 }, { "epoch": 24.297904191616766, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1267, "step": 16231 }, { "epoch": 24.29940119760479, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1358, "step": 16232 }, { "epoch": 24.300898203592816, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1317, "step": 16233 }, { "epoch": 24.302395209580837, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1323, "step": 16234 }, { "epoch": 24.30389221556886, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1279, "step": 16235 }, { "epoch": 24.305389221556887, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1399, "step": 16236 }, { "epoch": 24.30688622754491, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1268, "step": 16237 }, { "epoch": 24.308383233532933, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1271, "step": 16238 }, { "epoch": 24.309880239520957, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 16239 }, { "epoch": 24.311377245508982, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1354, "step": 16240 }, { "epoch": 24.312874251497007, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1338, "step": 16241 }, { "epoch": 24.31437125748503, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.133, "step": 16242 }, { "epoch": 24.315868263473053, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1351, "step": 16243 }, { "epoch": 24.317365269461078, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1308, "step": 16244 }, { "epoch": 24.318862275449103, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1284, "step": 16245 }, { "epoch": 24.320359281437124, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.131, "step": 16246 }, { "epoch": 24.32185628742515, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1301, "step": 16247 }, { "epoch": 24.323353293413174, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1278, "step": 16248 }, { "epoch": 24.3248502994012, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1248, "step": 16249 }, { "epoch": 24.32634730538922, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1233, "step": 16250 }, { "epoch": 24.327844311377245, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.132, "step": 16251 }, { "epoch": 24.32934131736527, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1326, "step": 16252 }, { "epoch": 24.330838323353294, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16253 }, { "epoch": 24.33233532934132, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.129, "step": 16254 }, { "epoch": 24.33383233532934, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.138, "step": 16255 }, { "epoch": 24.335329341317365, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1275, "step": 16256 }, { "epoch": 24.33682634730539, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1324, "step": 16257 }, { "epoch": 24.338323353293415, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1315, "step": 16258 }, { "epoch": 24.339820359281436, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1333, "step": 16259 }, { "epoch": 24.34131736526946, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1218, "step": 16260 }, { "epoch": 24.342814371257486, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 16261 }, { "epoch": 24.34431137724551, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1271, "step": 16262 }, { "epoch": 24.345808383233532, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1349, "step": 16263 }, { "epoch": 24.347305389221557, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.135, "step": 16264 }, { "epoch": 24.34880239520958, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1379, "step": 16265 }, { "epoch": 24.350299401197606, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1309, "step": 16266 }, { "epoch": 24.351796407185628, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1334, "step": 16267 }, { "epoch": 24.353293413173652, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1338, "step": 16268 }, { "epoch": 24.354790419161677, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1218, "step": 16269 }, { "epoch": 24.356287425149702, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1315, "step": 16270 }, { "epoch": 24.357784431137723, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1316, "step": 16271 }, { "epoch": 24.35928143712575, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 16272 }, { "epoch": 24.360778443113773, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1348, "step": 16273 }, { "epoch": 24.362275449101798, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 16274 }, { "epoch": 24.36377245508982, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1319, "step": 16275 }, { "epoch": 24.365269461077844, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1359, "step": 16276 }, { "epoch": 24.36676646706587, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.133, "step": 16277 }, { "epoch": 24.368263473053894, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1357, "step": 16278 }, { "epoch": 24.369760479041915, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1316, "step": 16279 }, { "epoch": 24.37125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1324, "step": 16280 }, { "epoch": 24.372754491017965, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.132, "step": 16281 }, { "epoch": 24.37425149700599, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1287, "step": 16282 }, { "epoch": 24.37574850299401, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1388, "step": 16283 }, { "epoch": 24.377245508982035, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1295, "step": 16284 }, { "epoch": 24.37874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1332, "step": 16285 }, { "epoch": 24.380239520958085, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1418, "step": 16286 }, { "epoch": 24.381736526946106, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1325, "step": 16287 }, { "epoch": 24.38323353293413, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1295, "step": 16288 }, { "epoch": 24.384730538922156, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1377, "step": 16289 }, { "epoch": 24.38622754491018, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1309, "step": 16290 }, { "epoch": 24.387724550898202, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1366, "step": 16291 }, { "epoch": 24.389221556886227, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16292 }, { "epoch": 24.39071856287425, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1268, "step": 16293 }, { "epoch": 24.392215568862277, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16294 }, { "epoch": 24.393712574850298, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1316, "step": 16295 }, { "epoch": 24.395209580838323, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1305, "step": 16296 }, { "epoch": 24.396706586826348, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1285, "step": 16297 }, { "epoch": 24.398203592814372, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1339, "step": 16298 }, { "epoch": 24.399700598802394, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1318, "step": 16299 }, { "epoch": 24.40119760479042, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1291, "step": 16300 }, { "epoch": 24.402694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1276, "step": 16301 }, { "epoch": 24.404191616766468, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16302 }, { "epoch": 24.40568862275449, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.13, "step": 16303 }, { "epoch": 24.407185628742514, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1296, "step": 16304 }, { "epoch": 24.40868263473054, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1272, "step": 16305 }, { "epoch": 24.410179640718564, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1334, "step": 16306 }, { "epoch": 24.411676646706585, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1299, "step": 16307 }, { "epoch": 24.41317365269461, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 16308 }, { "epoch": 24.414670658682635, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.131, "step": 16309 }, { "epoch": 24.41616766467066, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1375, "step": 16310 }, { "epoch": 24.41766467065868, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1304, "step": 16311 }, { "epoch": 24.419161676646706, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1371, "step": 16312 }, { "epoch": 24.42065868263473, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16313 }, { "epoch": 24.422155688622755, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1357, "step": 16314 }, { "epoch": 24.42365269461078, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1323, "step": 16315 }, { "epoch": 24.4251497005988, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1318, "step": 16316 }, { "epoch": 24.426646706586826, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1295, "step": 16317 }, { "epoch": 24.42814371257485, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 16318 }, { "epoch": 24.429640718562876, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1331, "step": 16319 }, { "epoch": 24.431137724550897, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 16320 }, { "epoch": 24.432634730538922, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1335, "step": 16321 }, { "epoch": 24.434131736526947, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.137, "step": 16322 }, { "epoch": 24.43562874251497, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1308, "step": 16323 }, { "epoch": 24.437125748502993, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1263, "step": 16324 }, { "epoch": 24.438622754491018, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1303, "step": 16325 }, { "epoch": 24.440119760479043, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 16326 }, { "epoch": 24.441616766467067, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1323, "step": 16327 }, { "epoch": 24.44311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1241, "step": 16328 }, { "epoch": 24.444610778443113, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1342, "step": 16329 }, { "epoch": 24.44610778443114, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1315, "step": 16330 }, { "epoch": 24.447604790419163, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1342, "step": 16331 }, { "epoch": 24.449101796407184, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1277, "step": 16332 }, { "epoch": 24.45059880239521, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1267, "step": 16333 }, { "epoch": 24.452095808383234, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1289, "step": 16334 }, { "epoch": 24.45359281437126, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.137, "step": 16335 }, { "epoch": 24.45508982035928, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1286, "step": 16336 }, { "epoch": 24.456586826347305, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1351, "step": 16337 }, { "epoch": 24.45808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1293, "step": 16338 }, { "epoch": 24.459580838323355, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.133, "step": 16339 }, { "epoch": 24.461077844311376, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.135, "step": 16340 }, { "epoch": 24.4625748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1297, "step": 16341 }, { "epoch": 24.464071856287426, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1347, "step": 16342 }, { "epoch": 24.46556886227545, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.134, "step": 16343 }, { "epoch": 24.46706586826347, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1314, "step": 16344 }, { "epoch": 24.468562874251496, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16345 }, { "epoch": 24.47005988023952, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1301, "step": 16346 }, { "epoch": 24.471556886227546, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1275, "step": 16347 }, { "epoch": 24.473053892215567, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1307, "step": 16348 }, { "epoch": 24.474550898203592, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1275, "step": 16349 }, { "epoch": 24.476047904191617, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1284, "step": 16350 }, { "epoch": 24.477544910179642, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1351, "step": 16351 }, { "epoch": 24.479041916167663, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1312, "step": 16352 }, { "epoch": 24.480538922155688, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1287, "step": 16353 }, { "epoch": 24.482035928143713, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 16354 }, { "epoch": 24.483532934131738, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1326, "step": 16355 }, { "epoch": 24.48502994011976, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1322, "step": 16356 }, { "epoch": 24.486526946107784, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 16357 }, { "epoch": 24.48802395209581, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1235, "step": 16358 }, { "epoch": 24.489520958083833, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1293, "step": 16359 }, { "epoch": 24.491017964071855, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1311, "step": 16360 }, { "epoch": 24.49251497005988, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.135, "step": 16361 }, { "epoch": 24.494011976047904, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16362 }, { "epoch": 24.49550898203593, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1331, "step": 16363 }, { "epoch": 24.49700598802395, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1357, "step": 16364 }, { "epoch": 24.498502994011975, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1312, "step": 16365 }, { "epoch": 24.5, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.129, "step": 16366 }, { "epoch": 24.501497005988025, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.128, "step": 16367 }, { "epoch": 24.50299401197605, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16368 }, { "epoch": 24.50449101796407, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1376, "step": 16369 }, { "epoch": 24.505988023952096, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1236, "step": 16370 }, { "epoch": 24.50748502994012, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16371 }, { "epoch": 24.508982035928145, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1359, "step": 16372 }, { "epoch": 24.510479041916167, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.13, "step": 16373 }, { "epoch": 24.51197604790419, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 16374 }, { "epoch": 24.513473053892216, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1351, "step": 16375 }, { "epoch": 24.51497005988024, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1366, "step": 16376 }, { "epoch": 24.516467065868262, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1292, "step": 16377 }, { "epoch": 24.517964071856287, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1368, "step": 16378 }, { "epoch": 24.519461077844312, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1278, "step": 16379 }, { "epoch": 24.520958083832337, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1317, "step": 16380 }, { "epoch": 24.522455089820358, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1233, "step": 16381 }, { "epoch": 24.523952095808383, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1365, "step": 16382 }, { "epoch": 24.525449101796408, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1354, "step": 16383 }, { "epoch": 24.526946107784433, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1375, "step": 16384 }, { "epoch": 24.528443113772454, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1379, "step": 16385 }, { "epoch": 24.52994011976048, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 16386 }, { "epoch": 24.531437125748504, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.13, "step": 16387 }, { "epoch": 24.53293413173653, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1256, "step": 16388 }, { "epoch": 24.53443113772455, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16389 }, { "epoch": 24.535928143712574, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1322, "step": 16390 }, { "epoch": 24.5374251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1245, "step": 16391 }, { "epoch": 24.538922155688624, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.131, "step": 16392 }, { "epoch": 24.540419161676645, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1331, "step": 16393 }, { "epoch": 24.54191616766467, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1312, "step": 16394 }, { "epoch": 24.543413173652695, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1355, "step": 16395 }, { "epoch": 24.54491017964072, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1298, "step": 16396 }, { "epoch": 24.54640718562874, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1319, "step": 16397 }, { "epoch": 24.547904191616766, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.135, "step": 16398 }, { "epoch": 24.54940119760479, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1321, "step": 16399 }, { "epoch": 24.550898203592816, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1305, "step": 16400 }, { "epoch": 24.552395209580837, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1302, "step": 16401 }, { "epoch": 24.55389221556886, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 16402 }, { "epoch": 24.555389221556887, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1299, "step": 16403 }, { "epoch": 24.55688622754491, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1394, "step": 16404 }, { "epoch": 24.558383233532933, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1275, "step": 16405 }, { "epoch": 24.559880239520957, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1351, "step": 16406 }, { "epoch": 24.561377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1283, "step": 16407 }, { "epoch": 24.562874251497007, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1256, "step": 16408 }, { "epoch": 24.56437125748503, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 16409 }, { "epoch": 24.565868263473053, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1301, "step": 16410 }, { "epoch": 24.567365269461078, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1291, "step": 16411 }, { "epoch": 24.568862275449103, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.136, "step": 16412 }, { "epoch": 24.570359281437124, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1335, "step": 16413 }, { "epoch": 24.57185628742515, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1313, "step": 16414 }, { "epoch": 24.573353293413174, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1336, "step": 16415 }, { "epoch": 24.5748502994012, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 16416 }, { "epoch": 24.57634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1408, "step": 16417 }, { "epoch": 24.577844311377245, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 16418 }, { "epoch": 24.57934131736527, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1288, "step": 16419 }, { "epoch": 24.580838323353294, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1346, "step": 16420 }, { "epoch": 24.58233532934132, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1391, "step": 16421 }, { "epoch": 24.58383233532934, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1282, "step": 16422 }, { "epoch": 24.585329341317365, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1349, "step": 16423 }, { "epoch": 24.58682634730539, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1326, "step": 16424 }, { "epoch": 24.58832335329341, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1351, "step": 16425 }, { "epoch": 24.589820359281436, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16426 }, { "epoch": 24.59131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.13, "step": 16427 }, { "epoch": 24.592814371257486, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1369, "step": 16428 }, { "epoch": 24.59431137724551, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.133, "step": 16429 }, { "epoch": 24.595808383233532, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1347, "step": 16430 }, { "epoch": 24.597305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1326, "step": 16431 }, { "epoch": 24.59880239520958, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 16432 }, { "epoch": 24.600299401197606, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16433 }, { "epoch": 24.601796407185628, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1307, "step": 16434 }, { "epoch": 24.603293413173652, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1327, "step": 16435 }, { "epoch": 24.604790419161677, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1296, "step": 16436 }, { "epoch": 24.606287425149702, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1364, "step": 16437 }, { "epoch": 24.607784431137723, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 16438 }, { "epoch": 24.60928143712575, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16439 }, { "epoch": 24.610778443113773, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1278, "step": 16440 }, { "epoch": 24.612275449101798, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 16441 }, { "epoch": 24.61377245508982, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 16442 }, { "epoch": 24.615269461077844, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1354, "step": 16443 }, { "epoch": 24.61676646706587, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 16444 }, { "epoch": 24.618263473053894, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16445 }, { "epoch": 24.619760479041915, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 16446 }, { "epoch": 24.62125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1327, "step": 16447 }, { "epoch": 24.622754491017965, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1234, "step": 16448 }, { "epoch": 24.62425149700599, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1284, "step": 16449 }, { "epoch": 24.62574850299401, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1232, "step": 16450 }, { "epoch": 24.627245508982035, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1368, "step": 16451 }, { "epoch": 24.62874251497006, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1279, "step": 16452 }, { "epoch": 24.630239520958085, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1312, "step": 16453 }, { "epoch": 24.631736526946106, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1264, "step": 16454 }, { "epoch": 24.63323353293413, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1352, "step": 16455 }, { "epoch": 24.634730538922156, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1266, "step": 16456 }, { "epoch": 24.63622754491018, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.13, "step": 16457 }, { "epoch": 24.637724550898202, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1387, "step": 16458 }, { "epoch": 24.639221556886227, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1427, "step": 16459 }, { "epoch": 24.64071856287425, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1294, "step": 16460 }, { "epoch": 24.642215568862277, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1283, "step": 16461 }, { "epoch": 24.643712574850298, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.127, "step": 16462 }, { "epoch": 24.645209580838323, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1338, "step": 16463 }, { "epoch": 24.646706586826348, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1334, "step": 16464 }, { "epoch": 24.648203592814372, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1369, "step": 16465 }, { "epoch": 24.649700598802394, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1355, "step": 16466 }, { "epoch": 24.65119760479042, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.126, "step": 16467 }, { "epoch": 24.652694610778443, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1279, "step": 16468 }, { "epoch": 24.654191616766468, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.126, "step": 16469 }, { "epoch": 24.65568862275449, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1289, "step": 16470 }, { "epoch": 24.657185628742514, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1343, "step": 16471 }, { "epoch": 24.65868263473054, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1372, "step": 16472 }, { "epoch": 24.660179640718564, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1258, "step": 16473 }, { "epoch": 24.66167664670659, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1281, "step": 16474 }, { "epoch": 24.66317365269461, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1355, "step": 16475 }, { "epoch": 24.664670658682635, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16476 }, { "epoch": 24.66616766467066, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.126, "step": 16477 }, { "epoch": 24.66766467065868, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1405, "step": 16478 }, { "epoch": 24.669161676646706, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 16479 }, { "epoch": 24.67065868263473, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1366, "step": 16480 }, { "epoch": 24.672155688622755, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1308, "step": 16481 }, { "epoch": 24.67365269461078, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 16482 }, { "epoch": 24.6751497005988, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1287, "step": 16483 }, { "epoch": 24.676646706586826, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1346, "step": 16484 }, { "epoch": 24.67814371257485, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1283, "step": 16485 }, { "epoch": 24.679640718562876, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1326, "step": 16486 }, { "epoch": 24.681137724550897, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16487 }, { "epoch": 24.682634730538922, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1354, "step": 16488 }, { "epoch": 24.684131736526947, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.127, "step": 16489 }, { "epoch": 24.68562874251497, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1304, "step": 16490 }, { "epoch": 24.687125748502993, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.129, "step": 16491 }, { "epoch": 24.688622754491018, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1301, "step": 16492 }, { "epoch": 24.690119760479043, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1353, "step": 16493 }, { "epoch": 24.691616766467067, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1327, "step": 16494 }, { "epoch": 24.69311377245509, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1362, "step": 16495 }, { "epoch": 24.694610778443113, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1291, "step": 16496 }, { "epoch": 24.69610778443114, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1285, "step": 16497 }, { "epoch": 24.697604790419163, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1361, "step": 16498 }, { "epoch": 24.699101796407184, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 16499 }, { "epoch": 24.70059880239521, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16500 }, { "epoch": 24.702095808383234, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1296, "step": 16501 }, { "epoch": 24.70359281437126, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1275, "step": 16502 }, { "epoch": 24.70508982035928, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1326, "step": 16503 }, { "epoch": 24.706586826347305, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1373, "step": 16504 }, { "epoch": 24.70808383233533, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1286, "step": 16505 }, { "epoch": 24.709580838323355, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16506 }, { "epoch": 24.711077844311376, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1304, "step": 16507 }, { "epoch": 24.7125748502994, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1391, "step": 16508 }, { "epoch": 24.714071856287426, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1317, "step": 16509 }, { "epoch": 24.71556886227545, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1309, "step": 16510 }, { "epoch": 24.71706586826347, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1255, "step": 16511 }, { "epoch": 24.718562874251496, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1279, "step": 16512 }, { "epoch": 24.72005988023952, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1267, "step": 16513 }, { "epoch": 24.721556886227546, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1263, "step": 16514 }, { "epoch": 24.723053892215567, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1301, "step": 16515 }, { "epoch": 24.724550898203592, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.126, "step": 16516 }, { "epoch": 24.726047904191617, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1292, "step": 16517 }, { "epoch": 24.727544910179642, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 16518 }, { "epoch": 24.729041916167663, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.126, "step": 16519 }, { "epoch": 24.730538922155688, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1345, "step": 16520 }, { "epoch": 24.732035928143713, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1329, "step": 16521 }, { "epoch": 24.733532934131738, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16522 }, { "epoch": 24.73502994011976, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1347, "step": 16523 }, { "epoch": 24.736526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1328, "step": 16524 }, { "epoch": 24.73802395209581, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1369, "step": 16525 }, { "epoch": 24.739520958083833, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1289, "step": 16526 }, { "epoch": 24.741017964071855, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1331, "step": 16527 }, { "epoch": 24.74251497005988, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1363, "step": 16528 }, { "epoch": 24.744011976047904, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1382, "step": 16529 }, { "epoch": 24.74550898203593, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.133, "step": 16530 }, { "epoch": 24.74700598802395, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1335, "step": 16531 }, { "epoch": 24.748502994011975, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1324, "step": 16532 }, { "epoch": 24.75, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1332, "step": 16533 }, { "epoch": 24.751497005988025, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 16534 }, { "epoch": 24.75299401197605, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1291, "step": 16535 }, { "epoch": 24.75449101796407, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1397, "step": 16536 }, { "epoch": 24.755988023952096, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.13, "step": 16537 }, { "epoch": 24.75748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1363, "step": 16538 }, { "epoch": 24.758982035928145, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1329, "step": 16539 }, { "epoch": 24.760479041916167, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1374, "step": 16540 }, { "epoch": 24.76197604790419, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1284, "step": 16541 }, { "epoch": 24.763473053892216, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1226, "step": 16542 }, { "epoch": 24.76497005988024, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1267, "step": 16543 }, { "epoch": 24.766467065868262, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1299, "step": 16544 }, { "epoch": 24.767964071856287, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1316, "step": 16545 }, { "epoch": 24.769461077844312, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1265, "step": 16546 }, { "epoch": 24.770958083832337, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1351, "step": 16547 }, { "epoch": 24.772455089820358, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.136, "step": 16548 }, { "epoch": 24.773952095808383, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1347, "step": 16549 }, { "epoch": 24.775449101796408, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.132, "step": 16550 }, { "epoch": 24.776946107784433, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 16551 }, { "epoch": 24.778443113772454, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 16552 }, { "epoch": 24.77994011976048, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1296, "step": 16553 }, { "epoch": 24.781437125748504, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16554 }, { "epoch": 24.78293413173653, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 16555 }, { "epoch": 24.78443113772455, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1348, "step": 16556 }, { "epoch": 24.785928143712574, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1298, "step": 16557 }, { "epoch": 24.7874251497006, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1346, "step": 16558 }, { "epoch": 24.788922155688624, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1344, "step": 16559 }, { "epoch": 24.790419161676645, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1229, "step": 16560 }, { "epoch": 24.79191616766467, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1299, "step": 16561 }, { "epoch": 24.793413173652695, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1395, "step": 16562 }, { "epoch": 24.79491017964072, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1364, "step": 16563 }, { "epoch": 24.79640718562874, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1314, "step": 16564 }, { "epoch": 24.797904191616766, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.127, "step": 16565 }, { "epoch": 24.79940119760479, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1307, "step": 16566 }, { "epoch": 24.800898203592816, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1417, "step": 16567 }, { "epoch": 24.802395209580837, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1295, "step": 16568 }, { "epoch": 24.80389221556886, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.125, "step": 16569 }, { "epoch": 24.805389221556887, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1296, "step": 16570 }, { "epoch": 24.80688622754491, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1301, "step": 16571 }, { "epoch": 24.808383233532933, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1286, "step": 16572 }, { "epoch": 24.809880239520957, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16573 }, { "epoch": 24.811377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1343, "step": 16574 }, { "epoch": 24.812874251497007, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1302, "step": 16575 }, { "epoch": 24.81437125748503, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1326, "step": 16576 }, { "epoch": 24.815868263473053, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16577 }, { "epoch": 24.817365269461078, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1328, "step": 16578 }, { "epoch": 24.818862275449103, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1294, "step": 16579 }, { "epoch": 24.820359281437124, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1296, "step": 16580 }, { "epoch": 24.82185628742515, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1323, "step": 16581 }, { "epoch": 24.823353293413174, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1335, "step": 16582 }, { "epoch": 24.8248502994012, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.125, "step": 16583 }, { "epoch": 24.82634730538922, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1285, "step": 16584 }, { "epoch": 24.827844311377245, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1341, "step": 16585 }, { "epoch": 24.82934131736527, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16586 }, { "epoch": 24.830838323353294, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1311, "step": 16587 }, { "epoch": 24.83233532934132, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1216, "step": 16588 }, { "epoch": 24.83383233532934, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.134, "step": 16589 }, { "epoch": 24.835329341317365, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.136, "step": 16590 }, { "epoch": 24.83682634730539, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1226, "step": 16591 }, { "epoch": 24.83832335329341, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1269, "step": 16592 }, { "epoch": 24.839820359281436, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1312, "step": 16593 }, { "epoch": 24.84131736526946, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1373, "step": 16594 }, { "epoch": 24.842814371257486, "grad_norm": 0.2158203125, "learning_rate": 0.0008, "loss": 1.1371, "step": 16595 }, { "epoch": 24.84431137724551, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1348, "step": 16596 }, { "epoch": 24.845808383233532, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 16597 }, { "epoch": 24.847305389221557, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.132, "step": 16598 }, { "epoch": 24.84880239520958, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1322, "step": 16599 }, { "epoch": 24.850299401197606, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1357, "step": 16600 }, { "epoch": 24.851796407185628, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1265, "step": 16601 }, { "epoch": 24.853293413173652, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1291, "step": 16602 }, { "epoch": 24.854790419161677, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1311, "step": 16603 }, { "epoch": 24.856287425149702, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1322, "step": 16604 }, { "epoch": 24.857784431137723, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1277, "step": 16605 }, { "epoch": 24.85928143712575, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1359, "step": 16606 }, { "epoch": 24.860778443113773, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.127, "step": 16607 }, { "epoch": 24.862275449101798, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1371, "step": 16608 }, { "epoch": 24.86377245508982, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1261, "step": 16609 }, { "epoch": 24.865269461077844, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1325, "step": 16610 }, { "epoch": 24.86676646706587, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1262, "step": 16611 }, { "epoch": 24.868263473053894, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.124, "step": 16612 }, { "epoch": 24.869760479041915, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 16613 }, { "epoch": 24.87125748502994, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.131, "step": 16614 }, { "epoch": 24.872754491017965, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1275, "step": 16615 }, { "epoch": 24.87425149700599, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1371, "step": 16616 }, { "epoch": 24.87574850299401, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1221, "step": 16617 }, { "epoch": 24.877245508982035, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1313, "step": 16618 }, { "epoch": 24.87874251497006, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1273, "step": 16619 }, { "epoch": 24.880239520958085, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.128, "step": 16620 }, { "epoch": 24.881736526946106, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1309, "step": 16621 }, { "epoch": 24.88323353293413, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 16622 }, { "epoch": 24.884730538922156, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1328, "step": 16623 }, { "epoch": 24.88622754491018, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1279, "step": 16624 }, { "epoch": 24.887724550898202, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.125, "step": 16625 }, { "epoch": 24.889221556886227, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.122, "step": 16626 }, { "epoch": 24.89071856287425, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1247, "step": 16627 }, { "epoch": 24.892215568862277, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1303, "step": 16628 }, { "epoch": 24.893712574850298, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1268, "step": 16629 }, { "epoch": 24.895209580838323, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.137, "step": 16630 }, { "epoch": 24.896706586826348, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1307, "step": 16631 }, { "epoch": 24.898203592814372, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16632 }, { "epoch": 24.899700598802394, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1271, "step": 16633 }, { "epoch": 24.90119760479042, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1254, "step": 16634 }, { "epoch": 24.902694610778443, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1335, "step": 16635 }, { "epoch": 24.904191616766468, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1305, "step": 16636 }, { "epoch": 24.90568862275449, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 16637 }, { "epoch": 24.907185628742514, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1285, "step": 16638 }, { "epoch": 24.90868263473054, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1309, "step": 16639 }, { "epoch": 24.910179640718564, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1343, "step": 16640 }, { "epoch": 24.91167664670659, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1342, "step": 16641 }, { "epoch": 24.91317365269461, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1319, "step": 16642 }, { "epoch": 24.914670658682635, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 16643 }, { "epoch": 24.91616766467066, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1365, "step": 16644 }, { "epoch": 24.91766467065868, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1391, "step": 16645 }, { "epoch": 24.919161676646706, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1333, "step": 16646 }, { "epoch": 24.92065868263473, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1291, "step": 16647 }, { "epoch": 24.922155688622755, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1334, "step": 16648 }, { "epoch": 24.92365269461078, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1292, "step": 16649 }, { "epoch": 24.9251497005988, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 16650 }, { "epoch": 24.926646706586826, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1284, "step": 16651 }, { "epoch": 24.92814371257485, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1298, "step": 16652 }, { "epoch": 24.929640718562876, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1277, "step": 16653 }, { "epoch": 24.931137724550897, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16654 }, { "epoch": 24.932634730538922, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1248, "step": 16655 }, { "epoch": 24.934131736526947, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1242, "step": 16656 }, { "epoch": 24.93562874251497, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16657 }, { "epoch": 24.937125748502993, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.13, "step": 16658 }, { "epoch": 24.938622754491018, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1315, "step": 16659 }, { "epoch": 24.940119760479043, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1357, "step": 16660 }, { "epoch": 24.941616766467067, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1276, "step": 16661 }, { "epoch": 24.94311377245509, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1314, "step": 16662 }, { "epoch": 24.944610778443113, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.133, "step": 16663 }, { "epoch": 24.94610778443114, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1302, "step": 16664 }, { "epoch": 24.947604790419163, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1415, "step": 16665 }, { "epoch": 24.949101796407184, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1353, "step": 16666 }, { "epoch": 24.95059880239521, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 16667 }, { "epoch": 24.952095808383234, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1312, "step": 16668 }, { "epoch": 24.95359281437126, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1244, "step": 16669 }, { "epoch": 24.95508982035928, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1316, "step": 16670 }, { "epoch": 24.956586826347305, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1308, "step": 16671 }, { "epoch": 24.95808383233533, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 16672 }, { "epoch": 24.959580838323355, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1353, "step": 16673 }, { "epoch": 24.961077844311376, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1314, "step": 16674 }, { "epoch": 24.9625748502994, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1415, "step": 16675 }, { "epoch": 24.964071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1347, "step": 16676 }, { "epoch": 24.96556886227545, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1303, "step": 16677 }, { "epoch": 24.96706586826347, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1248, "step": 16678 }, { "epoch": 24.968562874251496, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1353, "step": 16679 }, { "epoch": 24.97005988023952, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.133, "step": 16680 }, { "epoch": 24.971556886227546, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1297, "step": 16681 }, { "epoch": 24.973053892215567, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1283, "step": 16682 }, { "epoch": 24.974550898203592, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1239, "step": 16683 }, { "epoch": 24.976047904191617, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1342, "step": 16684 }, { "epoch": 24.977544910179642, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1347, "step": 16685 }, { "epoch": 24.979041916167663, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16686 }, { "epoch": 24.980538922155688, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 16687 }, { "epoch": 24.982035928143713, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 16688 }, { "epoch": 24.983532934131738, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.136, "step": 16689 }, { "epoch": 24.98502994011976, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1225, "step": 16690 }, { "epoch": 24.986526946107784, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1394, "step": 16691 }, { "epoch": 24.98802395209581, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1288, "step": 16692 }, { "epoch": 24.989520958083833, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.133, "step": 16693 }, { "epoch": 24.991017964071855, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 16694 }, { "epoch": 24.99251497005988, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 16695 }, { "epoch": 24.994011976047904, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1326, "step": 16696 }, { "epoch": 24.99550898203593, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1337, "step": 16697 }, { "epoch": 24.99700598802395, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1274, "step": 16698 }, { "epoch": 24.998502994011975, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1336, "step": 16699 }, { "epoch": 25.0, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1276, "step": 16700 }, { "epoch": 25.001497005988025, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1316, "step": 16701 }, { "epoch": 25.00299401197605, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1349, "step": 16702 }, { "epoch": 25.00449101796407, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1254, "step": 16703 }, { "epoch": 25.005988023952096, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1375, "step": 16704 }, { "epoch": 25.00748502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1238, "step": 16705 }, { "epoch": 25.008982035928145, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1344, "step": 16706 }, { "epoch": 25.010479041916167, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1347, "step": 16707 }, { "epoch": 25.01197604790419, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1298, "step": 16708 }, { "epoch": 25.013473053892216, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1357, "step": 16709 }, { "epoch": 25.01497005988024, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1278, "step": 16710 }, { "epoch": 25.016467065868262, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1295, "step": 16711 }, { "epoch": 25.017964071856287, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1348, "step": 16712 }, { "epoch": 25.019461077844312, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1286, "step": 16713 }, { "epoch": 25.020958083832337, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1335, "step": 16714 }, { "epoch": 25.022455089820358, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 16715 }, { "epoch": 25.023952095808383, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1297, "step": 16716 }, { "epoch": 25.025449101796408, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.128, "step": 16717 }, { "epoch": 25.026946107784433, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.134, "step": 16718 }, { "epoch": 25.028443113772454, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1298, "step": 16719 }, { "epoch": 25.02994011976048, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1363, "step": 16720 }, { "epoch": 25.031437125748504, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1333, "step": 16721 }, { "epoch": 25.03293413173653, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1356, "step": 16722 }, { "epoch": 25.03443113772455, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1341, "step": 16723 }, { "epoch": 25.035928143712574, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1289, "step": 16724 }, { "epoch": 25.0374251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 16725 }, { "epoch": 25.038922155688624, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1298, "step": 16726 }, { "epoch": 25.040419161676645, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1301, "step": 16727 }, { "epoch": 25.04191616766467, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1306, "step": 16728 }, { "epoch": 25.043413173652695, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16729 }, { "epoch": 25.04491017964072, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1321, "step": 16730 }, { "epoch": 25.04640718562874, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 16731 }, { "epoch": 25.047904191616766, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1332, "step": 16732 }, { "epoch": 25.04940119760479, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1279, "step": 16733 }, { "epoch": 25.050898203592816, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1341, "step": 16734 }, { "epoch": 25.052395209580837, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1299, "step": 16735 }, { "epoch": 25.05389221556886, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 16736 }, { "epoch": 25.055389221556887, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16737 }, { "epoch": 25.05688622754491, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1336, "step": 16738 }, { "epoch": 25.058383233532933, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 16739 }, { "epoch": 25.059880239520957, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1305, "step": 16740 }, { "epoch": 25.061377245508982, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.13, "step": 16741 }, { "epoch": 25.062874251497007, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1313, "step": 16742 }, { "epoch": 25.06437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1309, "step": 16743 }, { "epoch": 25.065868263473053, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1306, "step": 16744 }, { "epoch": 25.067365269461078, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1291, "step": 16745 }, { "epoch": 25.068862275449103, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16746 }, { "epoch": 25.070359281437124, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1276, "step": 16747 }, { "epoch": 25.07185628742515, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1266, "step": 16748 }, { "epoch": 25.073353293413174, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 16749 }, { "epoch": 25.0748502994012, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16750 }, { "epoch": 25.07634730538922, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1331, "step": 16751 }, { "epoch": 25.077844311377245, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1332, "step": 16752 }, { "epoch": 25.07934131736527, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16753 }, { "epoch": 25.080838323353294, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1265, "step": 16754 }, { "epoch": 25.082335329341316, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 16755 }, { "epoch": 25.08383233532934, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.132, "step": 16756 }, { "epoch": 25.085329341317365, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1296, "step": 16757 }, { "epoch": 25.08682634730539, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1316, "step": 16758 }, { "epoch": 25.088323353293415, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16759 }, { "epoch": 25.089820359281436, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 16760 }, { "epoch": 25.09131736526946, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1282, "step": 16761 }, { "epoch": 25.092814371257486, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1332, "step": 16762 }, { "epoch": 25.09431137724551, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1296, "step": 16763 }, { "epoch": 25.095808383233532, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1324, "step": 16764 }, { "epoch": 25.097305389221557, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1325, "step": 16765 }, { "epoch": 25.09880239520958, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1348, "step": 16766 }, { "epoch": 25.100299401197606, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1407, "step": 16767 }, { "epoch": 25.101796407185628, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1314, "step": 16768 }, { "epoch": 25.103293413173652, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1264, "step": 16769 }, { "epoch": 25.104790419161677, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1301, "step": 16770 }, { "epoch": 25.106287425149702, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1318, "step": 16771 }, { "epoch": 25.107784431137723, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1376, "step": 16772 }, { "epoch": 25.10928143712575, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1371, "step": 16773 }, { "epoch": 25.110778443113773, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1307, "step": 16774 }, { "epoch": 25.112275449101798, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.126, "step": 16775 }, { "epoch": 25.11377245508982, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1388, "step": 16776 }, { "epoch": 25.115269461077844, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1288, "step": 16777 }, { "epoch": 25.11676646706587, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1281, "step": 16778 }, { "epoch": 25.118263473053894, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16779 }, { "epoch": 25.119760479041915, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1292, "step": 16780 }, { "epoch": 25.12125748502994, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 16781 }, { "epoch": 25.122754491017965, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16782 }, { "epoch": 25.12425149700599, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 16783 }, { "epoch": 25.12574850299401, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1362, "step": 16784 }, { "epoch": 25.127245508982035, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1281, "step": 16785 }, { "epoch": 25.12874251497006, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1348, "step": 16786 }, { "epoch": 25.130239520958085, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1308, "step": 16787 }, { "epoch": 25.131736526946106, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1259, "step": 16788 }, { "epoch": 25.13323353293413, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1295, "step": 16789 }, { "epoch": 25.134730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1271, "step": 16790 }, { "epoch": 25.13622754491018, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.13, "step": 16791 }, { "epoch": 25.137724550898202, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1292, "step": 16792 }, { "epoch": 25.139221556886227, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1308, "step": 16793 }, { "epoch": 25.14071856287425, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1315, "step": 16794 }, { "epoch": 25.142215568862277, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16795 }, { "epoch": 25.143712574850298, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1298, "step": 16796 }, { "epoch": 25.145209580838323, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1263, "step": 16797 }, { "epoch": 25.146706586826348, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1275, "step": 16798 }, { "epoch": 25.148203592814372, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1372, "step": 16799 }, { "epoch": 25.149700598802394, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1287, "step": 16800 }, { "epoch": 25.15119760479042, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 16801 }, { "epoch": 25.152694610778443, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1288, "step": 16802 }, { "epoch": 25.154191616766468, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1328, "step": 16803 }, { "epoch": 25.15568862275449, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1328, "step": 16804 }, { "epoch": 25.157185628742514, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1263, "step": 16805 }, { "epoch": 25.15868263473054, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1287, "step": 16806 }, { "epoch": 25.160179640718564, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1336, "step": 16807 }, { "epoch": 25.161676646706585, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1313, "step": 16808 }, { "epoch": 25.16317365269461, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1325, "step": 16809 }, { "epoch": 25.164670658682635, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.125, "step": 16810 }, { "epoch": 25.16616766467066, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1311, "step": 16811 }, { "epoch": 25.16766467065868, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.132, "step": 16812 }, { "epoch": 25.169161676646706, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1323, "step": 16813 }, { "epoch": 25.17065868263473, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1283, "step": 16814 }, { "epoch": 25.172155688622755, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1264, "step": 16815 }, { "epoch": 25.17365269461078, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1295, "step": 16816 }, { "epoch": 25.1751497005988, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1292, "step": 16817 }, { "epoch": 25.176646706586826, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1383, "step": 16818 }, { "epoch": 25.17814371257485, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.131, "step": 16819 }, { "epoch": 25.179640718562876, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1325, "step": 16820 }, { "epoch": 25.181137724550897, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1285, "step": 16821 }, { "epoch": 25.182634730538922, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1315, "step": 16822 }, { "epoch": 25.184131736526947, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1343, "step": 16823 }, { "epoch": 25.18562874251497, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.135, "step": 16824 }, { "epoch": 25.187125748502993, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1235, "step": 16825 }, { "epoch": 25.188622754491018, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1372, "step": 16826 }, { "epoch": 25.190119760479043, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1267, "step": 16827 }, { "epoch": 25.191616766467067, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1294, "step": 16828 }, { "epoch": 25.19311377245509, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1292, "step": 16829 }, { "epoch": 25.194610778443113, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1326, "step": 16830 }, { "epoch": 25.19610778443114, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1241, "step": 16831 }, { "epoch": 25.197604790419163, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1293, "step": 16832 }, { "epoch": 25.199101796407184, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1345, "step": 16833 }, { "epoch": 25.20059880239521, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1348, "step": 16834 }, { "epoch": 25.202095808383234, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1323, "step": 16835 }, { "epoch": 25.20359281437126, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 16836 }, { "epoch": 25.20508982035928, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1271, "step": 16837 }, { "epoch": 25.206586826347305, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1319, "step": 16838 }, { "epoch": 25.20808383233533, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1293, "step": 16839 }, { "epoch": 25.209580838323355, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1265, "step": 16840 }, { "epoch": 25.211077844311376, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1386, "step": 16841 }, { "epoch": 25.2125748502994, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1331, "step": 16842 }, { "epoch": 25.214071856287426, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1351, "step": 16843 }, { "epoch": 25.21556886227545, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1295, "step": 16844 }, { "epoch": 25.21706586826347, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1346, "step": 16845 }, { "epoch": 25.218562874251496, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1382, "step": 16846 }, { "epoch": 25.22005988023952, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 16847 }, { "epoch": 25.221556886227546, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.135, "step": 16848 }, { "epoch": 25.223053892215567, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1346, "step": 16849 }, { "epoch": 25.224550898203592, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.135, "step": 16850 }, { "epoch": 25.226047904191617, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1269, "step": 16851 }, { "epoch": 25.227544910179642, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1318, "step": 16852 }, { "epoch": 25.229041916167663, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1294, "step": 16853 }, { "epoch": 25.230538922155688, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1248, "step": 16854 }, { "epoch": 25.232035928143713, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1367, "step": 16855 }, { "epoch": 25.233532934131738, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1358, "step": 16856 }, { "epoch": 25.23502994011976, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.123, "step": 16857 }, { "epoch": 25.236526946107784, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1298, "step": 16858 }, { "epoch": 25.23802395209581, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1261, "step": 16859 }, { "epoch": 25.239520958083833, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1335, "step": 16860 }, { "epoch": 25.241017964071855, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1282, "step": 16861 }, { "epoch": 25.24251497005988, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 16862 }, { "epoch": 25.244011976047904, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1352, "step": 16863 }, { "epoch": 25.24550898203593, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1285, "step": 16864 }, { "epoch": 25.24700598802395, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1258, "step": 16865 }, { "epoch": 25.248502994011975, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16866 }, { "epoch": 25.25, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1288, "step": 16867 }, { "epoch": 25.251497005988025, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 16868 }, { "epoch": 25.25299401197605, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1359, "step": 16869 }, { "epoch": 25.25449101796407, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1259, "step": 16870 }, { "epoch": 25.255988023952096, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1273, "step": 16871 }, { "epoch": 25.25748502994012, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1357, "step": 16872 }, { "epoch": 25.258982035928145, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1213, "step": 16873 }, { "epoch": 25.260479041916167, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1328, "step": 16874 }, { "epoch": 25.26197604790419, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1249, "step": 16875 }, { "epoch": 25.263473053892216, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1282, "step": 16876 }, { "epoch": 25.26497005988024, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1326, "step": 16877 }, { "epoch": 25.266467065868262, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1355, "step": 16878 }, { "epoch": 25.267964071856287, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1308, "step": 16879 }, { "epoch": 25.269461077844312, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1336, "step": 16880 }, { "epoch": 25.270958083832337, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.132, "step": 16881 }, { "epoch": 25.272455089820358, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16882 }, { "epoch": 25.273952095808383, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1322, "step": 16883 }, { "epoch": 25.275449101796408, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1285, "step": 16884 }, { "epoch": 25.276946107784433, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1277, "step": 16885 }, { "epoch": 25.278443113772454, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1329, "step": 16886 }, { "epoch": 25.27994011976048, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1236, "step": 16887 }, { "epoch": 25.281437125748504, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1322, "step": 16888 }, { "epoch": 25.28293413173653, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1283, "step": 16889 }, { "epoch": 25.28443113772455, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1338, "step": 16890 }, { "epoch": 25.285928143712574, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1252, "step": 16891 }, { "epoch": 25.2874251497006, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1356, "step": 16892 }, { "epoch": 25.288922155688624, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1237, "step": 16893 }, { "epoch": 25.290419161676645, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 16894 }, { "epoch": 25.29191616766467, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1259, "step": 16895 }, { "epoch": 25.293413173652695, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1321, "step": 16896 }, { "epoch": 25.29491017964072, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 16897 }, { "epoch": 25.29640718562874, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 16898 }, { "epoch": 25.297904191616766, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1316, "step": 16899 }, { "epoch": 25.29940119760479, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1298, "step": 16900 }, { "epoch": 25.300898203592816, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1268, "step": 16901 }, { "epoch": 25.302395209580837, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1319, "step": 16902 }, { "epoch": 25.30389221556886, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16903 }, { "epoch": 25.305389221556887, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1363, "step": 16904 }, { "epoch": 25.30688622754491, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16905 }, { "epoch": 25.308383233532933, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1255, "step": 16906 }, { "epoch": 25.309880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1249, "step": 16907 }, { "epoch": 25.311377245508982, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 16908 }, { "epoch": 25.312874251497007, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1265, "step": 16909 }, { "epoch": 25.31437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1255, "step": 16910 }, { "epoch": 25.315868263473053, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1327, "step": 16911 }, { "epoch": 25.317365269461078, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1253, "step": 16912 }, { "epoch": 25.318862275449103, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16913 }, { "epoch": 25.320359281437124, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16914 }, { "epoch": 25.32185628742515, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1326, "step": 16915 }, { "epoch": 25.323353293413174, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1315, "step": 16916 }, { "epoch": 25.3248502994012, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1278, "step": 16917 }, { "epoch": 25.32634730538922, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.129, "step": 16918 }, { "epoch": 25.327844311377245, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 16919 }, { "epoch": 25.32934131736527, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1361, "step": 16920 }, { "epoch": 25.330838323353294, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.134, "step": 16921 }, { "epoch": 25.33233532934132, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1351, "step": 16922 }, { "epoch": 25.33383233532934, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1295, "step": 16923 }, { "epoch": 25.335329341317365, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1301, "step": 16924 }, { "epoch": 25.33682634730539, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1295, "step": 16925 }, { "epoch": 25.338323353293415, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1343, "step": 16926 }, { "epoch": 25.339820359281436, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1388, "step": 16927 }, { "epoch": 25.34131736526946, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1315, "step": 16928 }, { "epoch": 25.342814371257486, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1326, "step": 16929 }, { "epoch": 25.34431137724551, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1296, "step": 16930 }, { "epoch": 25.345808383233532, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1277, "step": 16931 }, { "epoch": 25.347305389221557, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1288, "step": 16932 }, { "epoch": 25.34880239520958, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1344, "step": 16933 }, { "epoch": 25.350299401197606, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.132, "step": 16934 }, { "epoch": 25.351796407185628, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1297, "step": 16935 }, { "epoch": 25.353293413173652, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1327, "step": 16936 }, { "epoch": 25.354790419161677, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1306, "step": 16937 }, { "epoch": 25.356287425149702, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1175, "step": 16938 }, { "epoch": 25.357784431137723, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1326, "step": 16939 }, { "epoch": 25.35928143712575, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1332, "step": 16940 }, { "epoch": 25.360778443113773, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.124, "step": 16941 }, { "epoch": 25.362275449101798, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1351, "step": 16942 }, { "epoch": 25.36377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.131, "step": 16943 }, { "epoch": 25.365269461077844, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 16944 }, { "epoch": 25.36676646706587, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1302, "step": 16945 }, { "epoch": 25.368263473053894, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1313, "step": 16946 }, { "epoch": 25.369760479041915, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1378, "step": 16947 }, { "epoch": 25.37125748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1338, "step": 16948 }, { "epoch": 25.372754491017965, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1336, "step": 16949 }, { "epoch": 25.37425149700599, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16950 }, { "epoch": 25.37574850299401, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1314, "step": 16951 }, { "epoch": 25.377245508982035, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1418, "step": 16952 }, { "epoch": 25.37874251497006, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.133, "step": 16953 }, { "epoch": 25.380239520958085, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1282, "step": 16954 }, { "epoch": 25.381736526946106, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1269, "step": 16955 }, { "epoch": 25.38323353293413, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1308, "step": 16956 }, { "epoch": 25.384730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1264, "step": 16957 }, { "epoch": 25.38622754491018, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1279, "step": 16958 }, { "epoch": 25.387724550898202, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1354, "step": 16959 }, { "epoch": 25.389221556886227, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1263, "step": 16960 }, { "epoch": 25.39071856287425, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1293, "step": 16961 }, { "epoch": 25.392215568862277, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1316, "step": 16962 }, { "epoch": 25.393712574850298, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1294, "step": 16963 }, { "epoch": 25.395209580838323, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 16964 }, { "epoch": 25.396706586826348, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 16965 }, { "epoch": 25.398203592814372, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1264, "step": 16966 }, { "epoch": 25.399700598802394, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.13, "step": 16967 }, { "epoch": 25.40119760479042, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1349, "step": 16968 }, { "epoch": 25.402694610778443, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 16969 }, { "epoch": 25.404191616766468, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.132, "step": 16970 }, { "epoch": 25.40568862275449, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1342, "step": 16971 }, { "epoch": 25.407185628742514, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1369, "step": 16972 }, { "epoch": 25.40868263473054, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16973 }, { "epoch": 25.410179640718564, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1286, "step": 16974 }, { "epoch": 25.411676646706585, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1245, "step": 16975 }, { "epoch": 25.41317365269461, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1266, "step": 16976 }, { "epoch": 25.414670658682635, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1234, "step": 16977 }, { "epoch": 25.41616766467066, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1348, "step": 16978 }, { "epoch": 25.41766467065868, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1322, "step": 16979 }, { "epoch": 25.419161676646706, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1342, "step": 16980 }, { "epoch": 25.42065868263473, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1244, "step": 16981 }, { "epoch": 25.422155688622755, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 16982 }, { "epoch": 25.42365269461078, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 16983 }, { "epoch": 25.4251497005988, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1269, "step": 16984 }, { "epoch": 25.426646706586826, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1374, "step": 16985 }, { "epoch": 25.42814371257485, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.132, "step": 16986 }, { "epoch": 25.429640718562876, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1277, "step": 16987 }, { "epoch": 25.431137724550897, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.131, "step": 16988 }, { "epoch": 25.432634730538922, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 16989 }, { "epoch": 25.434131736526947, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1222, "step": 16990 }, { "epoch": 25.43562874251497, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1278, "step": 16991 }, { "epoch": 25.437125748502993, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1283, "step": 16992 }, { "epoch": 25.438622754491018, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1362, "step": 16993 }, { "epoch": 25.440119760479043, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1287, "step": 16994 }, { "epoch": 25.441616766467067, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1265, "step": 16995 }, { "epoch": 25.44311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1327, "step": 16996 }, { "epoch": 25.444610778443113, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1338, "step": 16997 }, { "epoch": 25.44610778443114, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1346, "step": 16998 }, { "epoch": 25.447604790419163, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1363, "step": 16999 }, { "epoch": 25.449101796407184, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1299, "step": 17000 }, { "epoch": 25.45059880239521, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1269, "step": 17001 }, { "epoch": 25.452095808383234, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.124, "step": 17002 }, { "epoch": 25.45359281437126, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1279, "step": 17003 }, { "epoch": 25.45508982035928, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1257, "step": 17004 }, { "epoch": 25.456586826347305, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 17005 }, { "epoch": 25.45808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1327, "step": 17006 }, { "epoch": 25.459580838323355, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1307, "step": 17007 }, { "epoch": 25.461077844311376, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1383, "step": 17008 }, { "epoch": 25.4625748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1313, "step": 17009 }, { "epoch": 25.464071856287426, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 17010 }, { "epoch": 25.46556886227545, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1295, "step": 17011 }, { "epoch": 25.46706586826347, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 17012 }, { "epoch": 25.468562874251496, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.132, "step": 17013 }, { "epoch": 25.47005988023952, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.129, "step": 17014 }, { "epoch": 25.471556886227546, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 17015 }, { "epoch": 25.473053892215567, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1345, "step": 17016 }, { "epoch": 25.474550898203592, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1318, "step": 17017 }, { "epoch": 25.476047904191617, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1271, "step": 17018 }, { "epoch": 25.477544910179642, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1344, "step": 17019 }, { "epoch": 25.479041916167663, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1336, "step": 17020 }, { "epoch": 25.480538922155688, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1295, "step": 17021 }, { "epoch": 25.482035928143713, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1327, "step": 17022 }, { "epoch": 25.483532934131738, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1332, "step": 17023 }, { "epoch": 25.48502994011976, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1297, "step": 17024 }, { "epoch": 25.486526946107784, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.132, "step": 17025 }, { "epoch": 25.48802395209581, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1253, "step": 17026 }, { "epoch": 25.489520958083833, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1257, "step": 17027 }, { "epoch": 25.491017964071855, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 17028 }, { "epoch": 25.49251497005988, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1345, "step": 17029 }, { "epoch": 25.494011976047904, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1273, "step": 17030 }, { "epoch": 25.49550898203593, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 17031 }, { "epoch": 25.49700598802395, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1347, "step": 17032 }, { "epoch": 25.498502994011975, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1325, "step": 17033 }, { "epoch": 25.5, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1303, "step": 17034 }, { "epoch": 25.501497005988025, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1276, "step": 17035 }, { "epoch": 25.50299401197605, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.138, "step": 17036 }, { "epoch": 25.50449101796407, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17037 }, { "epoch": 25.505988023952096, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1333, "step": 17038 }, { "epoch": 25.50748502994012, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1399, "step": 17039 }, { "epoch": 25.508982035928145, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.127, "step": 17040 }, { "epoch": 25.510479041916167, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1374, "step": 17041 }, { "epoch": 25.51197604790419, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1287, "step": 17042 }, { "epoch": 25.513473053892216, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1355, "step": 17043 }, { "epoch": 25.51497005988024, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1305, "step": 17044 }, { "epoch": 25.516467065868262, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1302, "step": 17045 }, { "epoch": 25.517964071856287, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1263, "step": 17046 }, { "epoch": 25.519461077844312, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1361, "step": 17047 }, { "epoch": 25.520958083832337, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1307, "step": 17048 }, { "epoch": 25.522455089820358, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1246, "step": 17049 }, { "epoch": 25.523952095808383, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17050 }, { "epoch": 25.525449101796408, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1338, "step": 17051 }, { "epoch": 25.526946107784433, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1245, "step": 17052 }, { "epoch": 25.528443113772454, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1306, "step": 17053 }, { "epoch": 25.52994011976048, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1285, "step": 17054 }, { "epoch": 25.531437125748504, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1239, "step": 17055 }, { "epoch": 25.53293413173653, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1336, "step": 17056 }, { "epoch": 25.53443113772455, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1316, "step": 17057 }, { "epoch": 25.535928143712574, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1427, "step": 17058 }, { "epoch": 25.5374251497006, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1287, "step": 17059 }, { "epoch": 25.538922155688624, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1327, "step": 17060 }, { "epoch": 25.540419161676645, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1254, "step": 17061 }, { "epoch": 25.54191616766467, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1262, "step": 17062 }, { "epoch": 25.543413173652695, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 17063 }, { "epoch": 25.54491017964072, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1315, "step": 17064 }, { "epoch": 25.54640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1342, "step": 17065 }, { "epoch": 25.547904191616766, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1312, "step": 17066 }, { "epoch": 25.54940119760479, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 17067 }, { "epoch": 25.550898203592816, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17068 }, { "epoch": 25.552395209580837, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1332, "step": 17069 }, { "epoch": 25.55389221556886, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.132, "step": 17070 }, { "epoch": 25.555389221556887, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1266, "step": 17071 }, { "epoch": 25.55688622754491, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1333, "step": 17072 }, { "epoch": 25.558383233532933, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1335, "step": 17073 }, { "epoch": 25.559880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1255, "step": 17074 }, { "epoch": 25.561377245508982, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1279, "step": 17075 }, { "epoch": 25.562874251497007, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1346, "step": 17076 }, { "epoch": 25.56437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 17077 }, { "epoch": 25.565868263473053, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 17078 }, { "epoch": 25.567365269461078, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.127, "step": 17079 }, { "epoch": 25.568862275449103, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.129, "step": 17080 }, { "epoch": 25.570359281437124, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1321, "step": 17081 }, { "epoch": 25.57185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1324, "step": 17082 }, { "epoch": 25.573353293413174, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1281, "step": 17083 }, { "epoch": 25.5748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1283, "step": 17084 }, { "epoch": 25.57634730538922, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1255, "step": 17085 }, { "epoch": 25.577844311377245, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1438, "step": 17086 }, { "epoch": 25.57934131736527, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.129, "step": 17087 }, { "epoch": 25.580838323353294, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1304, "step": 17088 }, { "epoch": 25.58233532934132, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1317, "step": 17089 }, { "epoch": 25.58383233532934, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1254, "step": 17090 }, { "epoch": 25.585329341317365, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1252, "step": 17091 }, { "epoch": 25.58682634730539, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1224, "step": 17092 }, { "epoch": 25.58832335329341, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1329, "step": 17093 }, { "epoch": 25.589820359281436, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1378, "step": 17094 }, { "epoch": 25.59131736526946, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1312, "step": 17095 }, { "epoch": 25.592814371257486, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1242, "step": 17096 }, { "epoch": 25.59431137724551, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1266, "step": 17097 }, { "epoch": 25.595808383233532, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17098 }, { "epoch": 25.597305389221557, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17099 }, { "epoch": 25.59880239520958, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1278, "step": 17100 }, { "epoch": 25.600299401197606, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1256, "step": 17101 }, { "epoch": 25.601796407185628, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1284, "step": 17102 }, { "epoch": 25.603293413173652, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1312, "step": 17103 }, { "epoch": 25.604790419161677, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1262, "step": 17104 }, { "epoch": 25.606287425149702, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.13, "step": 17105 }, { "epoch": 25.607784431137723, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 17106 }, { "epoch": 25.60928143712575, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1374, "step": 17107 }, { "epoch": 25.610778443113773, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1285, "step": 17108 }, { "epoch": 25.612275449101798, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1239, "step": 17109 }, { "epoch": 25.61377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1271, "step": 17110 }, { "epoch": 25.615269461077844, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 17111 }, { "epoch": 25.61676646706587, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1234, "step": 17112 }, { "epoch": 25.618263473053894, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 17113 }, { "epoch": 25.619760479041915, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1349, "step": 17114 }, { "epoch": 25.62125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1261, "step": 17115 }, { "epoch": 25.622754491017965, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1335, "step": 17116 }, { "epoch": 25.62425149700599, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 17117 }, { "epoch": 25.62574850299401, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1334, "step": 17118 }, { "epoch": 25.627245508982035, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1244, "step": 17119 }, { "epoch": 25.62874251497006, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1277, "step": 17120 }, { "epoch": 25.630239520958085, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1315, "step": 17121 }, { "epoch": 25.631736526946106, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17122 }, { "epoch": 25.63323353293413, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 17123 }, { "epoch": 25.634730538922156, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1309, "step": 17124 }, { "epoch": 25.63622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1281, "step": 17125 }, { "epoch": 25.637724550898202, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 17126 }, { "epoch": 25.639221556886227, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1288, "step": 17127 }, { "epoch": 25.64071856287425, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1336, "step": 17128 }, { "epoch": 25.642215568862277, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 17129 }, { "epoch": 25.643712574850298, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1178, "step": 17130 }, { "epoch": 25.645209580838323, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 17131 }, { "epoch": 25.646706586826348, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1338, "step": 17132 }, { "epoch": 25.648203592814372, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 17133 }, { "epoch": 25.649700598802394, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.128, "step": 17134 }, { "epoch": 25.65119760479042, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1328, "step": 17135 }, { "epoch": 25.652694610778443, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17136 }, { "epoch": 25.654191616766468, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1288, "step": 17137 }, { "epoch": 25.65568862275449, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 17138 }, { "epoch": 25.657185628742514, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1227, "step": 17139 }, { "epoch": 25.65868263473054, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.127, "step": 17140 }, { "epoch": 25.660179640718564, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1266, "step": 17141 }, { "epoch": 25.66167664670659, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1261, "step": 17142 }, { "epoch": 25.66317365269461, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1325, "step": 17143 }, { "epoch": 25.664670658682635, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1357, "step": 17144 }, { "epoch": 25.66616766467066, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1329, "step": 17145 }, { "epoch": 25.66766467065868, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1334, "step": 17146 }, { "epoch": 25.669161676646706, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1265, "step": 17147 }, { "epoch": 25.67065868263473, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1328, "step": 17148 }, { "epoch": 25.672155688622755, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1293, "step": 17149 }, { "epoch": 25.67365269461078, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1258, "step": 17150 }, { "epoch": 25.6751497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1304, "step": 17151 }, { "epoch": 25.676646706586826, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1286, "step": 17152 }, { "epoch": 25.67814371257485, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1258, "step": 17153 }, { "epoch": 25.679640718562876, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1297, "step": 17154 }, { "epoch": 25.681137724550897, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1302, "step": 17155 }, { "epoch": 25.682634730538922, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1313, "step": 17156 }, { "epoch": 25.684131736526947, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1255, "step": 17157 }, { "epoch": 25.68562874251497, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1256, "step": 17158 }, { "epoch": 25.687125748502993, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1368, "step": 17159 }, { "epoch": 25.688622754491018, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17160 }, { "epoch": 25.690119760479043, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1356, "step": 17161 }, { "epoch": 25.691616766467067, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1264, "step": 17162 }, { "epoch": 25.69311377245509, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 17163 }, { "epoch": 25.694610778443113, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1319, "step": 17164 }, { "epoch": 25.69610778443114, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1251, "step": 17165 }, { "epoch": 25.697604790419163, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1293, "step": 17166 }, { "epoch": 25.699101796407184, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1339, "step": 17167 }, { "epoch": 25.70059880239521, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1321, "step": 17168 }, { "epoch": 25.702095808383234, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17169 }, { "epoch": 25.70359281437126, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1273, "step": 17170 }, { "epoch": 25.70508982035928, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1276, "step": 17171 }, { "epoch": 25.706586826347305, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1249, "step": 17172 }, { "epoch": 25.70808383233533, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17173 }, { "epoch": 25.709580838323355, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1356, "step": 17174 }, { "epoch": 25.711077844311376, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1249, "step": 17175 }, { "epoch": 25.7125748502994, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1291, "step": 17176 }, { "epoch": 25.714071856287426, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1345, "step": 17177 }, { "epoch": 25.71556886227545, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1309, "step": 17178 }, { "epoch": 25.71706586826347, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1315, "step": 17179 }, { "epoch": 25.718562874251496, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.132, "step": 17180 }, { "epoch": 25.72005988023952, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17181 }, { "epoch": 25.721556886227546, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17182 }, { "epoch": 25.723053892215567, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.13, "step": 17183 }, { "epoch": 25.724550898203592, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17184 }, { "epoch": 25.726047904191617, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1383, "step": 17185 }, { "epoch": 25.727544910179642, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1267, "step": 17186 }, { "epoch": 25.729041916167663, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1337, "step": 17187 }, { "epoch": 25.730538922155688, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.133, "step": 17188 }, { "epoch": 25.732035928143713, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1354, "step": 17189 }, { "epoch": 25.733532934131738, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 17190 }, { "epoch": 25.73502994011976, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1286, "step": 17191 }, { "epoch": 25.736526946107784, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1326, "step": 17192 }, { "epoch": 25.73802395209581, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 17193 }, { "epoch": 25.739520958083833, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1351, "step": 17194 }, { "epoch": 25.741017964071855, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1385, "step": 17195 }, { "epoch": 25.74251497005988, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1279, "step": 17196 }, { "epoch": 25.744011976047904, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1302, "step": 17197 }, { "epoch": 25.74550898203593, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 17198 }, { "epoch": 25.74700598802395, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1324, "step": 17199 }, { "epoch": 25.748502994011975, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 17200 }, { "epoch": 25.75, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1268, "step": 17201 }, { "epoch": 25.751497005988025, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1305, "step": 17202 }, { "epoch": 25.75299401197605, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1344, "step": 17203 }, { "epoch": 25.75449101796407, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1356, "step": 17204 }, { "epoch": 25.755988023952096, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1378, "step": 17205 }, { "epoch": 25.75748502994012, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17206 }, { "epoch": 25.758982035928145, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1314, "step": 17207 }, { "epoch": 25.760479041916167, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 17208 }, { "epoch": 25.76197604790419, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1293, "step": 17209 }, { "epoch": 25.763473053892216, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1254, "step": 17210 }, { "epoch": 25.76497005988024, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1252, "step": 17211 }, { "epoch": 25.766467065868262, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1305, "step": 17212 }, { "epoch": 25.767964071856287, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1236, "step": 17213 }, { "epoch": 25.769461077844312, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17214 }, { "epoch": 25.770958083832337, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1343, "step": 17215 }, { "epoch": 25.772455089820358, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1339, "step": 17216 }, { "epoch": 25.773952095808383, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1323, "step": 17217 }, { "epoch": 25.775449101796408, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1335, "step": 17218 }, { "epoch": 25.776946107784433, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.125, "step": 17219 }, { "epoch": 25.778443113772454, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1355, "step": 17220 }, { "epoch": 25.77994011976048, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1307, "step": 17221 }, { "epoch": 25.781437125748504, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.132, "step": 17222 }, { "epoch": 25.78293413173653, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17223 }, { "epoch": 25.78443113772455, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.128, "step": 17224 }, { "epoch": 25.785928143712574, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1295, "step": 17225 }, { "epoch": 25.7874251497006, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1303, "step": 17226 }, { "epoch": 25.788922155688624, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1317, "step": 17227 }, { "epoch": 25.790419161676645, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 17228 }, { "epoch": 25.79191616766467, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1231, "step": 17229 }, { "epoch": 25.793413173652695, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.126, "step": 17230 }, { "epoch": 25.79491017964072, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1301, "step": 17231 }, { "epoch": 25.79640718562874, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17232 }, { "epoch": 25.797904191616766, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 17233 }, { "epoch": 25.79940119760479, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1326, "step": 17234 }, { "epoch": 25.800898203592816, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 17235 }, { "epoch": 25.802395209580837, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1231, "step": 17236 }, { "epoch": 25.80389221556886, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1385, "step": 17237 }, { "epoch": 25.805389221556887, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1206, "step": 17238 }, { "epoch": 25.80688622754491, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17239 }, { "epoch": 25.808383233532933, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1258, "step": 17240 }, { "epoch": 25.809880239520957, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1284, "step": 17241 }, { "epoch": 25.811377245508982, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17242 }, { "epoch": 25.812874251497007, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1251, "step": 17243 }, { "epoch": 25.81437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1252, "step": 17244 }, { "epoch": 25.815868263473053, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1307, "step": 17245 }, { "epoch": 25.817365269461078, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1345, "step": 17246 }, { "epoch": 25.818862275449103, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1264, "step": 17247 }, { "epoch": 25.820359281437124, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.136, "step": 17248 }, { "epoch": 25.82185628742515, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1264, "step": 17249 }, { "epoch": 25.823353293413174, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1258, "step": 17250 }, { "epoch": 25.8248502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1274, "step": 17251 }, { "epoch": 25.82634730538922, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1327, "step": 17252 }, { "epoch": 25.827844311377245, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1272, "step": 17253 }, { "epoch": 25.82934131736527, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1324, "step": 17254 }, { "epoch": 25.830838323353294, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1285, "step": 17255 }, { "epoch": 25.83233532934132, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1336, "step": 17256 }, { "epoch": 25.83383233532934, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1296, "step": 17257 }, { "epoch": 25.835329341317365, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.125, "step": 17258 }, { "epoch": 25.83682634730539, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1331, "step": 17259 }, { "epoch": 25.83832335329341, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1283, "step": 17260 }, { "epoch": 25.839820359281436, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1315, "step": 17261 }, { "epoch": 25.84131736526946, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17262 }, { "epoch": 25.842814371257486, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1333, "step": 17263 }, { "epoch": 25.84431137724551, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17264 }, { "epoch": 25.845808383233532, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1256, "step": 17265 }, { "epoch": 25.847305389221557, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.129, "step": 17266 }, { "epoch": 25.84880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1306, "step": 17267 }, { "epoch": 25.850299401197606, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1335, "step": 17268 }, { "epoch": 25.851796407185628, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.128, "step": 17269 }, { "epoch": 25.853293413173652, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1299, "step": 17270 }, { "epoch": 25.854790419161677, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 17271 }, { "epoch": 25.856287425149702, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1302, "step": 17272 }, { "epoch": 25.857784431137723, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1257, "step": 17273 }, { "epoch": 25.85928143712575, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1348, "step": 17274 }, { "epoch": 25.860778443113773, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1249, "step": 17275 }, { "epoch": 25.862275449101798, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1297, "step": 17276 }, { "epoch": 25.86377245508982, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1283, "step": 17277 }, { "epoch": 25.865269461077844, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.135, "step": 17278 }, { "epoch": 25.86676646706587, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1297, "step": 17279 }, { "epoch": 25.868263473053894, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1291, "step": 17280 }, { "epoch": 25.869760479041915, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.123, "step": 17281 }, { "epoch": 25.87125748502994, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1332, "step": 17282 }, { "epoch": 25.872754491017965, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1292, "step": 17283 }, { "epoch": 25.87425149700599, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1335, "step": 17284 }, { "epoch": 25.87574850299401, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1314, "step": 17285 }, { "epoch": 25.877245508982035, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1348, "step": 17286 }, { "epoch": 25.87874251497006, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1315, "step": 17287 }, { "epoch": 25.880239520958085, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1255, "step": 17288 }, { "epoch": 25.881736526946106, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1314, "step": 17289 }, { "epoch": 25.88323353293413, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17290 }, { "epoch": 25.884730538922156, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1356, "step": 17291 }, { "epoch": 25.88622754491018, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 17292 }, { "epoch": 25.887724550898202, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1316, "step": 17293 }, { "epoch": 25.889221556886227, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1271, "step": 17294 }, { "epoch": 25.89071856287425, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1361, "step": 17295 }, { "epoch": 25.892215568862277, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1303, "step": 17296 }, { "epoch": 25.893712574850298, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1289, "step": 17297 }, { "epoch": 25.895209580838323, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17298 }, { "epoch": 25.896706586826348, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1302, "step": 17299 }, { "epoch": 25.898203592814372, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1304, "step": 17300 }, { "epoch": 25.899700598802394, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1271, "step": 17301 }, { "epoch": 25.90119760479042, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1331, "step": 17302 }, { "epoch": 25.902694610778443, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1258, "step": 17303 }, { "epoch": 25.904191616766468, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17304 }, { "epoch": 25.90568862275449, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1342, "step": 17305 }, { "epoch": 25.907185628742514, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1322, "step": 17306 }, { "epoch": 25.90868263473054, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1282, "step": 17307 }, { "epoch": 25.910179640718564, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.126, "step": 17308 }, { "epoch": 25.91167664670659, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1308, "step": 17309 }, { "epoch": 25.91317365269461, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1336, "step": 17310 }, { "epoch": 25.914670658682635, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1238, "step": 17311 }, { "epoch": 25.91616766467066, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17312 }, { "epoch": 25.91766467065868, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 17313 }, { "epoch": 25.919161676646706, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1341, "step": 17314 }, { "epoch": 25.92065868263473, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 17315 }, { "epoch": 25.922155688622755, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1368, "step": 17316 }, { "epoch": 25.92365269461078, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1334, "step": 17317 }, { "epoch": 25.9251497005988, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1306, "step": 17318 }, { "epoch": 25.926646706586826, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1323, "step": 17319 }, { "epoch": 25.92814371257485, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1288, "step": 17320 }, { "epoch": 25.929640718562876, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1331, "step": 17321 }, { "epoch": 25.931137724550897, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1336, "step": 17322 }, { "epoch": 25.932634730538922, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1354, "step": 17323 }, { "epoch": 25.934131736526947, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 17324 }, { "epoch": 25.93562874251497, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1333, "step": 17325 }, { "epoch": 25.937125748502993, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1231, "step": 17326 }, { "epoch": 25.938622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1277, "step": 17327 }, { "epoch": 25.940119760479043, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 17328 }, { "epoch": 25.941616766467067, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1327, "step": 17329 }, { "epoch": 25.94311377245509, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1331, "step": 17330 }, { "epoch": 25.944610778443113, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1261, "step": 17331 }, { "epoch": 25.94610778443114, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1326, "step": 17332 }, { "epoch": 25.947604790419163, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1319, "step": 17333 }, { "epoch": 25.949101796407184, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1307, "step": 17334 }, { "epoch": 25.95059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1307, "step": 17335 }, { "epoch": 25.952095808383234, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1318, "step": 17336 }, { "epoch": 25.95359281437126, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1264, "step": 17337 }, { "epoch": 25.95508982035928, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1214, "step": 17338 }, { "epoch": 25.956586826347305, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 17339 }, { "epoch": 25.95808383233533, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1256, "step": 17340 }, { "epoch": 25.959580838323355, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1297, "step": 17341 }, { "epoch": 25.961077844311376, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 17342 }, { "epoch": 25.9625748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17343 }, { "epoch": 25.964071856287426, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1281, "step": 17344 }, { "epoch": 25.96556886227545, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1353, "step": 17345 }, { "epoch": 25.96706586826347, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.132, "step": 17346 }, { "epoch": 25.968562874251496, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1264, "step": 17347 }, { "epoch": 25.97005988023952, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1249, "step": 17348 }, { "epoch": 25.971556886227546, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1272, "step": 17349 }, { "epoch": 25.973053892215567, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17350 }, { "epoch": 25.974550898203592, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1375, "step": 17351 }, { "epoch": 25.976047904191617, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1305, "step": 17352 }, { "epoch": 25.977544910179642, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1278, "step": 17353 }, { "epoch": 25.979041916167663, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1323, "step": 17354 }, { "epoch": 25.980538922155688, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1306, "step": 17355 }, { "epoch": 25.982035928143713, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17356 }, { "epoch": 25.983532934131738, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.133, "step": 17357 }, { "epoch": 25.98502994011976, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1274, "step": 17358 }, { "epoch": 25.986526946107784, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1324, "step": 17359 }, { "epoch": 25.98802395209581, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1272, "step": 17360 }, { "epoch": 25.989520958083833, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1338, "step": 17361 }, { "epoch": 25.991017964071855, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1399, "step": 17362 }, { "epoch": 25.99251497005988, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1264, "step": 17363 }, { "epoch": 25.994011976047904, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 17364 }, { "epoch": 25.99550898203593, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1327, "step": 17365 }, { "epoch": 25.99700598802395, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1256, "step": 17366 }, { "epoch": 25.998502994011975, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1322, "step": 17367 }, { "epoch": 26.0, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1285, "step": 17368 }, { "epoch": 26.001497005988025, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.131, "step": 17369 }, { "epoch": 26.00299401197605, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17370 }, { "epoch": 26.00449101796407, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1265, "step": 17371 }, { "epoch": 26.005988023952096, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1318, "step": 17372 }, { "epoch": 26.00748502994012, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17373 }, { "epoch": 26.008982035928145, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1292, "step": 17374 }, { "epoch": 26.010479041916167, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1325, "step": 17375 }, { "epoch": 26.01197604790419, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1347, "step": 17376 }, { "epoch": 26.013473053892216, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 17377 }, { "epoch": 26.01497005988024, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1299, "step": 17378 }, { "epoch": 26.016467065868262, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1286, "step": 17379 }, { "epoch": 26.017964071856287, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.135, "step": 17380 }, { "epoch": 26.019461077844312, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1266, "step": 17381 }, { "epoch": 26.020958083832337, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1285, "step": 17382 }, { "epoch": 26.022455089820358, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1353, "step": 17383 }, { "epoch": 26.023952095808383, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1243, "step": 17384 }, { "epoch": 26.025449101796408, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1319, "step": 17385 }, { "epoch": 26.026946107784433, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17386 }, { "epoch": 26.028443113772454, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1323, "step": 17387 }, { "epoch": 26.02994011976048, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1289, "step": 17388 }, { "epoch": 26.031437125748504, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1308, "step": 17389 }, { "epoch": 26.03293413173653, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.132, "step": 17390 }, { "epoch": 26.03443113772455, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17391 }, { "epoch": 26.035928143712574, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1299, "step": 17392 }, { "epoch": 26.0374251497006, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1356, "step": 17393 }, { "epoch": 26.038922155688624, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17394 }, { "epoch": 26.040419161676645, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1346, "step": 17395 }, { "epoch": 26.04191616766467, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1294, "step": 17396 }, { "epoch": 26.043413173652695, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 17397 }, { "epoch": 26.04491017964072, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.126, "step": 17398 }, { "epoch": 26.04640718562874, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1264, "step": 17399 }, { "epoch": 26.047904191616766, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1303, "step": 17400 }, { "epoch": 26.04940119760479, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1326, "step": 17401 }, { "epoch": 26.050898203592816, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1297, "step": 17402 }, { "epoch": 26.052395209580837, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1317, "step": 17403 }, { "epoch": 26.05389221556886, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17404 }, { "epoch": 26.055389221556887, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1265, "step": 17405 }, { "epoch": 26.05688622754491, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1263, "step": 17406 }, { "epoch": 26.058383233532933, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.129, "step": 17407 }, { "epoch": 26.059880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1268, "step": 17408 }, { "epoch": 26.061377245508982, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1406, "step": 17409 }, { "epoch": 26.062874251497007, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1373, "step": 17410 }, { "epoch": 26.06437125748503, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1281, "step": 17411 }, { "epoch": 26.065868263473053, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1355, "step": 17412 }, { "epoch": 26.067365269461078, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.138, "step": 17413 }, { "epoch": 26.068862275449103, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1351, "step": 17414 }, { "epoch": 26.070359281437124, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1329, "step": 17415 }, { "epoch": 26.07185628742515, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1254, "step": 17416 }, { "epoch": 26.073353293413174, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1266, "step": 17417 }, { "epoch": 26.0748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1336, "step": 17418 }, { "epoch": 26.07634730538922, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1233, "step": 17419 }, { "epoch": 26.077844311377245, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17420 }, { "epoch": 26.07934131736527, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17421 }, { "epoch": 26.080838323353294, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1355, "step": 17422 }, { "epoch": 26.082335329341316, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17423 }, { "epoch": 26.08383233532934, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1256, "step": 17424 }, { "epoch": 26.085329341317365, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17425 }, { "epoch": 26.08682634730539, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1348, "step": 17426 }, { "epoch": 26.088323353293415, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1282, "step": 17427 }, { "epoch": 26.089820359281436, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1325, "step": 17428 }, { "epoch": 26.09131736526946, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1326, "step": 17429 }, { "epoch": 26.092814371257486, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17430 }, { "epoch": 26.09431137724551, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1271, "step": 17431 }, { "epoch": 26.095808383233532, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1339, "step": 17432 }, { "epoch": 26.097305389221557, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.13, "step": 17433 }, { "epoch": 26.09880239520958, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.133, "step": 17434 }, { "epoch": 26.100299401197606, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1243, "step": 17435 }, { "epoch": 26.101796407185628, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1248, "step": 17436 }, { "epoch": 26.103293413173652, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1259, "step": 17437 }, { "epoch": 26.104790419161677, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1285, "step": 17438 }, { "epoch": 26.106287425149702, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1307, "step": 17439 }, { "epoch": 26.107784431137723, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.136, "step": 17440 }, { "epoch": 26.10928143712575, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1274, "step": 17441 }, { "epoch": 26.110778443113773, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1281, "step": 17442 }, { "epoch": 26.112275449101798, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17443 }, { "epoch": 26.11377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 17444 }, { "epoch": 26.115269461077844, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1364, "step": 17445 }, { "epoch": 26.11676646706587, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.129, "step": 17446 }, { "epoch": 26.118263473053894, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1265, "step": 17447 }, { "epoch": 26.119760479041915, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1282, "step": 17448 }, { "epoch": 26.12125748502994, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1282, "step": 17449 }, { "epoch": 26.122754491017965, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1324, "step": 17450 }, { "epoch": 26.12425149700599, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1235, "step": 17451 }, { "epoch": 26.12574850299401, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1277, "step": 17452 }, { "epoch": 26.127245508982035, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 17453 }, { "epoch": 26.12874251497006, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1358, "step": 17454 }, { "epoch": 26.130239520958085, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.123, "step": 17455 }, { "epoch": 26.131736526946106, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1269, "step": 17456 }, { "epoch": 26.13323353293413, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1241, "step": 17457 }, { "epoch": 26.134730538922156, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 17458 }, { "epoch": 26.13622754491018, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1325, "step": 17459 }, { "epoch": 26.137724550898202, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1222, "step": 17460 }, { "epoch": 26.139221556886227, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1313, "step": 17461 }, { "epoch": 26.14071856287425, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 17462 }, { "epoch": 26.142215568862277, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1215, "step": 17463 }, { "epoch": 26.143712574850298, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17464 }, { "epoch": 26.145209580838323, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1283, "step": 17465 }, { "epoch": 26.146706586826348, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1235, "step": 17466 }, { "epoch": 26.148203592814372, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1324, "step": 17467 }, { "epoch": 26.149700598802394, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17468 }, { "epoch": 26.15119760479042, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1316, "step": 17469 }, { "epoch": 26.152694610778443, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1306, "step": 17470 }, { "epoch": 26.154191616766468, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1312, "step": 17471 }, { "epoch": 26.15568862275449, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 17472 }, { "epoch": 26.157185628742514, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1287, "step": 17473 }, { "epoch": 26.15868263473054, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1323, "step": 17474 }, { "epoch": 26.160179640718564, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1283, "step": 17475 }, { "epoch": 26.161676646706585, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1227, "step": 17476 }, { "epoch": 26.16317365269461, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 17477 }, { "epoch": 26.164670658682635, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1377, "step": 17478 }, { "epoch": 26.16616766467066, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1251, "step": 17479 }, { "epoch": 26.16766467065868, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1277, "step": 17480 }, { "epoch": 26.169161676646706, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17481 }, { "epoch": 26.17065868263473, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1331, "step": 17482 }, { "epoch": 26.172155688622755, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.131, "step": 17483 }, { "epoch": 26.17365269461078, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17484 }, { "epoch": 26.1751497005988, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1392, "step": 17485 }, { "epoch": 26.176646706586826, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.125, "step": 17486 }, { "epoch": 26.17814371257485, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1278, "step": 17487 }, { "epoch": 26.179640718562876, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17488 }, { "epoch": 26.181137724550897, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.128, "step": 17489 }, { "epoch": 26.182634730538922, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 17490 }, { "epoch": 26.184131736526947, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17491 }, { "epoch": 26.18562874251497, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1349, "step": 17492 }, { "epoch": 26.187125748502993, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 17493 }, { "epoch": 26.188622754491018, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1219, "step": 17494 }, { "epoch": 26.190119760479043, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1342, "step": 17495 }, { "epoch": 26.191616766467067, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1353, "step": 17496 }, { "epoch": 26.19311377245509, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1326, "step": 17497 }, { "epoch": 26.194610778443113, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1342, "step": 17498 }, { "epoch": 26.19610778443114, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.129, "step": 17499 }, { "epoch": 26.197604790419163, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1315, "step": 17500 }, { "epoch": 26.199101796407184, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1249, "step": 17501 }, { "epoch": 26.20059880239521, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1357, "step": 17502 }, { "epoch": 26.202095808383234, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1232, "step": 17503 }, { "epoch": 26.20359281437126, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1295, "step": 17504 }, { "epoch": 26.20508982035928, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.131, "step": 17505 }, { "epoch": 26.206586826347305, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1188, "step": 17506 }, { "epoch": 26.20808383233533, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1195, "step": 17507 }, { "epoch": 26.209580838323355, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1269, "step": 17508 }, { "epoch": 26.211077844311376, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1328, "step": 17509 }, { "epoch": 26.2125748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1295, "step": 17510 }, { "epoch": 26.214071856287426, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 17511 }, { "epoch": 26.21556886227545, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1367, "step": 17512 }, { "epoch": 26.21706586826347, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1297, "step": 17513 }, { "epoch": 26.218562874251496, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1343, "step": 17514 }, { "epoch": 26.22005988023952, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1269, "step": 17515 }, { "epoch": 26.221556886227546, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17516 }, { "epoch": 26.223053892215567, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1324, "step": 17517 }, { "epoch": 26.224550898203592, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1315, "step": 17518 }, { "epoch": 26.226047904191617, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1238, "step": 17519 }, { "epoch": 26.227544910179642, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1407, "step": 17520 }, { "epoch": 26.229041916167663, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1289, "step": 17521 }, { "epoch": 26.230538922155688, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1244, "step": 17522 }, { "epoch": 26.232035928143713, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1334, "step": 17523 }, { "epoch": 26.233532934131738, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.133, "step": 17524 }, { "epoch": 26.23502994011976, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1325, "step": 17525 }, { "epoch": 26.236526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1354, "step": 17526 }, { "epoch": 26.23802395209581, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.125, "step": 17527 }, { "epoch": 26.239520958083833, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1258, "step": 17528 }, { "epoch": 26.241017964071855, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.133, "step": 17529 }, { "epoch": 26.24251497005988, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1335, "step": 17530 }, { "epoch": 26.244011976047904, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1321, "step": 17531 }, { "epoch": 26.24550898203593, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1305, "step": 17532 }, { "epoch": 26.24700598802395, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1337, "step": 17533 }, { "epoch": 26.248502994011975, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1356, "step": 17534 }, { "epoch": 26.25, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1355, "step": 17535 }, { "epoch": 26.251497005988025, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.125, "step": 17536 }, { "epoch": 26.25299401197605, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1312, "step": 17537 }, { "epoch": 26.25449101796407, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 17538 }, { "epoch": 26.255988023952096, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1271, "step": 17539 }, { "epoch": 26.25748502994012, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17540 }, { "epoch": 26.258982035928145, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1315, "step": 17541 }, { "epoch": 26.260479041916167, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 17542 }, { "epoch": 26.26197604790419, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1265, "step": 17543 }, { "epoch": 26.263473053892216, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.128, "step": 17544 }, { "epoch": 26.26497005988024, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 17545 }, { "epoch": 26.266467065868262, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 17546 }, { "epoch": 26.267964071856287, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1325, "step": 17547 }, { "epoch": 26.269461077844312, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 17548 }, { "epoch": 26.270958083832337, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1273, "step": 17549 }, { "epoch": 26.272455089820358, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1246, "step": 17550 }, { "epoch": 26.273952095808383, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1297, "step": 17551 }, { "epoch": 26.275449101796408, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1289, "step": 17552 }, { "epoch": 26.276946107784433, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1316, "step": 17553 }, { "epoch": 26.278443113772454, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1214, "step": 17554 }, { "epoch": 26.27994011976048, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1295, "step": 17555 }, { "epoch": 26.281437125748504, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1275, "step": 17556 }, { "epoch": 26.28293413173653, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1323, "step": 17557 }, { "epoch": 26.28443113772455, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1324, "step": 17558 }, { "epoch": 26.285928143712574, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1363, "step": 17559 }, { "epoch": 26.2874251497006, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1294, "step": 17560 }, { "epoch": 26.288922155688624, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1266, "step": 17561 }, { "epoch": 26.290419161676645, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1277, "step": 17562 }, { "epoch": 26.29191616766467, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1292, "step": 17563 }, { "epoch": 26.293413173652695, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.126, "step": 17564 }, { "epoch": 26.29491017964072, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1291, "step": 17565 }, { "epoch": 26.29640718562874, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1278, "step": 17566 }, { "epoch": 26.297904191616766, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1287, "step": 17567 }, { "epoch": 26.29940119760479, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.134, "step": 17568 }, { "epoch": 26.300898203592816, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.129, "step": 17569 }, { "epoch": 26.302395209580837, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1264, "step": 17570 }, { "epoch": 26.30389221556886, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1282, "step": 17571 }, { "epoch": 26.305389221556887, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17572 }, { "epoch": 26.30688622754491, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1308, "step": 17573 }, { "epoch": 26.308383233532933, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17574 }, { "epoch": 26.309880239520957, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 17575 }, { "epoch": 26.311377245508982, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1364, "step": 17576 }, { "epoch": 26.312874251497007, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.132, "step": 17577 }, { "epoch": 26.31437125748503, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1261, "step": 17578 }, { "epoch": 26.315868263473053, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17579 }, { "epoch": 26.317365269461078, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1276, "step": 17580 }, { "epoch": 26.318862275449103, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1312, "step": 17581 }, { "epoch": 26.320359281437124, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1288, "step": 17582 }, { "epoch": 26.32185628742515, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1337, "step": 17583 }, { "epoch": 26.323353293413174, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 17584 }, { "epoch": 26.3248502994012, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1306, "step": 17585 }, { "epoch": 26.32634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1289, "step": 17586 }, { "epoch": 26.327844311377245, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1229, "step": 17587 }, { "epoch": 26.32934131736527, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1372, "step": 17588 }, { "epoch": 26.330838323353294, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1275, "step": 17589 }, { "epoch": 26.33233532934132, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1227, "step": 17590 }, { "epoch": 26.33383233532934, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 17591 }, { "epoch": 26.335329341317365, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1309, "step": 17592 }, { "epoch": 26.33682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1253, "step": 17593 }, { "epoch": 26.338323353293415, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 17594 }, { "epoch": 26.339820359281436, "grad_norm": 0.04931640625, "learning_rate": 0.0008, "loss": 1.1258, "step": 17595 }, { "epoch": 26.34131736526946, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1299, "step": 17596 }, { "epoch": 26.342814371257486, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1241, "step": 17597 }, { "epoch": 26.34431137724551, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1345, "step": 17598 }, { "epoch": 26.345808383233532, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1357, "step": 17599 }, { "epoch": 26.347305389221557, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1284, "step": 17600 }, { "epoch": 26.34880239520958, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1323, "step": 17601 }, { "epoch": 26.350299401197606, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1301, "step": 17602 }, { "epoch": 26.351796407185628, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1338, "step": 17603 }, { "epoch": 26.353293413173652, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 17604 }, { "epoch": 26.354790419161677, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 17605 }, { "epoch": 26.356287425149702, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1313, "step": 17606 }, { "epoch": 26.357784431137723, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 17607 }, { "epoch": 26.35928143712575, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17608 }, { "epoch": 26.360778443113773, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1342, "step": 17609 }, { "epoch": 26.362275449101798, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1388, "step": 17610 }, { "epoch": 26.36377245508982, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1306, "step": 17611 }, { "epoch": 26.365269461077844, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17612 }, { "epoch": 26.36676646706587, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1275, "step": 17613 }, { "epoch": 26.368263473053894, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.127, "step": 17614 }, { "epoch": 26.369760479041915, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1312, "step": 17615 }, { "epoch": 26.37125748502994, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1304, "step": 17616 }, { "epoch": 26.372754491017965, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17617 }, { "epoch": 26.37425149700599, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1327, "step": 17618 }, { "epoch": 26.37574850299401, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1284, "step": 17619 }, { "epoch": 26.377245508982035, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1322, "step": 17620 }, { "epoch": 26.37874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.125, "step": 17621 }, { "epoch": 26.380239520958085, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1326, "step": 17622 }, { "epoch": 26.381736526946106, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17623 }, { "epoch": 26.38323353293413, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17624 }, { "epoch": 26.384730538922156, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1341, "step": 17625 }, { "epoch": 26.38622754491018, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1325, "step": 17626 }, { "epoch": 26.387724550898202, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1297, "step": 17627 }, { "epoch": 26.389221556886227, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1197, "step": 17628 }, { "epoch": 26.39071856287425, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1303, "step": 17629 }, { "epoch": 26.392215568862277, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1266, "step": 17630 }, { "epoch": 26.393712574850298, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1276, "step": 17631 }, { "epoch": 26.395209580838323, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1293, "step": 17632 }, { "epoch": 26.396706586826348, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1238, "step": 17633 }, { "epoch": 26.398203592814372, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1274, "step": 17634 }, { "epoch": 26.399700598802394, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1341, "step": 17635 }, { "epoch": 26.40119760479042, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1381, "step": 17636 }, { "epoch": 26.402694610778443, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1357, "step": 17637 }, { "epoch": 26.404191616766468, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1306, "step": 17638 }, { "epoch": 26.40568862275449, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.129, "step": 17639 }, { "epoch": 26.407185628742514, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1305, "step": 17640 }, { "epoch": 26.40868263473054, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1225, "step": 17641 }, { "epoch": 26.410179640718564, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1279, "step": 17642 }, { "epoch": 26.411676646706585, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1278, "step": 17643 }, { "epoch": 26.41317365269461, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1346, "step": 17644 }, { "epoch": 26.414670658682635, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1244, "step": 17645 }, { "epoch": 26.41616766467066, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1283, "step": 17646 }, { "epoch": 26.41766467065868, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1292, "step": 17647 }, { "epoch": 26.419161676646706, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1304, "step": 17648 }, { "epoch": 26.42065868263473, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1291, "step": 17649 }, { "epoch": 26.422155688622755, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1293, "step": 17650 }, { "epoch": 26.42365269461078, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.136, "step": 17651 }, { "epoch": 26.4251497005988, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17652 }, { "epoch": 26.426646706586826, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.131, "step": 17653 }, { "epoch": 26.42814371257485, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1271, "step": 17654 }, { "epoch": 26.429640718562876, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1285, "step": 17655 }, { "epoch": 26.431137724550897, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17656 }, { "epoch": 26.432634730538922, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17657 }, { "epoch": 26.434131736526947, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17658 }, { "epoch": 26.43562874251497, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1293, "step": 17659 }, { "epoch": 26.437125748502993, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1292, "step": 17660 }, { "epoch": 26.438622754491018, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1313, "step": 17661 }, { "epoch": 26.440119760479043, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1285, "step": 17662 }, { "epoch": 26.441616766467067, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1247, "step": 17663 }, { "epoch": 26.44311377245509, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1249, "step": 17664 }, { "epoch": 26.444610778443113, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1302, "step": 17665 }, { "epoch": 26.44610778443114, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 17666 }, { "epoch": 26.447604790419163, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1333, "step": 17667 }, { "epoch": 26.449101796407184, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.127, "step": 17668 }, { "epoch": 26.45059880239521, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1399, "step": 17669 }, { "epoch": 26.452095808383234, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1307, "step": 17670 }, { "epoch": 26.45359281437126, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1192, "step": 17671 }, { "epoch": 26.45508982035928, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1243, "step": 17672 }, { "epoch": 26.456586826347305, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1309, "step": 17673 }, { "epoch": 26.45808383233533, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1252, "step": 17674 }, { "epoch": 26.459580838323355, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1321, "step": 17675 }, { "epoch": 26.461077844311376, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1336, "step": 17676 }, { "epoch": 26.4625748502994, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 17677 }, { "epoch": 26.464071856287426, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 17678 }, { "epoch": 26.46556886227545, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 17679 }, { "epoch": 26.46706586826347, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1284, "step": 17680 }, { "epoch": 26.468562874251496, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1287, "step": 17681 }, { "epoch": 26.47005988023952, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 17682 }, { "epoch": 26.471556886227546, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 17683 }, { "epoch": 26.473053892215567, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1289, "step": 17684 }, { "epoch": 26.474550898203592, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.13, "step": 17685 }, { "epoch": 26.476047904191617, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1363, "step": 17686 }, { "epoch": 26.477544910179642, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1362, "step": 17687 }, { "epoch": 26.479041916167663, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1348, "step": 17688 }, { "epoch": 26.480538922155688, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17689 }, { "epoch": 26.482035928143713, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17690 }, { "epoch": 26.483532934131738, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1244, "step": 17691 }, { "epoch": 26.48502994011976, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.137, "step": 17692 }, { "epoch": 26.486526946107784, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1316, "step": 17693 }, { "epoch": 26.48802395209581, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1304, "step": 17694 }, { "epoch": 26.489520958083833, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.132, "step": 17695 }, { "epoch": 26.491017964071855, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1239, "step": 17696 }, { "epoch": 26.49251497005988, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1267, "step": 17697 }, { "epoch": 26.494011976047904, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 17698 }, { "epoch": 26.49550898203593, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1254, "step": 17699 }, { "epoch": 26.49700598802395, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1311, "step": 17700 }, { "epoch": 26.498502994011975, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1229, "step": 17701 }, { "epoch": 26.5, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1263, "step": 17702 }, { "epoch": 26.501497005988025, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1323, "step": 17703 }, { "epoch": 26.50299401197605, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 17704 }, { "epoch": 26.50449101796407, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1305, "step": 17705 }, { "epoch": 26.505988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1266, "step": 17706 }, { "epoch": 26.50748502994012, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1334, "step": 17707 }, { "epoch": 26.508982035928145, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 17708 }, { "epoch": 26.510479041916167, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1297, "step": 17709 }, { "epoch": 26.51197604790419, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17710 }, { "epoch": 26.513473053892216, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1307, "step": 17711 }, { "epoch": 26.51497005988024, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1209, "step": 17712 }, { "epoch": 26.516467065868262, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 17713 }, { "epoch": 26.517964071856287, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 17714 }, { "epoch": 26.519461077844312, "grad_norm": 0.048095703125, "learning_rate": 0.0008, "loss": 1.1302, "step": 17715 }, { "epoch": 26.520958083832337, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.135, "step": 17716 }, { "epoch": 26.522455089820358, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1345, "step": 17717 }, { "epoch": 26.523952095808383, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1277, "step": 17718 }, { "epoch": 26.525449101796408, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1238, "step": 17719 }, { "epoch": 26.526946107784433, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 17720 }, { "epoch": 26.528443113772454, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1299, "step": 17721 }, { "epoch": 26.52994011976048, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1267, "step": 17722 }, { "epoch": 26.531437125748504, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1307, "step": 17723 }, { "epoch": 26.53293413173653, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1271, "step": 17724 }, { "epoch": 26.53443113772455, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1272, "step": 17725 }, { "epoch": 26.535928143712574, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1336, "step": 17726 }, { "epoch": 26.5374251497006, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 17727 }, { "epoch": 26.538922155688624, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1268, "step": 17728 }, { "epoch": 26.540419161676645, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1366, "step": 17729 }, { "epoch": 26.54191616766467, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.131, "step": 17730 }, { "epoch": 26.543413173652695, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 17731 }, { "epoch": 26.54491017964072, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17732 }, { "epoch": 26.54640718562874, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1364, "step": 17733 }, { "epoch": 26.547904191616766, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1292, "step": 17734 }, { "epoch": 26.54940119760479, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1225, "step": 17735 }, { "epoch": 26.550898203592816, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1306, "step": 17736 }, { "epoch": 26.552395209580837, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1287, "step": 17737 }, { "epoch": 26.55389221556886, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1243, "step": 17738 }, { "epoch": 26.555389221556887, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1272, "step": 17739 }, { "epoch": 26.55688622754491, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1323, "step": 17740 }, { "epoch": 26.558383233532933, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1317, "step": 17741 }, { "epoch": 26.559880239520957, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1235, "step": 17742 }, { "epoch": 26.561377245508982, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1374, "step": 17743 }, { "epoch": 26.562874251497007, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 17744 }, { "epoch": 26.56437125748503, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1323, "step": 17745 }, { "epoch": 26.565868263473053, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17746 }, { "epoch": 26.567365269461078, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1275, "step": 17747 }, { "epoch": 26.568862275449103, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1246, "step": 17748 }, { "epoch": 26.570359281437124, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1344, "step": 17749 }, { "epoch": 26.57185628742515, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1286, "step": 17750 }, { "epoch": 26.573353293413174, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1366, "step": 17751 }, { "epoch": 26.5748502994012, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1367, "step": 17752 }, { "epoch": 26.57634730538922, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1307, "step": 17753 }, { "epoch": 26.577844311377245, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1316, "step": 17754 }, { "epoch": 26.57934131736527, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1312, "step": 17755 }, { "epoch": 26.580838323353294, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1276, "step": 17756 }, { "epoch": 26.58233532934132, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1281, "step": 17757 }, { "epoch": 26.58383233532934, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1274, "step": 17758 }, { "epoch": 26.585329341317365, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.127, "step": 17759 }, { "epoch": 26.58682634730539, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1192, "step": 17760 }, { "epoch": 26.58832335329341, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1253, "step": 17761 }, { "epoch": 26.589820359281436, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1252, "step": 17762 }, { "epoch": 26.59131736526946, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1328, "step": 17763 }, { "epoch": 26.592814371257486, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1259, "step": 17764 }, { "epoch": 26.59431137724551, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1304, "step": 17765 }, { "epoch": 26.595808383233532, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1341, "step": 17766 }, { "epoch": 26.597305389221557, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1287, "step": 17767 }, { "epoch": 26.59880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1319, "step": 17768 }, { "epoch": 26.600299401197606, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1285, "step": 17769 }, { "epoch": 26.601796407185628, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1278, "step": 17770 }, { "epoch": 26.603293413173652, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1231, "step": 17771 }, { "epoch": 26.604790419161677, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1283, "step": 17772 }, { "epoch": 26.606287425149702, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1292, "step": 17773 }, { "epoch": 26.607784431137723, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1323, "step": 17774 }, { "epoch": 26.60928143712575, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1312, "step": 17775 }, { "epoch": 26.610778443113773, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.128, "step": 17776 }, { "epoch": 26.612275449101798, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1403, "step": 17777 }, { "epoch": 26.61377245508982, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1337, "step": 17778 }, { "epoch": 26.615269461077844, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1372, "step": 17779 }, { "epoch": 26.61676646706587, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1323, "step": 17780 }, { "epoch": 26.618263473053894, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1346, "step": 17781 }, { "epoch": 26.619760479041915, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17782 }, { "epoch": 26.62125748502994, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17783 }, { "epoch": 26.622754491017965, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1258, "step": 17784 }, { "epoch": 26.62425149700599, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1309, "step": 17785 }, { "epoch": 26.62574850299401, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1286, "step": 17786 }, { "epoch": 26.627245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1359, "step": 17787 }, { "epoch": 26.62874251497006, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1277, "step": 17788 }, { "epoch": 26.630239520958085, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1256, "step": 17789 }, { "epoch": 26.631736526946106, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1367, "step": 17790 }, { "epoch": 26.63323353293413, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1343, "step": 17791 }, { "epoch": 26.634730538922156, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1364, "step": 17792 }, { "epoch": 26.63622754491018, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1296, "step": 17793 }, { "epoch": 26.637724550898202, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1281, "step": 17794 }, { "epoch": 26.639221556886227, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1303, "step": 17795 }, { "epoch": 26.64071856287425, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1285, "step": 17796 }, { "epoch": 26.642215568862277, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1318, "step": 17797 }, { "epoch": 26.643712574850298, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1249, "step": 17798 }, { "epoch": 26.645209580838323, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 17799 }, { "epoch": 26.646706586826348, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1314, "step": 17800 }, { "epoch": 26.648203592814372, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1275, "step": 17801 }, { "epoch": 26.649700598802394, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1276, "step": 17802 }, { "epoch": 26.65119760479042, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17803 }, { "epoch": 26.652694610778443, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.133, "step": 17804 }, { "epoch": 26.654191616766468, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1262, "step": 17805 }, { "epoch": 26.65568862275449, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1276, "step": 17806 }, { "epoch": 26.657185628742514, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1247, "step": 17807 }, { "epoch": 26.65868263473054, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17808 }, { "epoch": 26.660179640718564, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1324, "step": 17809 }, { "epoch": 26.66167664670659, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1361, "step": 17810 }, { "epoch": 26.66317365269461, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1352, "step": 17811 }, { "epoch": 26.664670658682635, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1299, "step": 17812 }, { "epoch": 26.66616766467066, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1315, "step": 17813 }, { "epoch": 26.66766467065868, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.127, "step": 17814 }, { "epoch": 26.669161676646706, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.129, "step": 17815 }, { "epoch": 26.67065868263473, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 17816 }, { "epoch": 26.672155688622755, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1255, "step": 17817 }, { "epoch": 26.67365269461078, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1288, "step": 17818 }, { "epoch": 26.6751497005988, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1275, "step": 17819 }, { "epoch": 26.676646706586826, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1286, "step": 17820 }, { "epoch": 26.67814371257485, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 17821 }, { "epoch": 26.679640718562876, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17822 }, { "epoch": 26.681137724550897, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 17823 }, { "epoch": 26.682634730538922, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1323, "step": 17824 }, { "epoch": 26.684131736526947, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.133, "step": 17825 }, { "epoch": 26.68562874251497, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1261, "step": 17826 }, { "epoch": 26.687125748502993, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1314, "step": 17827 }, { "epoch": 26.688622754491018, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17828 }, { "epoch": 26.690119760479043, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.128, "step": 17829 }, { "epoch": 26.691616766467067, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17830 }, { "epoch": 26.69311377245509, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17831 }, { "epoch": 26.694610778443113, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1294, "step": 17832 }, { "epoch": 26.69610778443114, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 17833 }, { "epoch": 26.697604790419163, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1257, "step": 17834 }, { "epoch": 26.699101796407184, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1292, "step": 17835 }, { "epoch": 26.70059880239521, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1315, "step": 17836 }, { "epoch": 26.702095808383234, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1318, "step": 17837 }, { "epoch": 26.70359281437126, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.13, "step": 17838 }, { "epoch": 26.70508982035928, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17839 }, { "epoch": 26.706586826347305, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1207, "step": 17840 }, { "epoch": 26.70808383233533, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1316, "step": 17841 }, { "epoch": 26.709580838323355, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1348, "step": 17842 }, { "epoch": 26.711077844311376, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17843 }, { "epoch": 26.7125748502994, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1289, "step": 17844 }, { "epoch": 26.714071856287426, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1289, "step": 17845 }, { "epoch": 26.71556886227545, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1335, "step": 17846 }, { "epoch": 26.71706586826347, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1234, "step": 17847 }, { "epoch": 26.718562874251496, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1265, "step": 17848 }, { "epoch": 26.72005988023952, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17849 }, { "epoch": 26.721556886227546, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1299, "step": 17850 }, { "epoch": 26.723053892215567, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 17851 }, { "epoch": 26.724550898203592, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1303, "step": 17852 }, { "epoch": 26.726047904191617, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1334, "step": 17853 }, { "epoch": 26.727544910179642, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1218, "step": 17854 }, { "epoch": 26.729041916167663, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1332, "step": 17855 }, { "epoch": 26.730538922155688, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1365, "step": 17856 }, { "epoch": 26.732035928143713, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1268, "step": 17857 }, { "epoch": 26.733532934131738, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1346, "step": 17858 }, { "epoch": 26.73502994011976, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 17859 }, { "epoch": 26.736526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 17860 }, { "epoch": 26.73802395209581, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 17861 }, { "epoch": 26.739520958083833, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1349, "step": 17862 }, { "epoch": 26.741017964071855, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.131, "step": 17863 }, { "epoch": 26.74251497005988, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1275, "step": 17864 }, { "epoch": 26.744011976047904, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1274, "step": 17865 }, { "epoch": 26.74550898203593, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1314, "step": 17866 }, { "epoch": 26.74700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1288, "step": 17867 }, { "epoch": 26.748502994011975, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1281, "step": 17868 }, { "epoch": 26.75, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1285, "step": 17869 }, { "epoch": 26.751497005988025, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1322, "step": 17870 }, { "epoch": 26.75299401197605, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1263, "step": 17871 }, { "epoch": 26.75449101796407, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1254, "step": 17872 }, { "epoch": 26.755988023952096, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1273, "step": 17873 }, { "epoch": 26.75748502994012, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1254, "step": 17874 }, { "epoch": 26.758982035928145, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1311, "step": 17875 }, { "epoch": 26.760479041916167, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.129, "step": 17876 }, { "epoch": 26.76197604790419, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1242, "step": 17877 }, { "epoch": 26.763473053892216, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1249, "step": 17878 }, { "epoch": 26.76497005988024, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1251, "step": 17879 }, { "epoch": 26.766467065868262, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1287, "step": 17880 }, { "epoch": 26.767964071856287, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1273, "step": 17881 }, { "epoch": 26.769461077844312, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1372, "step": 17882 }, { "epoch": 26.770958083832337, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17883 }, { "epoch": 26.772455089820358, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 17884 }, { "epoch": 26.773952095808383, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1341, "step": 17885 }, { "epoch": 26.775449101796408, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1331, "step": 17886 }, { "epoch": 26.776946107784433, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1191, "step": 17887 }, { "epoch": 26.778443113772454, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1312, "step": 17888 }, { "epoch": 26.77994011976048, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.13, "step": 17889 }, { "epoch": 26.781437125748504, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17890 }, { "epoch": 26.78293413173653, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1288, "step": 17891 }, { "epoch": 26.78443113772455, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1274, "step": 17892 }, { "epoch": 26.785928143712574, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 17893 }, { "epoch": 26.7874251497006, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1267, "step": 17894 }, { "epoch": 26.788922155688624, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1296, "step": 17895 }, { "epoch": 26.790419161676645, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1311, "step": 17896 }, { "epoch": 26.79191616766467, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 17897 }, { "epoch": 26.793413173652695, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1342, "step": 17898 }, { "epoch": 26.79491017964072, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 17899 }, { "epoch": 26.79640718562874, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1248, "step": 17900 }, { "epoch": 26.797904191616766, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.134, "step": 17901 }, { "epoch": 26.79940119760479, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1293, "step": 17902 }, { "epoch": 26.800898203592816, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 17903 }, { "epoch": 26.802395209580837, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1254, "step": 17904 }, { "epoch": 26.80389221556886, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1317, "step": 17905 }, { "epoch": 26.805389221556887, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1313, "step": 17906 }, { "epoch": 26.80688622754491, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.122, "step": 17907 }, { "epoch": 26.808383233532933, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 17908 }, { "epoch": 26.809880239520957, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1265, "step": 17909 }, { "epoch": 26.811377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 17910 }, { "epoch": 26.812874251497007, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.125, "step": 17911 }, { "epoch": 26.81437125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1298, "step": 17912 }, { "epoch": 26.815868263473053, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1304, "step": 17913 }, { "epoch": 26.817365269461078, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 17914 }, { "epoch": 26.818862275449103, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.127, "step": 17915 }, { "epoch": 26.820359281437124, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 17916 }, { "epoch": 26.82185628742515, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1282, "step": 17917 }, { "epoch": 26.823353293413174, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1282, "step": 17918 }, { "epoch": 26.8248502994012, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1272, "step": 17919 }, { "epoch": 26.82634730538922, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1263, "step": 17920 }, { "epoch": 26.827844311377245, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.129, "step": 17921 }, { "epoch": 26.82934131736527, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1329, "step": 17922 }, { "epoch": 26.830838323353294, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1306, "step": 17923 }, { "epoch": 26.83233532934132, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 17924 }, { "epoch": 26.83383233532934, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1291, "step": 17925 }, { "epoch": 26.835329341317365, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1281, "step": 17926 }, { "epoch": 26.83682634730539, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1325, "step": 17927 }, { "epoch": 26.83832335329341, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.13, "step": 17928 }, { "epoch": 26.839820359281436, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17929 }, { "epoch": 26.84131736526946, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1232, "step": 17930 }, { "epoch": 26.842814371257486, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1348, "step": 17931 }, { "epoch": 26.84431137724551, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1358, "step": 17932 }, { "epoch": 26.845808383233532, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17933 }, { "epoch": 26.847305389221557, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1327, "step": 17934 }, { "epoch": 26.84880239520958, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1321, "step": 17935 }, { "epoch": 26.850299401197606, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 17936 }, { "epoch": 26.851796407185628, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1336, "step": 17937 }, { "epoch": 26.853293413173652, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1253, "step": 17938 }, { "epoch": 26.854790419161677, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17939 }, { "epoch": 26.856287425149702, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1375, "step": 17940 }, { "epoch": 26.857784431137723, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1306, "step": 17941 }, { "epoch": 26.85928143712575, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1288, "step": 17942 }, { "epoch": 26.860778443113773, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1278, "step": 17943 }, { "epoch": 26.862275449101798, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 17944 }, { "epoch": 26.86377245508982, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1325, "step": 17945 }, { "epoch": 26.865269461077844, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1321, "step": 17946 }, { "epoch": 26.86676646706587, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1271, "step": 17947 }, { "epoch": 26.868263473053894, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1279, "step": 17948 }, { "epoch": 26.869760479041915, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1284, "step": 17949 }, { "epoch": 26.87125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1283, "step": 17950 }, { "epoch": 26.872754491017965, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1253, "step": 17951 }, { "epoch": 26.87425149700599, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1284, "step": 17952 }, { "epoch": 26.87574850299401, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1279, "step": 17953 }, { "epoch": 26.877245508982035, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1291, "step": 17954 }, { "epoch": 26.87874251497006, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1281, "step": 17955 }, { "epoch": 26.880239520958085, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1318, "step": 17956 }, { "epoch": 26.881736526946106, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.129, "step": 17957 }, { "epoch": 26.88323353293413, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 17958 }, { "epoch": 26.884730538922156, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1257, "step": 17959 }, { "epoch": 26.88622754491018, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17960 }, { "epoch": 26.887724550898202, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.131, "step": 17961 }, { "epoch": 26.889221556886227, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1274, "step": 17962 }, { "epoch": 26.89071856287425, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.131, "step": 17963 }, { "epoch": 26.892215568862277, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1294, "step": 17964 }, { "epoch": 26.893712574850298, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1323, "step": 17965 }, { "epoch": 26.895209580838323, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 17966 }, { "epoch": 26.896706586826348, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1356, "step": 17967 }, { "epoch": 26.898203592814372, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1316, "step": 17968 }, { "epoch": 26.899700598802394, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17969 }, { "epoch": 26.90119760479042, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.131, "step": 17970 }, { "epoch": 26.902694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1335, "step": 17971 }, { "epoch": 26.904191616766468, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1383, "step": 17972 }, { "epoch": 26.90568862275449, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 17973 }, { "epoch": 26.907185628742514, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1295, "step": 17974 }, { "epoch": 26.90868263473054, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1299, "step": 17975 }, { "epoch": 26.910179640718564, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1281, "step": 17976 }, { "epoch": 26.91167664670659, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.123, "step": 17977 }, { "epoch": 26.91317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1354, "step": 17978 }, { "epoch": 26.914670658682635, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 17979 }, { "epoch": 26.91616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1314, "step": 17980 }, { "epoch": 26.91766467065868, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.127, "step": 17981 }, { "epoch": 26.919161676646706, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1312, "step": 17982 }, { "epoch": 26.92065868263473, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1256, "step": 17983 }, { "epoch": 26.922155688622755, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.128, "step": 17984 }, { "epoch": 26.92365269461078, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1317, "step": 17985 }, { "epoch": 26.9251497005988, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1238, "step": 17986 }, { "epoch": 26.926646706586826, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17987 }, { "epoch": 26.92814371257485, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1315, "step": 17988 }, { "epoch": 26.929640718562876, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1298, "step": 17989 }, { "epoch": 26.931137724550897, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1283, "step": 17990 }, { "epoch": 26.932634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1299, "step": 17991 }, { "epoch": 26.934131736526947, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1281, "step": 17992 }, { "epoch": 26.93562874251497, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.125, "step": 17993 }, { "epoch": 26.937125748502993, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1279, "step": 17994 }, { "epoch": 26.938622754491018, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1262, "step": 17995 }, { "epoch": 26.940119760479043, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1315, "step": 17996 }, { "epoch": 26.941616766467067, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1318, "step": 17997 }, { "epoch": 26.94311377245509, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1314, "step": 17998 }, { "epoch": 26.944610778443113, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1294, "step": 17999 }, { "epoch": 26.94610778443114, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1291, "step": 18000 }, { "epoch": 26.947604790419163, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1255, "step": 18001 }, { "epoch": 26.949101796407184, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1312, "step": 18002 }, { "epoch": 26.95059880239521, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.133, "step": 18003 }, { "epoch": 26.952095808383234, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1319, "step": 18004 }, { "epoch": 26.95359281437126, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1332, "step": 18005 }, { "epoch": 26.95508982035928, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1308, "step": 18006 }, { "epoch": 26.956586826347305, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.128, "step": 18007 }, { "epoch": 26.95808383233533, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1252, "step": 18008 }, { "epoch": 26.959580838323355, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1277, "step": 18009 }, { "epoch": 26.961077844311376, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1312, "step": 18010 }, { "epoch": 26.9625748502994, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.135, "step": 18011 }, { "epoch": 26.964071856287426, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1251, "step": 18012 }, { "epoch": 26.96556886227545, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1277, "step": 18013 }, { "epoch": 26.96706586826347, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1261, "step": 18014 }, { "epoch": 26.968562874251496, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1303, "step": 18015 }, { "epoch": 26.97005988023952, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1309, "step": 18016 }, { "epoch": 26.971556886227546, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.126, "step": 18017 }, { "epoch": 26.973053892215567, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1263, "step": 18018 }, { "epoch": 26.974550898203592, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1308, "step": 18019 }, { "epoch": 26.976047904191617, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1332, "step": 18020 }, { "epoch": 26.977544910179642, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1308, "step": 18021 }, { "epoch": 26.979041916167663, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.13, "step": 18022 }, { "epoch": 26.980538922155688, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1201, "step": 18023 }, { "epoch": 26.982035928143713, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1316, "step": 18024 }, { "epoch": 26.983532934131738, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1303, "step": 18025 }, { "epoch": 26.98502994011976, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.131, "step": 18026 }, { "epoch": 26.986526946107784, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1313, "step": 18027 }, { "epoch": 26.98802395209581, "grad_norm": 0.1689453125, "learning_rate": 0.0008, "loss": 1.1338, "step": 18028 }, { "epoch": 26.989520958083833, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1384, "step": 18029 }, { "epoch": 26.991017964071855, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1309, "step": 18030 }, { "epoch": 26.99251497005988, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1374, "step": 18031 }, { "epoch": 26.994011976047904, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1345, "step": 18032 }, { "epoch": 26.99550898203593, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.129, "step": 18033 }, { "epoch": 26.99700598802395, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1236, "step": 18034 }, { "epoch": 26.998502994011975, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1287, "step": 18035 }, { "epoch": 27.0, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 18036 }, { "epoch": 27.001497005988025, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1258, "step": 18037 }, { "epoch": 27.00299401197605, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.128, "step": 18038 }, { "epoch": 27.00449101796407, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18039 }, { "epoch": 27.005988023952096, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1239, "step": 18040 }, { "epoch": 27.00748502994012, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1299, "step": 18041 }, { "epoch": 27.008982035928145, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1332, "step": 18042 }, { "epoch": 27.010479041916167, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1286, "step": 18043 }, { "epoch": 27.01197604790419, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1332, "step": 18044 }, { "epoch": 27.013473053892216, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1307, "step": 18045 }, { "epoch": 27.01497005988024, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1291, "step": 18046 }, { "epoch": 27.016467065868262, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1269, "step": 18047 }, { "epoch": 27.017964071856287, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1231, "step": 18048 }, { "epoch": 27.019461077844312, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1301, "step": 18049 }, { "epoch": 27.020958083832337, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1255, "step": 18050 }, { "epoch": 27.022455089820358, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1317, "step": 18051 }, { "epoch": 27.023952095808383, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1326, "step": 18052 }, { "epoch": 27.025449101796408, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1266, "step": 18053 }, { "epoch": 27.026946107784433, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1216, "step": 18054 }, { "epoch": 27.028443113772454, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1325, "step": 18055 }, { "epoch": 27.02994011976048, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18056 }, { "epoch": 27.031437125748504, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 18057 }, { "epoch": 27.03293413173653, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1233, "step": 18058 }, { "epoch": 27.03443113772455, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1326, "step": 18059 }, { "epoch": 27.035928143712574, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1315, "step": 18060 }, { "epoch": 27.0374251497006, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1307, "step": 18061 }, { "epoch": 27.038922155688624, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1286, "step": 18062 }, { "epoch": 27.040419161676645, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1327, "step": 18063 }, { "epoch": 27.04191616766467, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 18064 }, { "epoch": 27.043413173652695, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1278, "step": 18065 }, { "epoch": 27.04491017964072, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1279, "step": 18066 }, { "epoch": 27.04640718562874, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18067 }, { "epoch": 27.047904191616766, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1297, "step": 18068 }, { "epoch": 27.04940119760479, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1291, "step": 18069 }, { "epoch": 27.050898203592816, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1242, "step": 18070 }, { "epoch": 27.052395209580837, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 18071 }, { "epoch": 27.05389221556886, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1327, "step": 18072 }, { "epoch": 27.055389221556887, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1207, "step": 18073 }, { "epoch": 27.05688622754491, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1328, "step": 18074 }, { "epoch": 27.058383233532933, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.13, "step": 18075 }, { "epoch": 27.059880239520957, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1281, "step": 18076 }, { "epoch": 27.061377245508982, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1277, "step": 18077 }, { "epoch": 27.062874251497007, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1242, "step": 18078 }, { "epoch": 27.06437125748503, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 18079 }, { "epoch": 27.065868263473053, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1243, "step": 18080 }, { "epoch": 27.067365269461078, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1336, "step": 18081 }, { "epoch": 27.068862275449103, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1327, "step": 18082 }, { "epoch": 27.070359281437124, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1378, "step": 18083 }, { "epoch": 27.07185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.127, "step": 18084 }, { "epoch": 27.073353293413174, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1301, "step": 18085 }, { "epoch": 27.0748502994012, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1294, "step": 18086 }, { "epoch": 27.07634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1299, "step": 18087 }, { "epoch": 27.077844311377245, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1333, "step": 18088 }, { "epoch": 27.07934131736527, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1283, "step": 18089 }, { "epoch": 27.080838323353294, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 18090 }, { "epoch": 27.082335329341316, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1255, "step": 18091 }, { "epoch": 27.08383233532934, "grad_norm": 0.048828125, "learning_rate": 0.0008, "loss": 1.1289, "step": 18092 }, { "epoch": 27.085329341317365, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1283, "step": 18093 }, { "epoch": 27.08682634730539, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1187, "step": 18094 }, { "epoch": 27.088323353293415, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1301, "step": 18095 }, { "epoch": 27.089820359281436, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1237, "step": 18096 }, { "epoch": 27.09131736526946, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1258, "step": 18097 }, { "epoch": 27.092814371257486, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1368, "step": 18098 }, { "epoch": 27.09431137724551, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1261, "step": 18099 }, { "epoch": 27.095808383233532, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1287, "step": 18100 }, { "epoch": 27.097305389221557, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18101 }, { "epoch": 27.09880239520958, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1313, "step": 18102 }, { "epoch": 27.100299401197606, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1221, "step": 18103 }, { "epoch": 27.101796407185628, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1338, "step": 18104 }, { "epoch": 27.103293413173652, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1314, "step": 18105 }, { "epoch": 27.104790419161677, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1314, "step": 18106 }, { "epoch": 27.106287425149702, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.13, "step": 18107 }, { "epoch": 27.107784431137723, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1324, "step": 18108 }, { "epoch": 27.10928143712575, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1329, "step": 18109 }, { "epoch": 27.110778443113773, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1309, "step": 18110 }, { "epoch": 27.112275449101798, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 18111 }, { "epoch": 27.11377245508982, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18112 }, { "epoch": 27.115269461077844, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1268, "step": 18113 }, { "epoch": 27.11676646706587, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1258, "step": 18114 }, { "epoch": 27.118263473053894, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 18115 }, { "epoch": 27.119760479041915, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1282, "step": 18116 }, { "epoch": 27.12125748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 18117 }, { "epoch": 27.122754491017965, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1284, "step": 18118 }, { "epoch": 27.12425149700599, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1295, "step": 18119 }, { "epoch": 27.12574850299401, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 18120 }, { "epoch": 27.127245508982035, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1183, "step": 18121 }, { "epoch": 27.12874251497006, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1308, "step": 18122 }, { "epoch": 27.130239520958085, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1334, "step": 18123 }, { "epoch": 27.131736526946106, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 18124 }, { "epoch": 27.13323353293413, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1295, "step": 18125 }, { "epoch": 27.134730538922156, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1275, "step": 18126 }, { "epoch": 27.13622754491018, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1258, "step": 18127 }, { "epoch": 27.137724550898202, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1257, "step": 18128 }, { "epoch": 27.139221556886227, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1309, "step": 18129 }, { "epoch": 27.14071856287425, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1246, "step": 18130 }, { "epoch": 27.142215568862277, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1315, "step": 18131 }, { "epoch": 27.143712574850298, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1227, "step": 18132 }, { "epoch": 27.145209580838323, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1281, "step": 18133 }, { "epoch": 27.146706586826348, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1327, "step": 18134 }, { "epoch": 27.148203592814372, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1332, "step": 18135 }, { "epoch": 27.149700598802394, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1327, "step": 18136 }, { "epoch": 27.15119760479042, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1283, "step": 18137 }, { "epoch": 27.152694610778443, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1356, "step": 18138 }, { "epoch": 27.154191616766468, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1286, "step": 18139 }, { "epoch": 27.15568862275449, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1348, "step": 18140 }, { "epoch": 27.157185628742514, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1337, "step": 18141 }, { "epoch": 27.15868263473054, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.128, "step": 18142 }, { "epoch": 27.160179640718564, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1307, "step": 18143 }, { "epoch": 27.161676646706585, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1224, "step": 18144 }, { "epoch": 27.16317365269461, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18145 }, { "epoch": 27.164670658682635, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1228, "step": 18146 }, { "epoch": 27.16616766467066, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1296, "step": 18147 }, { "epoch": 27.16766467065868, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1305, "step": 18148 }, { "epoch": 27.169161676646706, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18149 }, { "epoch": 27.17065868263473, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1241, "step": 18150 }, { "epoch": 27.172155688622755, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1311, "step": 18151 }, { "epoch": 27.17365269461078, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1293, "step": 18152 }, { "epoch": 27.1751497005988, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.128, "step": 18153 }, { "epoch": 27.176646706586826, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18154 }, { "epoch": 27.17814371257485, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1276, "step": 18155 }, { "epoch": 27.179640718562876, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1244, "step": 18156 }, { "epoch": 27.181137724550897, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1288, "step": 18157 }, { "epoch": 27.182634730538922, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1312, "step": 18158 }, { "epoch": 27.184131736526947, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1256, "step": 18159 }, { "epoch": 27.18562874251497, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1295, "step": 18160 }, { "epoch": 27.187125748502993, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1281, "step": 18161 }, { "epoch": 27.188622754491018, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1257, "step": 18162 }, { "epoch": 27.190119760479043, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1282, "step": 18163 }, { "epoch": 27.191616766467067, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1304, "step": 18164 }, { "epoch": 27.19311377245509, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1285, "step": 18165 }, { "epoch": 27.194610778443113, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1263, "step": 18166 }, { "epoch": 27.19610778443114, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1301, "step": 18167 }, { "epoch": 27.197604790419163, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1304, "step": 18168 }, { "epoch": 27.199101796407184, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1243, "step": 18169 }, { "epoch": 27.20059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1211, "step": 18170 }, { "epoch": 27.202095808383234, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 18171 }, { "epoch": 27.20359281437126, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1209, "step": 18172 }, { "epoch": 27.20508982035928, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 18173 }, { "epoch": 27.206586826347305, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1302, "step": 18174 }, { "epoch": 27.20808383233533, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1318, "step": 18175 }, { "epoch": 27.209580838323355, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1336, "step": 18176 }, { "epoch": 27.211077844311376, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1273, "step": 18177 }, { "epoch": 27.2125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1297, "step": 18178 }, { "epoch": 27.214071856287426, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 18179 }, { "epoch": 27.21556886227545, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1284, "step": 18180 }, { "epoch": 27.21706586826347, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1357, "step": 18181 }, { "epoch": 27.218562874251496, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.125, "step": 18182 }, { "epoch": 27.22005988023952, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1301, "step": 18183 }, { "epoch": 27.221556886227546, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1248, "step": 18184 }, { "epoch": 27.223053892215567, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1286, "step": 18185 }, { "epoch": 27.224550898203592, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1322, "step": 18186 }, { "epoch": 27.226047904191617, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1297, "step": 18187 }, { "epoch": 27.227544910179642, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1308, "step": 18188 }, { "epoch": 27.229041916167663, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1295, "step": 18189 }, { "epoch": 27.230538922155688, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1259, "step": 18190 }, { "epoch": 27.232035928143713, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1299, "step": 18191 }, { "epoch": 27.233532934131738, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1256, "step": 18192 }, { "epoch": 27.23502994011976, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1302, "step": 18193 }, { "epoch": 27.236526946107784, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1364, "step": 18194 }, { "epoch": 27.23802395209581, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1307, "step": 18195 }, { "epoch": 27.239520958083833, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1276, "step": 18196 }, { "epoch": 27.241017964071855, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.125, "step": 18197 }, { "epoch": 27.24251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1289, "step": 18198 }, { "epoch": 27.244011976047904, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1291, "step": 18199 }, { "epoch": 27.24550898203593, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18200 }, { "epoch": 27.24700598802395, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1314, "step": 18201 }, { "epoch": 27.248502994011975, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1356, "step": 18202 }, { "epoch": 27.25, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1352, "step": 18203 }, { "epoch": 27.251497005988025, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18204 }, { "epoch": 27.25299401197605, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1349, "step": 18205 }, { "epoch": 27.25449101796407, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 18206 }, { "epoch": 27.255988023952096, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1312, "step": 18207 }, { "epoch": 27.25748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1367, "step": 18208 }, { "epoch": 27.258982035928145, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.13, "step": 18209 }, { "epoch": 27.260479041916167, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1271, "step": 18210 }, { "epoch": 27.26197604790419, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1321, "step": 18211 }, { "epoch": 27.263473053892216, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.128, "step": 18212 }, { "epoch": 27.26497005988024, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1303, "step": 18213 }, { "epoch": 27.266467065868262, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18214 }, { "epoch": 27.267964071856287, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1249, "step": 18215 }, { "epoch": 27.269461077844312, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1325, "step": 18216 }, { "epoch": 27.270958083832337, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1271, "step": 18217 }, { "epoch": 27.272455089820358, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1273, "step": 18218 }, { "epoch": 27.273952095808383, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1327, "step": 18219 }, { "epoch": 27.275449101796408, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1268, "step": 18220 }, { "epoch": 27.276946107784433, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1355, "step": 18221 }, { "epoch": 27.278443113772454, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 18222 }, { "epoch": 27.27994011976048, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1248, "step": 18223 }, { "epoch": 27.281437125748504, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1295, "step": 18224 }, { "epoch": 27.28293413173653, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1338, "step": 18225 }, { "epoch": 27.28443113772455, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18226 }, { "epoch": 27.285928143712574, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1345, "step": 18227 }, { "epoch": 27.2874251497006, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1305, "step": 18228 }, { "epoch": 27.288922155688624, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1247, "step": 18229 }, { "epoch": 27.290419161676645, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.13, "step": 18230 }, { "epoch": 27.29191616766467, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1324, "step": 18231 }, { "epoch": 27.293413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 18232 }, { "epoch": 27.29491017964072, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1294, "step": 18233 }, { "epoch": 27.29640718562874, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1307, "step": 18234 }, { "epoch": 27.297904191616766, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1258, "step": 18235 }, { "epoch": 27.29940119760479, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1263, "step": 18236 }, { "epoch": 27.300898203592816, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1314, "step": 18237 }, { "epoch": 27.302395209580837, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1306, "step": 18238 }, { "epoch": 27.30389221556886, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1226, "step": 18239 }, { "epoch": 27.305389221556887, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1328, "step": 18240 }, { "epoch": 27.30688622754491, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1196, "step": 18241 }, { "epoch": 27.308383233532933, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1331, "step": 18242 }, { "epoch": 27.309880239520957, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1331, "step": 18243 }, { "epoch": 27.311377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1221, "step": 18244 }, { "epoch": 27.312874251497007, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1243, "step": 18245 }, { "epoch": 27.31437125748503, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 18246 }, { "epoch": 27.315868263473053, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 18247 }, { "epoch": 27.317365269461078, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1271, "step": 18248 }, { "epoch": 27.318862275449103, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 18249 }, { "epoch": 27.320359281437124, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1309, "step": 18250 }, { "epoch": 27.32185628742515, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1315, "step": 18251 }, { "epoch": 27.323353293413174, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1316, "step": 18252 }, { "epoch": 27.3248502994012, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18253 }, { "epoch": 27.32634730538922, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1237, "step": 18254 }, { "epoch": 27.327844311377245, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1295, "step": 18255 }, { "epoch": 27.32934131736527, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 18256 }, { "epoch": 27.330838323353294, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1233, "step": 18257 }, { "epoch": 27.33233532934132, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1283, "step": 18258 }, { "epoch": 27.33383233532934, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1255, "step": 18259 }, { "epoch": 27.335329341317365, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.129, "step": 18260 }, { "epoch": 27.33682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1328, "step": 18261 }, { "epoch": 27.338323353293415, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1291, "step": 18262 }, { "epoch": 27.339820359281436, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1272, "step": 18263 }, { "epoch": 27.34131736526946, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1321, "step": 18264 }, { "epoch": 27.342814371257486, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1302, "step": 18265 }, { "epoch": 27.34431137724551, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1281, "step": 18266 }, { "epoch": 27.345808383233532, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1248, "step": 18267 }, { "epoch": 27.347305389221557, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1326, "step": 18268 }, { "epoch": 27.34880239520958, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1274, "step": 18269 }, { "epoch": 27.350299401197606, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 18270 }, { "epoch": 27.351796407185628, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1309, "step": 18271 }, { "epoch": 27.353293413173652, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.13, "step": 18272 }, { "epoch": 27.354790419161677, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18273 }, { "epoch": 27.356287425149702, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1184, "step": 18274 }, { "epoch": 27.357784431137723, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1271, "step": 18275 }, { "epoch": 27.35928143712575, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1319, "step": 18276 }, { "epoch": 27.360778443113773, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1291, "step": 18277 }, { "epoch": 27.362275449101798, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.127, "step": 18278 }, { "epoch": 27.36377245508982, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1324, "step": 18279 }, { "epoch": 27.365269461077844, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1316, "step": 18280 }, { "epoch": 27.36676646706587, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.129, "step": 18281 }, { "epoch": 27.368263473053894, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1309, "step": 18282 }, { "epoch": 27.369760479041915, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1301, "step": 18283 }, { "epoch": 27.37125748502994, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1318, "step": 18284 }, { "epoch": 27.372754491017965, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1297, "step": 18285 }, { "epoch": 27.37425149700599, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.128, "step": 18286 }, { "epoch": 27.37574850299401, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1334, "step": 18287 }, { "epoch": 27.377245508982035, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1265, "step": 18288 }, { "epoch": 27.37874251497006, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1223, "step": 18289 }, { "epoch": 27.380239520958085, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1327, "step": 18290 }, { "epoch": 27.381736526946106, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1306, "step": 18291 }, { "epoch": 27.38323353293413, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.129, "step": 18292 }, { "epoch": 27.384730538922156, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1255, "step": 18293 }, { "epoch": 27.38622754491018, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1291, "step": 18294 }, { "epoch": 27.387724550898202, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18295 }, { "epoch": 27.389221556886227, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1281, "step": 18296 }, { "epoch": 27.39071856287425, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1252, "step": 18297 }, { "epoch": 27.392215568862277, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1264, "step": 18298 }, { "epoch": 27.393712574850298, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18299 }, { "epoch": 27.395209580838323, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1333, "step": 18300 }, { "epoch": 27.396706586826348, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.129, "step": 18301 }, { "epoch": 27.398203592814372, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1274, "step": 18302 }, { "epoch": 27.399700598802394, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1344, "step": 18303 }, { "epoch": 27.40119760479042, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 18304 }, { "epoch": 27.402694610778443, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1314, "step": 18305 }, { "epoch": 27.404191616766468, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1222, "step": 18306 }, { "epoch": 27.40568862275449, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1314, "step": 18307 }, { "epoch": 27.407185628742514, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1303, "step": 18308 }, { "epoch": 27.40868263473054, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 18309 }, { "epoch": 27.410179640718564, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1268, "step": 18310 }, { "epoch": 27.411676646706585, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 18311 }, { "epoch": 27.41317365269461, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1306, "step": 18312 }, { "epoch": 27.414670658682635, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1384, "step": 18313 }, { "epoch": 27.41616766467066, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1254, "step": 18314 }, { "epoch": 27.41766467065868, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 18315 }, { "epoch": 27.419161676646706, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1267, "step": 18316 }, { "epoch": 27.42065868263473, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 18317 }, { "epoch": 27.422155688622755, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 18318 }, { "epoch": 27.42365269461078, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 18319 }, { "epoch": 27.4251497005988, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.136, "step": 18320 }, { "epoch": 27.426646706586826, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1287, "step": 18321 }, { "epoch": 27.42814371257485, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1355, "step": 18322 }, { "epoch": 27.429640718562876, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1338, "step": 18323 }, { "epoch": 27.431137724550897, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1255, "step": 18324 }, { "epoch": 27.432634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1246, "step": 18325 }, { "epoch": 27.434131736526947, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1294, "step": 18326 }, { "epoch": 27.43562874251497, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1283, "step": 18327 }, { "epoch": 27.437125748502993, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.131, "step": 18328 }, { "epoch": 27.438622754491018, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1259, "step": 18329 }, { "epoch": 27.440119760479043, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 18330 }, { "epoch": 27.441616766467067, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1316, "step": 18331 }, { "epoch": 27.44311377245509, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 18332 }, { "epoch": 27.444610778443113, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1304, "step": 18333 }, { "epoch": 27.44610778443114, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18334 }, { "epoch": 27.447604790419163, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.125, "step": 18335 }, { "epoch": 27.449101796407184, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1277, "step": 18336 }, { "epoch": 27.45059880239521, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1208, "step": 18337 }, { "epoch": 27.452095808383234, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1319, "step": 18338 }, { "epoch": 27.45359281437126, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1258, "step": 18339 }, { "epoch": 27.45508982035928, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.134, "step": 18340 }, { "epoch": 27.456586826347305, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.127, "step": 18341 }, { "epoch": 27.45808383233533, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 18342 }, { "epoch": 27.459580838323355, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1251, "step": 18343 }, { "epoch": 27.461077844311376, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1248, "step": 18344 }, { "epoch": 27.4625748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1275, "step": 18345 }, { "epoch": 27.464071856287426, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.131, "step": 18346 }, { "epoch": 27.46556886227545, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1288, "step": 18347 }, { "epoch": 27.46706586826347, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1258, "step": 18348 }, { "epoch": 27.468562874251496, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1238, "step": 18349 }, { "epoch": 27.47005988023952, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1332, "step": 18350 }, { "epoch": 27.471556886227546, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1299, "step": 18351 }, { "epoch": 27.473053892215567, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1316, "step": 18352 }, { "epoch": 27.474550898203592, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1317, "step": 18353 }, { "epoch": 27.476047904191617, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1186, "step": 18354 }, { "epoch": 27.477544910179642, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1276, "step": 18355 }, { "epoch": 27.479041916167663, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1196, "step": 18356 }, { "epoch": 27.480538922155688, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1328, "step": 18357 }, { "epoch": 27.482035928143713, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1323, "step": 18358 }, { "epoch": 27.483532934131738, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.135, "step": 18359 }, { "epoch": 27.48502994011976, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 18360 }, { "epoch": 27.486526946107784, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18361 }, { "epoch": 27.48802395209581, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1281, "step": 18362 }, { "epoch": 27.489520958083833, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1302, "step": 18363 }, { "epoch": 27.491017964071855, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1256, "step": 18364 }, { "epoch": 27.49251497005988, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.127, "step": 18365 }, { "epoch": 27.494011976047904, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18366 }, { "epoch": 27.49550898203593, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1226, "step": 18367 }, { "epoch": 27.49700598802395, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1315, "step": 18368 }, { "epoch": 27.498502994011975, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1316, "step": 18369 }, { "epoch": 27.5, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.133, "step": 18370 }, { "epoch": 27.501497005988025, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 18371 }, { "epoch": 27.50299401197605, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.134, "step": 18372 }, { "epoch": 27.50449101796407, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1329, "step": 18373 }, { "epoch": 27.505988023952096, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1305, "step": 18374 }, { "epoch": 27.50748502994012, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1306, "step": 18375 }, { "epoch": 27.508982035928145, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1245, "step": 18376 }, { "epoch": 27.510479041916167, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1336, "step": 18377 }, { "epoch": 27.51197604790419, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 18378 }, { "epoch": 27.513473053892216, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1286, "step": 18379 }, { "epoch": 27.51497005988024, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 18380 }, { "epoch": 27.516467065868262, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 18381 }, { "epoch": 27.517964071856287, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1281, "step": 18382 }, { "epoch": 27.519461077844312, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1292, "step": 18383 }, { "epoch": 27.520958083832337, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.13, "step": 18384 }, { "epoch": 27.522455089820358, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1241, "step": 18385 }, { "epoch": 27.523952095808383, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.125, "step": 18386 }, { "epoch": 27.525449101796408, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1269, "step": 18387 }, { "epoch": 27.526946107784433, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1299, "step": 18388 }, { "epoch": 27.528443113772454, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 18389 }, { "epoch": 27.52994011976048, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1303, "step": 18390 }, { "epoch": 27.531437125748504, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.129, "step": 18391 }, { "epoch": 27.53293413173653, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1328, "step": 18392 }, { "epoch": 27.53443113772455, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 18393 }, { "epoch": 27.535928143712574, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1275, "step": 18394 }, { "epoch": 27.5374251497006, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1306, "step": 18395 }, { "epoch": 27.538922155688624, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1242, "step": 18396 }, { "epoch": 27.540419161676645, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1385, "step": 18397 }, { "epoch": 27.54191616766467, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.126, "step": 18398 }, { "epoch": 27.543413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1244, "step": 18399 }, { "epoch": 27.54491017964072, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1226, "step": 18400 }, { "epoch": 27.54640718562874, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1322, "step": 18401 }, { "epoch": 27.547904191616766, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1312, "step": 18402 }, { "epoch": 27.54940119760479, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1253, "step": 18403 }, { "epoch": 27.550898203592816, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1355, "step": 18404 }, { "epoch": 27.552395209580837, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1296, "step": 18405 }, { "epoch": 27.55389221556886, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1326, "step": 18406 }, { "epoch": 27.555389221556887, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1283, "step": 18407 }, { "epoch": 27.55688622754491, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1284, "step": 18408 }, { "epoch": 27.558383233532933, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.135, "step": 18409 }, { "epoch": 27.559880239520957, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1254, "step": 18410 }, { "epoch": 27.561377245508982, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1321, "step": 18411 }, { "epoch": 27.562874251497007, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1321, "step": 18412 }, { "epoch": 27.56437125748503, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1359, "step": 18413 }, { "epoch": 27.565868263473053, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.13, "step": 18414 }, { "epoch": 27.567365269461078, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 18415 }, { "epoch": 27.568862275449103, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.135, "step": 18416 }, { "epoch": 27.570359281437124, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1268, "step": 18417 }, { "epoch": 27.57185628742515, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1296, "step": 18418 }, { "epoch": 27.573353293413174, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 18419 }, { "epoch": 27.5748502994012, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.126, "step": 18420 }, { "epoch": 27.57634730538922, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18421 }, { "epoch": 27.577844311377245, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1265, "step": 18422 }, { "epoch": 27.57934131736527, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1228, "step": 18423 }, { "epoch": 27.580838323353294, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1394, "step": 18424 }, { "epoch": 27.58233532934132, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1254, "step": 18425 }, { "epoch": 27.58383233532934, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1237, "step": 18426 }, { "epoch": 27.585329341317365, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1269, "step": 18427 }, { "epoch": 27.58682634730539, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 18428 }, { "epoch": 27.58832335329341, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1392, "step": 18429 }, { "epoch": 27.589820359281436, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 18430 }, { "epoch": 27.59131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 18431 }, { "epoch": 27.592814371257486, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18432 }, { "epoch": 27.59431137724551, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1342, "step": 18433 }, { "epoch": 27.595808383233532, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1295, "step": 18434 }, { "epoch": 27.597305389221557, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1318, "step": 18435 }, { "epoch": 27.59880239520958, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 18436 }, { "epoch": 27.600299401197606, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1268, "step": 18437 }, { "epoch": 27.601796407185628, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1305, "step": 18438 }, { "epoch": 27.603293413173652, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1329, "step": 18439 }, { "epoch": 27.604790419161677, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18440 }, { "epoch": 27.606287425149702, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1314, "step": 18441 }, { "epoch": 27.607784431137723, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.132, "step": 18442 }, { "epoch": 27.60928143712575, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1252, "step": 18443 }, { "epoch": 27.610778443113773, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1315, "step": 18444 }, { "epoch": 27.612275449101798, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1313, "step": 18445 }, { "epoch": 27.61377245508982, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18446 }, { "epoch": 27.615269461077844, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1211, "step": 18447 }, { "epoch": 27.61676646706587, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.132, "step": 18448 }, { "epoch": 27.618263473053894, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1252, "step": 18449 }, { "epoch": 27.619760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 18450 }, { "epoch": 27.62125748502994, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1339, "step": 18451 }, { "epoch": 27.622754491017965, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1226, "step": 18452 }, { "epoch": 27.62425149700599, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1258, "step": 18453 }, { "epoch": 27.62574850299401, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1263, "step": 18454 }, { "epoch": 27.627245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1345, "step": 18455 }, { "epoch": 27.62874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1315, "step": 18456 }, { "epoch": 27.630239520958085, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1307, "step": 18457 }, { "epoch": 27.631736526946106, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1285, "step": 18458 }, { "epoch": 27.63323353293413, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1251, "step": 18459 }, { "epoch": 27.634730538922156, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1373, "step": 18460 }, { "epoch": 27.63622754491018, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1323, "step": 18461 }, { "epoch": 27.637724550898202, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1307, "step": 18462 }, { "epoch": 27.639221556886227, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1303, "step": 18463 }, { "epoch": 27.64071856287425, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 18464 }, { "epoch": 27.642215568862277, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1322, "step": 18465 }, { "epoch": 27.643712574850298, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1345, "step": 18466 }, { "epoch": 27.645209580838323, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1273, "step": 18467 }, { "epoch": 27.646706586826348, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1274, "step": 18468 }, { "epoch": 27.648203592814372, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1286, "step": 18469 }, { "epoch": 27.649700598802394, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.123, "step": 18470 }, { "epoch": 27.65119760479042, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1325, "step": 18471 }, { "epoch": 27.652694610778443, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1353, "step": 18472 }, { "epoch": 27.654191616766468, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1298, "step": 18473 }, { "epoch": 27.65568862275449, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.125, "step": 18474 }, { "epoch": 27.657185628742514, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1362, "step": 18475 }, { "epoch": 27.65868263473054, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1281, "step": 18476 }, { "epoch": 27.660179640718564, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18477 }, { "epoch": 27.66167664670659, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1262, "step": 18478 }, { "epoch": 27.66317365269461, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 18479 }, { "epoch": 27.664670658682635, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 18480 }, { "epoch": 27.66616766467066, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1283, "step": 18481 }, { "epoch": 27.66766467065868, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1321, "step": 18482 }, { "epoch": 27.669161676646706, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1276, "step": 18483 }, { "epoch": 27.67065868263473, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1378, "step": 18484 }, { "epoch": 27.672155688622755, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1366, "step": 18485 }, { "epoch": 27.67365269461078, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1271, "step": 18486 }, { "epoch": 27.6751497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1317, "step": 18487 }, { "epoch": 27.676646706586826, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1288, "step": 18488 }, { "epoch": 27.67814371257485, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1245, "step": 18489 }, { "epoch": 27.679640718562876, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1297, "step": 18490 }, { "epoch": 27.681137724550897, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1308, "step": 18491 }, { "epoch": 27.682634730538922, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1255, "step": 18492 }, { "epoch": 27.684131736526947, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1221, "step": 18493 }, { "epoch": 27.68562874251497, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1277, "step": 18494 }, { "epoch": 27.687125748502993, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1346, "step": 18495 }, { "epoch": 27.688622754491018, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.129, "step": 18496 }, { "epoch": 27.690119760479043, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1293, "step": 18497 }, { "epoch": 27.691616766467067, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 18498 }, { "epoch": 27.69311377245509, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1311, "step": 18499 }, { "epoch": 27.694610778443113, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.133, "step": 18500 }, { "epoch": 27.69610778443114, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1356, "step": 18501 }, { "epoch": 27.697604790419163, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.131, "step": 18502 }, { "epoch": 27.699101796407184, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 18503 }, { "epoch": 27.70059880239521, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 18504 }, { "epoch": 27.702095808383234, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1277, "step": 18505 }, { "epoch": 27.70359281437126, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1262, "step": 18506 }, { "epoch": 27.70508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1265, "step": 18507 }, { "epoch": 27.706586826347305, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1272, "step": 18508 }, { "epoch": 27.70808383233533, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 18509 }, { "epoch": 27.709580838323355, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1316, "step": 18510 }, { "epoch": 27.711077844311376, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18511 }, { "epoch": 27.7125748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 18512 }, { "epoch": 27.714071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1309, "step": 18513 }, { "epoch": 27.71556886227545, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 18514 }, { "epoch": 27.71706586826347, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1266, "step": 18515 }, { "epoch": 27.718562874251496, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1303, "step": 18516 }, { "epoch": 27.72005988023952, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1304, "step": 18517 }, { "epoch": 27.721556886227546, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1329, "step": 18518 }, { "epoch": 27.723053892215567, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1222, "step": 18519 }, { "epoch": 27.724550898203592, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.132, "step": 18520 }, { "epoch": 27.726047904191617, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 18521 }, { "epoch": 27.727544910179642, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1316, "step": 18522 }, { "epoch": 27.729041916167663, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 18523 }, { "epoch": 27.730538922155688, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1252, "step": 18524 }, { "epoch": 27.732035928143713, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1302, "step": 18525 }, { "epoch": 27.733532934131738, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1301, "step": 18526 }, { "epoch": 27.73502994011976, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1228, "step": 18527 }, { "epoch": 27.736526946107784, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 18528 }, { "epoch": 27.73802395209581, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 18529 }, { "epoch": 27.739520958083833, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1231, "step": 18530 }, { "epoch": 27.741017964071855, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1242, "step": 18531 }, { "epoch": 27.74251497005988, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1314, "step": 18532 }, { "epoch": 27.744011976047904, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1288, "step": 18533 }, { "epoch": 27.74550898203593, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1294, "step": 18534 }, { "epoch": 27.74700598802395, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1274, "step": 18535 }, { "epoch": 27.748502994011975, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1309, "step": 18536 }, { "epoch": 27.75, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1321, "step": 18537 }, { "epoch": 27.751497005988025, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1288, "step": 18538 }, { "epoch": 27.75299401197605, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.125, "step": 18539 }, { "epoch": 27.75449101796407, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1278, "step": 18540 }, { "epoch": 27.755988023952096, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1248, "step": 18541 }, { "epoch": 27.75748502994012, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1333, "step": 18542 }, { "epoch": 27.758982035928145, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1265, "step": 18543 }, { "epoch": 27.760479041916167, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1328, "step": 18544 }, { "epoch": 27.76197604790419, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1271, "step": 18545 }, { "epoch": 27.763473053892216, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 18546 }, { "epoch": 27.76497005988024, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1257, "step": 18547 }, { "epoch": 27.766467065868262, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1282, "step": 18548 }, { "epoch": 27.767964071856287, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1236, "step": 18549 }, { "epoch": 27.769461077844312, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1262, "step": 18550 }, { "epoch": 27.770958083832337, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1307, "step": 18551 }, { "epoch": 27.772455089820358, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1283, "step": 18552 }, { "epoch": 27.773952095808383, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 18553 }, { "epoch": 27.775449101796408, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1294, "step": 18554 }, { "epoch": 27.776946107784433, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1229, "step": 18555 }, { "epoch": 27.778443113772454, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1274, "step": 18556 }, { "epoch": 27.77994011976048, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18557 }, { "epoch": 27.781437125748504, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1245, "step": 18558 }, { "epoch": 27.78293413173653, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1306, "step": 18559 }, { "epoch": 27.78443113772455, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 18560 }, { "epoch": 27.785928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1257, "step": 18561 }, { "epoch": 27.7874251497006, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1258, "step": 18562 }, { "epoch": 27.788922155688624, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1275, "step": 18563 }, { "epoch": 27.790419161676645, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1285, "step": 18564 }, { "epoch": 27.79191616766467, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1358, "step": 18565 }, { "epoch": 27.793413173652695, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1273, "step": 18566 }, { "epoch": 27.79491017964072, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1337, "step": 18567 }, { "epoch": 27.79640718562874, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1339, "step": 18568 }, { "epoch": 27.797904191616766, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 18569 }, { "epoch": 27.79940119760479, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1283, "step": 18570 }, { "epoch": 27.800898203592816, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1285, "step": 18571 }, { "epoch": 27.802395209580837, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.14, "step": 18572 }, { "epoch": 27.80389221556886, "grad_norm": 0.050537109375, "learning_rate": 0.0008, "loss": 1.1277, "step": 18573 }, { "epoch": 27.805389221556887, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1319, "step": 18574 }, { "epoch": 27.80688622754491, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1351, "step": 18575 }, { "epoch": 27.808383233532933, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1266, "step": 18576 }, { "epoch": 27.809880239520957, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.127, "step": 18577 }, { "epoch": 27.811377245508982, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1289, "step": 18578 }, { "epoch": 27.812874251497007, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1252, "step": 18579 }, { "epoch": 27.81437125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1323, "step": 18580 }, { "epoch": 27.815868263473053, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1247, "step": 18581 }, { "epoch": 27.817365269461078, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18582 }, { "epoch": 27.818862275449103, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1343, "step": 18583 }, { "epoch": 27.820359281437124, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1274, "step": 18584 }, { "epoch": 27.82185628742515, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.125, "step": 18585 }, { "epoch": 27.823353293413174, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1323, "step": 18586 }, { "epoch": 27.8248502994012, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1316, "step": 18587 }, { "epoch": 27.82634730538922, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18588 }, { "epoch": 27.827844311377245, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1337, "step": 18589 }, { "epoch": 27.82934131736527, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1221, "step": 18590 }, { "epoch": 27.830838323353294, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1254, "step": 18591 }, { "epoch": 27.83233532934132, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1284, "step": 18592 }, { "epoch": 27.83383233532934, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18593 }, { "epoch": 27.835329341317365, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1319, "step": 18594 }, { "epoch": 27.83682634730539, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1292, "step": 18595 }, { "epoch": 27.83832335329341, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1305, "step": 18596 }, { "epoch": 27.839820359281436, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1263, "step": 18597 }, { "epoch": 27.84131736526946, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1301, "step": 18598 }, { "epoch": 27.842814371257486, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1308, "step": 18599 }, { "epoch": 27.84431137724551, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1342, "step": 18600 }, { "epoch": 27.845808383233532, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1293, "step": 18601 }, { "epoch": 27.847305389221557, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1276, "step": 18602 }, { "epoch": 27.84880239520958, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.124, "step": 18603 }, { "epoch": 27.850299401197606, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1263, "step": 18604 }, { "epoch": 27.851796407185628, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1282, "step": 18605 }, { "epoch": 27.853293413173652, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 18606 }, { "epoch": 27.854790419161677, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18607 }, { "epoch": 27.856287425149702, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.131, "step": 18608 }, { "epoch": 27.857784431137723, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1324, "step": 18609 }, { "epoch": 27.85928143712575, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1214, "step": 18610 }, { "epoch": 27.860778443113773, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1306, "step": 18611 }, { "epoch": 27.862275449101798, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1255, "step": 18612 }, { "epoch": 27.86377245508982, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1298, "step": 18613 }, { "epoch": 27.865269461077844, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 18614 }, { "epoch": 27.86676646706587, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1291, "step": 18615 }, { "epoch": 27.868263473053894, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 18616 }, { "epoch": 27.869760479041915, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1333, "step": 18617 }, { "epoch": 27.87125748502994, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 18618 }, { "epoch": 27.872754491017965, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1278, "step": 18619 }, { "epoch": 27.87425149700599, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1326, "step": 18620 }, { "epoch": 27.87574850299401, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1272, "step": 18621 }, { "epoch": 27.877245508982035, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1306, "step": 18622 }, { "epoch": 27.87874251497006, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1288, "step": 18623 }, { "epoch": 27.880239520958085, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1268, "step": 18624 }, { "epoch": 27.881736526946106, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1239, "step": 18625 }, { "epoch": 27.88323353293413, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.125, "step": 18626 }, { "epoch": 27.884730538922156, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1237, "step": 18627 }, { "epoch": 27.88622754491018, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1325, "step": 18628 }, { "epoch": 27.887724550898202, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1313, "step": 18629 }, { "epoch": 27.889221556886227, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1302, "step": 18630 }, { "epoch": 27.89071856287425, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1232, "step": 18631 }, { "epoch": 27.892215568862277, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1199, "step": 18632 }, { "epoch": 27.893712574850298, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1313, "step": 18633 }, { "epoch": 27.895209580838323, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1343, "step": 18634 }, { "epoch": 27.896706586826348, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1289, "step": 18635 }, { "epoch": 27.898203592814372, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1263, "step": 18636 }, { "epoch": 27.899700598802394, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1285, "step": 18637 }, { "epoch": 27.90119760479042, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1335, "step": 18638 }, { "epoch": 27.902694610778443, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.123, "step": 18639 }, { "epoch": 27.904191616766468, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1328, "step": 18640 }, { "epoch": 27.90568862275449, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1238, "step": 18641 }, { "epoch": 27.907185628742514, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1321, "step": 18642 }, { "epoch": 27.90868263473054, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18643 }, { "epoch": 27.910179640718564, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1277, "step": 18644 }, { "epoch": 27.91167664670659, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.127, "step": 18645 }, { "epoch": 27.91317365269461, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 18646 }, { "epoch": 27.914670658682635, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1342, "step": 18647 }, { "epoch": 27.91616766467066, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1335, "step": 18648 }, { "epoch": 27.91766467065868, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 18649 }, { "epoch": 27.919161676646706, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1268, "step": 18650 }, { "epoch": 27.92065868263473, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 18651 }, { "epoch": 27.922155688622755, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1289, "step": 18652 }, { "epoch": 27.92365269461078, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18653 }, { "epoch": 27.9251497005988, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1266, "step": 18654 }, { "epoch": 27.926646706586826, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1305, "step": 18655 }, { "epoch": 27.92814371257485, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1274, "step": 18656 }, { "epoch": 27.929640718562876, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1192, "step": 18657 }, { "epoch": 27.931137724550897, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1272, "step": 18658 }, { "epoch": 27.932634730538922, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1365, "step": 18659 }, { "epoch": 27.934131736526947, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1344, "step": 18660 }, { "epoch": 27.93562874251497, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1273, "step": 18661 }, { "epoch": 27.937125748502993, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 18662 }, { "epoch": 27.938622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1235, "step": 18663 }, { "epoch": 27.940119760479043, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1279, "step": 18664 }, { "epoch": 27.941616766467067, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.13, "step": 18665 }, { "epoch": 27.94311377245509, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1272, "step": 18666 }, { "epoch": 27.944610778443113, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1348, "step": 18667 }, { "epoch": 27.94610778443114, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18668 }, { "epoch": 27.947604790419163, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1301, "step": 18669 }, { "epoch": 27.949101796407184, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1246, "step": 18670 }, { "epoch": 27.95059880239521, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18671 }, { "epoch": 27.952095808383234, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1219, "step": 18672 }, { "epoch": 27.95359281437126, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1329, "step": 18673 }, { "epoch": 27.95508982035928, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1349, "step": 18674 }, { "epoch": 27.956586826347305, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.133, "step": 18675 }, { "epoch": 27.95808383233533, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.133, "step": 18676 }, { "epoch": 27.959580838323355, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1347, "step": 18677 }, { "epoch": 27.961077844311376, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 18678 }, { "epoch": 27.9625748502994, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1265, "step": 18679 }, { "epoch": 27.964071856287426, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1331, "step": 18680 }, { "epoch": 27.96556886227545, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1354, "step": 18681 }, { "epoch": 27.96706586826347, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1251, "step": 18682 }, { "epoch": 27.968562874251496, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1251, "step": 18683 }, { "epoch": 27.97005988023952, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1327, "step": 18684 }, { "epoch": 27.971556886227546, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1237, "step": 18685 }, { "epoch": 27.973053892215567, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1319, "step": 18686 }, { "epoch": 27.974550898203592, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.128, "step": 18687 }, { "epoch": 27.976047904191617, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1269, "step": 18688 }, { "epoch": 27.977544910179642, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.13, "step": 18689 }, { "epoch": 27.979041916167663, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1322, "step": 18690 }, { "epoch": 27.980538922155688, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1303, "step": 18691 }, { "epoch": 27.982035928143713, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1279, "step": 18692 }, { "epoch": 27.983532934131738, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1323, "step": 18693 }, { "epoch": 27.98502994011976, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1314, "step": 18694 }, { "epoch": 27.986526946107784, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.13, "step": 18695 }, { "epoch": 27.98802395209581, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1261, "step": 18696 }, { "epoch": 27.989520958083833, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1296, "step": 18697 }, { "epoch": 27.991017964071855, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1282, "step": 18698 }, { "epoch": 27.99251497005988, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.128, "step": 18699 }, { "epoch": 27.994011976047904, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1204, "step": 18700 }, { "epoch": 27.99550898203593, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1231, "step": 18701 }, { "epoch": 27.99700598802395, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1282, "step": 18702 }, { "epoch": 27.998502994011975, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1304, "step": 18703 }, { "epoch": 28.0, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1279, "step": 18704 }, { "epoch": 28.001497005988025, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1232, "step": 18705 }, { "epoch": 28.00299401197605, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1323, "step": 18706 }, { "epoch": 28.00449101796407, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 18707 }, { "epoch": 28.005988023952096, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1273, "step": 18708 }, { "epoch": 28.00748502994012, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1315, "step": 18709 }, { "epoch": 28.008982035928145, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1331, "step": 18710 }, { "epoch": 28.010479041916167, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1206, "step": 18711 }, { "epoch": 28.01197604790419, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 18712 }, { "epoch": 28.013473053892216, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1221, "step": 18713 }, { "epoch": 28.01497005988024, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1282, "step": 18714 }, { "epoch": 28.016467065868262, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18715 }, { "epoch": 28.017964071856287, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1291, "step": 18716 }, { "epoch": 28.019461077844312, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1277, "step": 18717 }, { "epoch": 28.020958083832337, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1318, "step": 18718 }, { "epoch": 28.022455089820358, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1326, "step": 18719 }, { "epoch": 28.023952095808383, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1267, "step": 18720 }, { "epoch": 28.025449101796408, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1298, "step": 18721 }, { "epoch": 28.026946107784433, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1218, "step": 18722 }, { "epoch": 28.028443113772454, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1254, "step": 18723 }, { "epoch": 28.02994011976048, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1258, "step": 18724 }, { "epoch": 28.031437125748504, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.127, "step": 18725 }, { "epoch": 28.03293413173653, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 18726 }, { "epoch": 28.03443113772455, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1261, "step": 18727 }, { "epoch": 28.035928143712574, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1338, "step": 18728 }, { "epoch": 28.0374251497006, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1299, "step": 18729 }, { "epoch": 28.038922155688624, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 18730 }, { "epoch": 28.040419161676645, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1199, "step": 18731 }, { "epoch": 28.04191616766467, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1233, "step": 18732 }, { "epoch": 28.043413173652695, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1278, "step": 18733 }, { "epoch": 28.04491017964072, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 18734 }, { "epoch": 28.04640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1338, "step": 18735 }, { "epoch": 28.047904191616766, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1233, "step": 18736 }, { "epoch": 28.04940119760479, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 18737 }, { "epoch": 28.050898203592816, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1309, "step": 18738 }, { "epoch": 28.052395209580837, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1214, "step": 18739 }, { "epoch": 28.05389221556886, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18740 }, { "epoch": 28.055389221556887, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1342, "step": 18741 }, { "epoch": 28.05688622754491, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1247, "step": 18742 }, { "epoch": 28.058383233532933, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.126, "step": 18743 }, { "epoch": 28.059880239520957, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1217, "step": 18744 }, { "epoch": 28.061377245508982, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1302, "step": 18745 }, { "epoch": 28.062874251497007, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1286, "step": 18746 }, { "epoch": 28.06437125748503, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1309, "step": 18747 }, { "epoch": 28.065868263473053, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1242, "step": 18748 }, { "epoch": 28.067365269461078, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1275, "step": 18749 }, { "epoch": 28.068862275449103, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.128, "step": 18750 }, { "epoch": 28.070359281437124, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1219, "step": 18751 }, { "epoch": 28.07185628742515, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1332, "step": 18752 }, { "epoch": 28.073353293413174, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1344, "step": 18753 }, { "epoch": 28.0748502994012, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1254, "step": 18754 }, { "epoch": 28.07634730538922, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 18755 }, { "epoch": 28.077844311377245, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 18756 }, { "epoch": 28.07934131736527, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1241, "step": 18757 }, { "epoch": 28.080838323353294, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1296, "step": 18758 }, { "epoch": 28.082335329341316, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1252, "step": 18759 }, { "epoch": 28.08383233532934, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1247, "step": 18760 }, { "epoch": 28.085329341317365, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.131, "step": 18761 }, { "epoch": 28.08682634730539, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1295, "step": 18762 }, { "epoch": 28.088323353293415, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1261, "step": 18763 }, { "epoch": 28.089820359281436, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1226, "step": 18764 }, { "epoch": 28.09131736526946, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1247, "step": 18765 }, { "epoch": 28.092814371257486, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.129, "step": 18766 }, { "epoch": 28.09431137724551, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1256, "step": 18767 }, { "epoch": 28.095808383233532, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18768 }, { "epoch": 28.097305389221557, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18769 }, { "epoch": 28.09880239520958, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1276, "step": 18770 }, { "epoch": 28.100299401197606, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1325, "step": 18771 }, { "epoch": 28.101796407185628, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1315, "step": 18772 }, { "epoch": 28.103293413173652, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1252, "step": 18773 }, { "epoch": 28.104790419161677, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1302, "step": 18774 }, { "epoch": 28.106287425149702, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1274, "step": 18775 }, { "epoch": 28.107784431137723, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1281, "step": 18776 }, { "epoch": 28.10928143712575, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1363, "step": 18777 }, { "epoch": 28.110778443113773, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 18778 }, { "epoch": 28.112275449101798, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1285, "step": 18779 }, { "epoch": 28.11377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.129, "step": 18780 }, { "epoch": 28.115269461077844, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1312, "step": 18781 }, { "epoch": 28.11676646706587, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1284, "step": 18782 }, { "epoch": 28.118263473053894, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1257, "step": 18783 }, { "epoch": 28.119760479041915, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1321, "step": 18784 }, { "epoch": 28.12125748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1288, "step": 18785 }, { "epoch": 28.122754491017965, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1318, "step": 18786 }, { "epoch": 28.12425149700599, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1298, "step": 18787 }, { "epoch": 28.12574850299401, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1277, "step": 18788 }, { "epoch": 28.127245508982035, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.129, "step": 18789 }, { "epoch": 28.12874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1262, "step": 18790 }, { "epoch": 28.130239520958085, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1332, "step": 18791 }, { "epoch": 28.131736526946106, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 18792 }, { "epoch": 28.13323353293413, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1237, "step": 18793 }, { "epoch": 28.134730538922156, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1268, "step": 18794 }, { "epoch": 28.13622754491018, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1279, "step": 18795 }, { "epoch": 28.137724550898202, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 18796 }, { "epoch": 28.139221556886227, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 18797 }, { "epoch": 28.14071856287425, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18798 }, { "epoch": 28.142215568862277, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.129, "step": 18799 }, { "epoch": 28.143712574850298, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1276, "step": 18800 }, { "epoch": 28.145209580838323, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.126, "step": 18801 }, { "epoch": 28.146706586826348, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18802 }, { "epoch": 28.148203592814372, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1283, "step": 18803 }, { "epoch": 28.149700598802394, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1273, "step": 18804 }, { "epoch": 28.15119760479042, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1278, "step": 18805 }, { "epoch": 28.152694610778443, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18806 }, { "epoch": 28.154191616766468, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1281, "step": 18807 }, { "epoch": 28.15568862275449, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1358, "step": 18808 }, { "epoch": 28.157185628742514, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1254, "step": 18809 }, { "epoch": 28.15868263473054, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1226, "step": 18810 }, { "epoch": 28.160179640718564, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1239, "step": 18811 }, { "epoch": 28.161676646706585, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1311, "step": 18812 }, { "epoch": 28.16317365269461, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.124, "step": 18813 }, { "epoch": 28.164670658682635, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1283, "step": 18814 }, { "epoch": 28.16616766467066, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1258, "step": 18815 }, { "epoch": 28.16766467065868, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1205, "step": 18816 }, { "epoch": 28.169161676646706, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1294, "step": 18817 }, { "epoch": 28.17065868263473, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1311, "step": 18818 }, { "epoch": 28.172155688622755, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1322, "step": 18819 }, { "epoch": 28.17365269461078, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 18820 }, { "epoch": 28.1751497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1234, "step": 18821 }, { "epoch": 28.176646706586826, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1275, "step": 18822 }, { "epoch": 28.17814371257485, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1262, "step": 18823 }, { "epoch": 28.179640718562876, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1298, "step": 18824 }, { "epoch": 28.181137724550897, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 18825 }, { "epoch": 28.182634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.127, "step": 18826 }, { "epoch": 28.184131736526947, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18827 }, { "epoch": 28.18562874251497, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.127, "step": 18828 }, { "epoch": 28.187125748502993, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1287, "step": 18829 }, { "epoch": 28.188622754491018, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1253, "step": 18830 }, { "epoch": 28.190119760479043, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.134, "step": 18831 }, { "epoch": 28.191616766467067, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1275, "step": 18832 }, { "epoch": 28.19311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1266, "step": 18833 }, { "epoch": 28.194610778443113, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.13, "step": 18834 }, { "epoch": 28.19610778443114, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1208, "step": 18835 }, { "epoch": 28.197604790419163, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1306, "step": 18836 }, { "epoch": 28.199101796407184, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1324, "step": 18837 }, { "epoch": 28.20059880239521, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1302, "step": 18838 }, { "epoch": 28.202095808383234, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 18839 }, { "epoch": 28.20359281437126, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1272, "step": 18840 }, { "epoch": 28.20508982035928, "grad_norm": 0.1689453125, "learning_rate": 0.0008, "loss": 1.1231, "step": 18841 }, { "epoch": 28.206586826347305, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.1236, "step": 18842 }, { "epoch": 28.20808383233533, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1297, "step": 18843 }, { "epoch": 28.209580838323355, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.1337, "step": 18844 }, { "epoch": 28.211077844311376, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 18845 }, { "epoch": 28.2125748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18846 }, { "epoch": 28.214071856287426, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1344, "step": 18847 }, { "epoch": 28.21556886227545, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1294, "step": 18848 }, { "epoch": 28.21706586826347, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1266, "step": 18849 }, { "epoch": 28.218562874251496, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1286, "step": 18850 }, { "epoch": 28.22005988023952, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1338, "step": 18851 }, { "epoch": 28.221556886227546, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1319, "step": 18852 }, { "epoch": 28.223053892215567, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1276, "step": 18853 }, { "epoch": 28.224550898203592, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.128, "step": 18854 }, { "epoch": 28.226047904191617, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1295, "step": 18855 }, { "epoch": 28.227544910179642, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 18856 }, { "epoch": 28.229041916167663, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1342, "step": 18857 }, { "epoch": 28.230538922155688, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1334, "step": 18858 }, { "epoch": 28.232035928143713, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1261, "step": 18859 }, { "epoch": 28.233532934131738, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 18860 }, { "epoch": 28.23502994011976, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18861 }, { "epoch": 28.236526946107784, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1333, "step": 18862 }, { "epoch": 28.23802395209581, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1341, "step": 18863 }, { "epoch": 28.239520958083833, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1307, "step": 18864 }, { "epoch": 28.241017964071855, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.129, "step": 18865 }, { "epoch": 28.24251497005988, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1241, "step": 18866 }, { "epoch": 28.244011976047904, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 18867 }, { "epoch": 28.24550898203593, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1245, "step": 18868 }, { "epoch": 28.24700598802395, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1203, "step": 18869 }, { "epoch": 28.248502994011975, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1193, "step": 18870 }, { "epoch": 28.25, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1323, "step": 18871 }, { "epoch": 28.251497005988025, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1299, "step": 18872 }, { "epoch": 28.25299401197605, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1313, "step": 18873 }, { "epoch": 28.25449101796407, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.129, "step": 18874 }, { "epoch": 28.255988023952096, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1266, "step": 18875 }, { "epoch": 28.25748502994012, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 18876 }, { "epoch": 28.258982035928145, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1395, "step": 18877 }, { "epoch": 28.260479041916167, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1279, "step": 18878 }, { "epoch": 28.26197604790419, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1342, "step": 18879 }, { "epoch": 28.263473053892216, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 18880 }, { "epoch": 28.26497005988024, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 18881 }, { "epoch": 28.266467065868262, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1279, "step": 18882 }, { "epoch": 28.267964071856287, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1329, "step": 18883 }, { "epoch": 28.269461077844312, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1328, "step": 18884 }, { "epoch": 28.270958083832337, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1332, "step": 18885 }, { "epoch": 28.272455089820358, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1317, "step": 18886 }, { "epoch": 28.273952095808383, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1336, "step": 18887 }, { "epoch": 28.275449101796408, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.129, "step": 18888 }, { "epoch": 28.276946107784433, "grad_norm": 0.0498046875, "learning_rate": 0.0008, "loss": 1.1249, "step": 18889 }, { "epoch": 28.278443113772454, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1302, "step": 18890 }, { "epoch": 28.27994011976048, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18891 }, { "epoch": 28.281437125748504, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1341, "step": 18892 }, { "epoch": 28.28293413173653, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1317, "step": 18893 }, { "epoch": 28.28443113772455, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1343, "step": 18894 }, { "epoch": 28.285928143712574, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1299, "step": 18895 }, { "epoch": 28.2874251497006, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1304, "step": 18896 }, { "epoch": 28.288922155688624, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.126, "step": 18897 }, { "epoch": 28.290419161676645, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1291, "step": 18898 }, { "epoch": 28.29191616766467, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1324, "step": 18899 }, { "epoch": 28.293413173652695, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1291, "step": 18900 }, { "epoch": 28.29491017964072, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1231, "step": 18901 }, { "epoch": 28.29640718562874, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1303, "step": 18902 }, { "epoch": 28.297904191616766, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1324, "step": 18903 }, { "epoch": 28.29940119760479, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1242, "step": 18904 }, { "epoch": 28.300898203592816, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1317, "step": 18905 }, { "epoch": 28.302395209580837, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1261, "step": 18906 }, { "epoch": 28.30389221556886, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1316, "step": 18907 }, { "epoch": 28.305389221556887, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1229, "step": 18908 }, { "epoch": 28.30688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1271, "step": 18909 }, { "epoch": 28.308383233532933, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1209, "step": 18910 }, { "epoch": 28.309880239520957, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.123, "step": 18911 }, { "epoch": 28.311377245508982, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1282, "step": 18912 }, { "epoch": 28.312874251497007, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 18913 }, { "epoch": 28.31437125748503, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 18914 }, { "epoch": 28.315868263473053, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1241, "step": 18915 }, { "epoch": 28.317365269461078, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1247, "step": 18916 }, { "epoch": 28.318862275449103, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.126, "step": 18917 }, { "epoch": 28.320359281437124, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1232, "step": 18918 }, { "epoch": 28.32185628742515, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1311, "step": 18919 }, { "epoch": 28.323353293413174, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1275, "step": 18920 }, { "epoch": 28.3248502994012, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18921 }, { "epoch": 28.32634730538922, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1341, "step": 18922 }, { "epoch": 28.327844311377245, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1329, "step": 18923 }, { "epoch": 28.32934131736527, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1331, "step": 18924 }, { "epoch": 28.330838323353294, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1241, "step": 18925 }, { "epoch": 28.33233532934132, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1299, "step": 18926 }, { "epoch": 28.33383233532934, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.125, "step": 18927 }, { "epoch": 28.335329341317365, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1263, "step": 18928 }, { "epoch": 28.33682634730539, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.127, "step": 18929 }, { "epoch": 28.338323353293415, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1255, "step": 18930 }, { "epoch": 28.339820359281436, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1279, "step": 18931 }, { "epoch": 28.34131736526946, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1312, "step": 18932 }, { "epoch": 28.342814371257486, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1342, "step": 18933 }, { "epoch": 28.34431137724551, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1254, "step": 18934 }, { "epoch": 28.345808383233532, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1294, "step": 18935 }, { "epoch": 28.347305389221557, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1344, "step": 18936 }, { "epoch": 28.34880239520958, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1282, "step": 18937 }, { "epoch": 28.350299401197606, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1302, "step": 18938 }, { "epoch": 28.351796407185628, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1278, "step": 18939 }, { "epoch": 28.353293413173652, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.1286, "step": 18940 }, { "epoch": 28.354790419161677, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1297, "step": 18941 }, { "epoch": 28.356287425149702, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1357, "step": 18942 }, { "epoch": 28.357784431137723, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 18943 }, { "epoch": 28.35928143712575, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.134, "step": 18944 }, { "epoch": 28.360778443113773, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1246, "step": 18945 }, { "epoch": 28.362275449101798, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1218, "step": 18946 }, { "epoch": 28.36377245508982, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.127, "step": 18947 }, { "epoch": 28.365269461077844, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1272, "step": 18948 }, { "epoch": 28.36676646706587, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1306, "step": 18949 }, { "epoch": 28.368263473053894, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1282, "step": 18950 }, { "epoch": 28.369760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.13, "step": 18951 }, { "epoch": 28.37125748502994, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1264, "step": 18952 }, { "epoch": 28.372754491017965, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1218, "step": 18953 }, { "epoch": 28.37425149700599, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.136, "step": 18954 }, { "epoch": 28.37574850299401, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1313, "step": 18955 }, { "epoch": 28.377245508982035, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1272, "step": 18956 }, { "epoch": 28.37874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1316, "step": 18957 }, { "epoch": 28.380239520958085, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1266, "step": 18958 }, { "epoch": 28.381736526946106, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18959 }, { "epoch": 28.38323353293413, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1272, "step": 18960 }, { "epoch": 28.384730538922156, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1381, "step": 18961 }, { "epoch": 28.38622754491018, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1235, "step": 18962 }, { "epoch": 28.387724550898202, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1255, "step": 18963 }, { "epoch": 28.389221556886227, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1319, "step": 18964 }, { "epoch": 28.39071856287425, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1316, "step": 18965 }, { "epoch": 28.392215568862277, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1299, "step": 18966 }, { "epoch": 28.393712574850298, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1252, "step": 18967 }, { "epoch": 28.395209580838323, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1249, "step": 18968 }, { "epoch": 28.396706586826348, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1311, "step": 18969 }, { "epoch": 28.398203592814372, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1344, "step": 18970 }, { "epoch": 28.399700598802394, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1287, "step": 18971 }, { "epoch": 28.40119760479042, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1228, "step": 18972 }, { "epoch": 28.402694610778443, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1268, "step": 18973 }, { "epoch": 28.404191616766468, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1253, "step": 18974 }, { "epoch": 28.40568862275449, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1242, "step": 18975 }, { "epoch": 28.407185628742514, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1315, "step": 18976 }, { "epoch": 28.40868263473054, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1262, "step": 18977 }, { "epoch": 28.410179640718564, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1278, "step": 18978 }, { "epoch": 28.411676646706585, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1249, "step": 18979 }, { "epoch": 28.41317365269461, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1234, "step": 18980 }, { "epoch": 28.414670658682635, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18981 }, { "epoch": 28.41616766467066, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 18982 }, { "epoch": 28.41766467065868, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1338, "step": 18983 }, { "epoch": 28.419161676646706, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1291, "step": 18984 }, { "epoch": 28.42065868263473, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1356, "step": 18985 }, { "epoch": 28.422155688622755, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1292, "step": 18986 }, { "epoch": 28.42365269461078, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 18987 }, { "epoch": 28.4251497005988, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1265, "step": 18988 }, { "epoch": 28.426646706586826, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.122, "step": 18989 }, { "epoch": 28.42814371257485, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 18990 }, { "epoch": 28.429640718562876, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1325, "step": 18991 }, { "epoch": 28.431137724550897, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18992 }, { "epoch": 28.432634730538922, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1293, "step": 18993 }, { "epoch": 28.434131736526947, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.126, "step": 18994 }, { "epoch": 28.43562874251497, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1293, "step": 18995 }, { "epoch": 28.437125748502993, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1227, "step": 18996 }, { "epoch": 28.438622754491018, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1259, "step": 18997 }, { "epoch": 28.440119760479043, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1282, "step": 18998 }, { "epoch": 28.441616766467067, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1255, "step": 18999 }, { "epoch": 28.44311377245509, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1347, "step": 19000 }, { "epoch": 28.444610778443113, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 19001 }, { "epoch": 28.44610778443114, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19002 }, { "epoch": 28.447604790419163, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1282, "step": 19003 }, { "epoch": 28.449101796407184, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 19004 }, { "epoch": 28.45059880239521, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 19005 }, { "epoch": 28.452095808383234, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19006 }, { "epoch": 28.45359281437126, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1247, "step": 19007 }, { "epoch": 28.45508982035928, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1253, "step": 19008 }, { "epoch": 28.456586826347305, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1341, "step": 19009 }, { "epoch": 28.45808383233533, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1342, "step": 19010 }, { "epoch": 28.459580838323355, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1289, "step": 19011 }, { "epoch": 28.461077844311376, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.126, "step": 19012 }, { "epoch": 28.4625748502994, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1277, "step": 19013 }, { "epoch": 28.464071856287426, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1224, "step": 19014 }, { "epoch": 28.46556886227545, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 19015 }, { "epoch": 28.46706586826347, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1289, "step": 19016 }, { "epoch": 28.468562874251496, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1279, "step": 19017 }, { "epoch": 28.47005988023952, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1257, "step": 19018 }, { "epoch": 28.471556886227546, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1242, "step": 19019 }, { "epoch": 28.473053892215567, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1311, "step": 19020 }, { "epoch": 28.474550898203592, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 19021 }, { "epoch": 28.476047904191617, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1259, "step": 19022 }, { "epoch": 28.477544910179642, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19023 }, { "epoch": 28.479041916167663, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1273, "step": 19024 }, { "epoch": 28.480538922155688, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1282, "step": 19025 }, { "epoch": 28.482035928143713, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1284, "step": 19026 }, { "epoch": 28.483532934131738, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1148, "step": 19027 }, { "epoch": 28.48502994011976, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1329, "step": 19028 }, { "epoch": 28.486526946107784, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1257, "step": 19029 }, { "epoch": 28.48802395209581, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.127, "step": 19030 }, { "epoch": 28.489520958083833, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1315, "step": 19031 }, { "epoch": 28.491017964071855, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1311, "step": 19032 }, { "epoch": 28.49251497005988, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1312, "step": 19033 }, { "epoch": 28.494011976047904, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1233, "step": 19034 }, { "epoch": 28.49550898203593, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1336, "step": 19035 }, { "epoch": 28.49700598802395, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1381, "step": 19036 }, { "epoch": 28.498502994011975, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1256, "step": 19037 }, { "epoch": 28.5, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1285, "step": 19038 }, { "epoch": 28.501497005988025, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1319, "step": 19039 }, { "epoch": 28.50299401197605, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 19040 }, { "epoch": 28.50449101796407, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1269, "step": 19041 }, { "epoch": 28.505988023952096, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1242, "step": 19042 }, { "epoch": 28.50748502994012, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1335, "step": 19043 }, { "epoch": 28.508982035928145, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1319, "step": 19044 }, { "epoch": 28.510479041916167, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1248, "step": 19045 }, { "epoch": 28.51197604790419, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1202, "step": 19046 }, { "epoch": 28.513473053892216, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1326, "step": 19047 }, { "epoch": 28.51497005988024, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1268, "step": 19048 }, { "epoch": 28.516467065868262, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1287, "step": 19049 }, { "epoch": 28.517964071856287, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1348, "step": 19050 }, { "epoch": 28.519461077844312, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1324, "step": 19051 }, { "epoch": 28.520958083832337, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1311, "step": 19052 }, { "epoch": 28.522455089820358, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1247, "step": 19053 }, { "epoch": 28.523952095808383, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1251, "step": 19054 }, { "epoch": 28.525449101796408, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1273, "step": 19055 }, { "epoch": 28.526946107784433, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19056 }, { "epoch": 28.528443113772454, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1338, "step": 19057 }, { "epoch": 28.52994011976048, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.128, "step": 19058 }, { "epoch": 28.531437125748504, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19059 }, { "epoch": 28.53293413173653, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1307, "step": 19060 }, { "epoch": 28.53443113772455, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1303, "step": 19061 }, { "epoch": 28.535928143712574, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1279, "step": 19062 }, { "epoch": 28.5374251497006, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1243, "step": 19063 }, { "epoch": 28.538922155688624, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1287, "step": 19064 }, { "epoch": 28.540419161676645, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.128, "step": 19065 }, { "epoch": 28.54191616766467, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1275, "step": 19066 }, { "epoch": 28.543413173652695, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1324, "step": 19067 }, { "epoch": 28.54491017964072, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1299, "step": 19068 }, { "epoch": 28.54640718562874, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1315, "step": 19069 }, { "epoch": 28.547904191616766, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1348, "step": 19070 }, { "epoch": 28.54940119760479, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1272, "step": 19071 }, { "epoch": 28.550898203592816, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1322, "step": 19072 }, { "epoch": 28.552395209580837, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1267, "step": 19073 }, { "epoch": 28.55389221556886, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19074 }, { "epoch": 28.555389221556887, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1312, "step": 19075 }, { "epoch": 28.55688622754491, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1273, "step": 19076 }, { "epoch": 28.558383233532933, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 19077 }, { "epoch": 28.559880239520957, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19078 }, { "epoch": 28.561377245508982, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 19079 }, { "epoch": 28.562874251497007, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1266, "step": 19080 }, { "epoch": 28.56437125748503, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1271, "step": 19081 }, { "epoch": 28.565868263473053, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 19082 }, { "epoch": 28.567365269461078, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1319, "step": 19083 }, { "epoch": 28.568862275449103, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1265, "step": 19084 }, { "epoch": 28.570359281437124, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19085 }, { "epoch": 28.57185628742515, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1309, "step": 19086 }, { "epoch": 28.573353293413174, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1226, "step": 19087 }, { "epoch": 28.5748502994012, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 19088 }, { "epoch": 28.57634730538922, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1292, "step": 19089 }, { "epoch": 28.577844311377245, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1283, "step": 19090 }, { "epoch": 28.57934131736527, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.13, "step": 19091 }, { "epoch": 28.580838323353294, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1181, "step": 19092 }, { "epoch": 28.58233532934132, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1325, "step": 19093 }, { "epoch": 28.58383233532934, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1319, "step": 19094 }, { "epoch": 28.585329341317365, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 19095 }, { "epoch": 28.58682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1259, "step": 19096 }, { "epoch": 28.58832335329341, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1261, "step": 19097 }, { "epoch": 28.589820359281436, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1246, "step": 19098 }, { "epoch": 28.59131736526946, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1325, "step": 19099 }, { "epoch": 28.592814371257486, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.129, "step": 19100 }, { "epoch": 28.59431137724551, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.133, "step": 19101 }, { "epoch": 28.595808383233532, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1265, "step": 19102 }, { "epoch": 28.597305389221557, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1261, "step": 19103 }, { "epoch": 28.59880239520958, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1318, "step": 19104 }, { "epoch": 28.600299401197606, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.132, "step": 19105 }, { "epoch": 28.601796407185628, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1326, "step": 19106 }, { "epoch": 28.603293413173652, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1291, "step": 19107 }, { "epoch": 28.604790419161677, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1245, "step": 19108 }, { "epoch": 28.606287425149702, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1215, "step": 19109 }, { "epoch": 28.607784431137723, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1306, "step": 19110 }, { "epoch": 28.60928143712575, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1274, "step": 19111 }, { "epoch": 28.610778443113773, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1276, "step": 19112 }, { "epoch": 28.612275449101798, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1251, "step": 19113 }, { "epoch": 28.61377245508982, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1298, "step": 19114 }, { "epoch": 28.615269461077844, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1325, "step": 19115 }, { "epoch": 28.61676646706587, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1262, "step": 19116 }, { "epoch": 28.618263473053894, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1294, "step": 19117 }, { "epoch": 28.619760479041915, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1265, "step": 19118 }, { "epoch": 28.62125748502994, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1231, "step": 19119 }, { "epoch": 28.622754491017965, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1271, "step": 19120 }, { "epoch": 28.62425149700599, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1284, "step": 19121 }, { "epoch": 28.62574850299401, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1325, "step": 19122 }, { "epoch": 28.627245508982035, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.134, "step": 19123 }, { "epoch": 28.62874251497006, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1235, "step": 19124 }, { "epoch": 28.630239520958085, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1362, "step": 19125 }, { "epoch": 28.631736526946106, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1268, "step": 19126 }, { "epoch": 28.63323353293413, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1332, "step": 19127 }, { "epoch": 28.634730538922156, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1262, "step": 19128 }, { "epoch": 28.63622754491018, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1274, "step": 19129 }, { "epoch": 28.637724550898202, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.13, "step": 19130 }, { "epoch": 28.639221556886227, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1291, "step": 19131 }, { "epoch": 28.64071856287425, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1234, "step": 19132 }, { "epoch": 28.642215568862277, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1251, "step": 19133 }, { "epoch": 28.643712574850298, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1302, "step": 19134 }, { "epoch": 28.645209580838323, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1266, "step": 19135 }, { "epoch": 28.646706586826348, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1287, "step": 19136 }, { "epoch": 28.648203592814372, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1343, "step": 19137 }, { "epoch": 28.649700598802394, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1244, "step": 19138 }, { "epoch": 28.65119760479042, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1208, "step": 19139 }, { "epoch": 28.652694610778443, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1259, "step": 19140 }, { "epoch": 28.654191616766468, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1247, "step": 19141 }, { "epoch": 28.65568862275449, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.128, "step": 19142 }, { "epoch": 28.657185628742514, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 19143 }, { "epoch": 28.65868263473054, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1319, "step": 19144 }, { "epoch": 28.660179640718564, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1346, "step": 19145 }, { "epoch": 28.66167664670659, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1307, "step": 19146 }, { "epoch": 28.66317365269461, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.126, "step": 19147 }, { "epoch": 28.664670658682635, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 19148 }, { "epoch": 28.66616766467066, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1296, "step": 19149 }, { "epoch": 28.66766467065868, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.129, "step": 19150 }, { "epoch": 28.669161676646706, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1205, "step": 19151 }, { "epoch": 28.67065868263473, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1235, "step": 19152 }, { "epoch": 28.672155688622755, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1245, "step": 19153 }, { "epoch": 28.67365269461078, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1248, "step": 19154 }, { "epoch": 28.6751497005988, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1242, "step": 19155 }, { "epoch": 28.676646706586826, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1283, "step": 19156 }, { "epoch": 28.67814371257485, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 19157 }, { "epoch": 28.679640718562876, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1248, "step": 19158 }, { "epoch": 28.681137724550897, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1333, "step": 19159 }, { "epoch": 28.682634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1288, "step": 19160 }, { "epoch": 28.684131736526947, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1317, "step": 19161 }, { "epoch": 28.68562874251497, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1318, "step": 19162 }, { "epoch": 28.687125748502993, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1252, "step": 19163 }, { "epoch": 28.688622754491018, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1235, "step": 19164 }, { "epoch": 28.690119760479043, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1276, "step": 19165 }, { "epoch": 28.691616766467067, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1286, "step": 19166 }, { "epoch": 28.69311377245509, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19167 }, { "epoch": 28.694610778443113, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1285, "step": 19168 }, { "epoch": 28.69610778443114, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1291, "step": 19169 }, { "epoch": 28.697604790419163, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1376, "step": 19170 }, { "epoch": 28.699101796407184, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1295, "step": 19171 }, { "epoch": 28.70059880239521, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1332, "step": 19172 }, { "epoch": 28.702095808383234, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1171, "step": 19173 }, { "epoch": 28.70359281437126, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1259, "step": 19174 }, { "epoch": 28.70508982035928, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1232, "step": 19175 }, { "epoch": 28.706586826347305, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1276, "step": 19176 }, { "epoch": 28.70808383233533, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.128, "step": 19177 }, { "epoch": 28.709580838323355, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1249, "step": 19178 }, { "epoch": 28.711077844311376, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1331, "step": 19179 }, { "epoch": 28.7125748502994, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1218, "step": 19180 }, { "epoch": 28.714071856287426, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1289, "step": 19181 }, { "epoch": 28.71556886227545, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19182 }, { "epoch": 28.71706586826347, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1339, "step": 19183 }, { "epoch": 28.718562874251496, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1243, "step": 19184 }, { "epoch": 28.72005988023952, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1258, "step": 19185 }, { "epoch": 28.721556886227546, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1217, "step": 19186 }, { "epoch": 28.723053892215567, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1315, "step": 19187 }, { "epoch": 28.724550898203592, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1229, "step": 19188 }, { "epoch": 28.726047904191617, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1295, "step": 19189 }, { "epoch": 28.727544910179642, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1219, "step": 19190 }, { "epoch": 28.729041916167663, "grad_norm": 0.047607421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19191 }, { "epoch": 28.730538922155688, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1299, "step": 19192 }, { "epoch": 28.732035928143713, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1309, "step": 19193 }, { "epoch": 28.733532934131738, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.131, "step": 19194 }, { "epoch": 28.73502994011976, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1289, "step": 19195 }, { "epoch": 28.736526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1352, "step": 19196 }, { "epoch": 28.73802395209581, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1278, "step": 19197 }, { "epoch": 28.739520958083833, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1247, "step": 19198 }, { "epoch": 28.741017964071855, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1344, "step": 19199 }, { "epoch": 28.74251497005988, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1279, "step": 19200 }, { "epoch": 28.744011976047904, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 19201 }, { "epoch": 28.74550898203593, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1318, "step": 19202 }, { "epoch": 28.74700598802395, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1258, "step": 19203 }, { "epoch": 28.748502994011975, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1308, "step": 19204 }, { "epoch": 28.75, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1375, "step": 19205 }, { "epoch": 28.751497005988025, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19206 }, { "epoch": 28.75299401197605, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1303, "step": 19207 }, { "epoch": 28.75449101796407, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.127, "step": 19208 }, { "epoch": 28.755988023952096, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1186, "step": 19209 }, { "epoch": 28.75748502994012, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1242, "step": 19210 }, { "epoch": 28.758982035928145, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19211 }, { "epoch": 28.760479041916167, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1362, "step": 19212 }, { "epoch": 28.76197604790419, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1288, "step": 19213 }, { "epoch": 28.763473053892216, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1269, "step": 19214 }, { "epoch": 28.76497005988024, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1367, "step": 19215 }, { "epoch": 28.766467065868262, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.13, "step": 19216 }, { "epoch": 28.767964071856287, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1323, "step": 19217 }, { "epoch": 28.769461077844312, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1276, "step": 19218 }, { "epoch": 28.770958083832337, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19219 }, { "epoch": 28.772455089820358, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1222, "step": 19220 }, { "epoch": 28.773952095808383, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1301, "step": 19221 }, { "epoch": 28.775449101796408, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.126, "step": 19222 }, { "epoch": 28.776946107784433, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1276, "step": 19223 }, { "epoch": 28.778443113772454, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1343, "step": 19224 }, { "epoch": 28.77994011976048, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.131, "step": 19225 }, { "epoch": 28.781437125748504, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1308, "step": 19226 }, { "epoch": 28.78293413173653, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1349, "step": 19227 }, { "epoch": 28.78443113772455, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1314, "step": 19228 }, { "epoch": 28.785928143712574, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.127, "step": 19229 }, { "epoch": 28.7874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1277, "step": 19230 }, { "epoch": 28.788922155688624, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1313, "step": 19231 }, { "epoch": 28.790419161676645, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.128, "step": 19232 }, { "epoch": 28.79191616766467, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1332, "step": 19233 }, { "epoch": 28.793413173652695, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1284, "step": 19234 }, { "epoch": 28.79491017964072, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1241, "step": 19235 }, { "epoch": 28.79640718562874, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 19236 }, { "epoch": 28.797904191616766, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1305, "step": 19237 }, { "epoch": 28.79940119760479, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1266, "step": 19238 }, { "epoch": 28.800898203592816, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 19239 }, { "epoch": 28.802395209580837, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.134, "step": 19240 }, { "epoch": 28.80389221556886, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1313, "step": 19241 }, { "epoch": 28.805389221556887, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1306, "step": 19242 }, { "epoch": 28.80688622754491, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.128, "step": 19243 }, { "epoch": 28.808383233532933, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1303, "step": 19244 }, { "epoch": 28.809880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1265, "step": 19245 }, { "epoch": 28.811377245508982, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.124, "step": 19246 }, { "epoch": 28.812874251497007, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19247 }, { "epoch": 28.81437125748503, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1383, "step": 19248 }, { "epoch": 28.815868263473053, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1251, "step": 19249 }, { "epoch": 28.817365269461078, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1222, "step": 19250 }, { "epoch": 28.818862275449103, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1258, "step": 19251 }, { "epoch": 28.820359281437124, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1359, "step": 19252 }, { "epoch": 28.82185628742515, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1335, "step": 19253 }, { "epoch": 28.823353293413174, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1321, "step": 19254 }, { "epoch": 28.8248502994012, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1297, "step": 19255 }, { "epoch": 28.82634730538922, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 19256 }, { "epoch": 28.827844311377245, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1319, "step": 19257 }, { "epoch": 28.82934131736527, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 19258 }, { "epoch": 28.830838323353294, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1267, "step": 19259 }, { "epoch": 28.83233532934132, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1302, "step": 19260 }, { "epoch": 28.83383233532934, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1211, "step": 19261 }, { "epoch": 28.835329341317365, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1238, "step": 19262 }, { "epoch": 28.83682634730539, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1257, "step": 19263 }, { "epoch": 28.83832335329341, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1337, "step": 19264 }, { "epoch": 28.839820359281436, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19265 }, { "epoch": 28.84131736526946, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1264, "step": 19266 }, { "epoch": 28.842814371257486, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1252, "step": 19267 }, { "epoch": 28.84431137724551, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1379, "step": 19268 }, { "epoch": 28.845808383233532, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1244, "step": 19269 }, { "epoch": 28.847305389221557, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1312, "step": 19270 }, { "epoch": 28.84880239520958, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1221, "step": 19271 }, { "epoch": 28.850299401197606, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1219, "step": 19272 }, { "epoch": 28.851796407185628, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1317, "step": 19273 }, { "epoch": 28.853293413173652, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1281, "step": 19274 }, { "epoch": 28.854790419161677, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 19275 }, { "epoch": 28.856287425149702, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1257, "step": 19276 }, { "epoch": 28.857784431137723, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1279, "step": 19277 }, { "epoch": 28.85928143712575, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 19278 }, { "epoch": 28.860778443113773, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.131, "step": 19279 }, { "epoch": 28.862275449101798, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1292, "step": 19280 }, { "epoch": 28.86377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1269, "step": 19281 }, { "epoch": 28.865269461077844, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1248, "step": 19282 }, { "epoch": 28.86676646706587, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1263, "step": 19283 }, { "epoch": 28.868263473053894, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19284 }, { "epoch": 28.869760479041915, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.131, "step": 19285 }, { "epoch": 28.87125748502994, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.13, "step": 19286 }, { "epoch": 28.872754491017965, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 19287 }, { "epoch": 28.87425149700599, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1314, "step": 19288 }, { "epoch": 28.87574850299401, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1293, "step": 19289 }, { "epoch": 28.877245508982035, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19290 }, { "epoch": 28.87874251497006, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.131, "step": 19291 }, { "epoch": 28.880239520958085, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1314, "step": 19292 }, { "epoch": 28.881736526946106, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1241, "step": 19293 }, { "epoch": 28.88323353293413, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1274, "step": 19294 }, { "epoch": 28.884730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1242, "step": 19295 }, { "epoch": 28.88622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 19296 }, { "epoch": 28.887724550898202, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1337, "step": 19297 }, { "epoch": 28.889221556886227, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1302, "step": 19298 }, { "epoch": 28.89071856287425, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1257, "step": 19299 }, { "epoch": 28.892215568862277, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 19300 }, { "epoch": 28.893712574850298, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1315, "step": 19301 }, { "epoch": 28.895209580838323, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 19302 }, { "epoch": 28.896706586826348, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1308, "step": 19303 }, { "epoch": 28.898203592814372, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1321, "step": 19304 }, { "epoch": 28.899700598802394, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.13, "step": 19305 }, { "epoch": 28.90119760479042, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1303, "step": 19306 }, { "epoch": 28.902694610778443, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1186, "step": 19307 }, { "epoch": 28.904191616766468, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1241, "step": 19308 }, { "epoch": 28.90568862275449, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1225, "step": 19309 }, { "epoch": 28.907185628742514, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.131, "step": 19310 }, { "epoch": 28.90868263473054, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1233, "step": 19311 }, { "epoch": 28.910179640718564, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1302, "step": 19312 }, { "epoch": 28.91167664670659, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.127, "step": 19313 }, { "epoch": 28.91317365269461, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 19314 }, { "epoch": 28.914670658682635, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1256, "step": 19315 }, { "epoch": 28.91616766467066, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1216, "step": 19316 }, { "epoch": 28.91766467065868, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.125, "step": 19317 }, { "epoch": 28.919161676646706, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.13, "step": 19318 }, { "epoch": 28.92065868263473, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1335, "step": 19319 }, { "epoch": 28.922155688622755, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1356, "step": 19320 }, { "epoch": 28.92365269461078, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 19321 }, { "epoch": 28.9251497005988, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1309, "step": 19322 }, { "epoch": 28.926646706586826, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1225, "step": 19323 }, { "epoch": 28.92814371257485, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.131, "step": 19324 }, { "epoch": 28.929640718562876, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1279, "step": 19325 }, { "epoch": 28.931137724550897, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1265, "step": 19326 }, { "epoch": 28.932634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1291, "step": 19327 }, { "epoch": 28.934131736526947, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1318, "step": 19328 }, { "epoch": 28.93562874251497, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1298, "step": 19329 }, { "epoch": 28.937125748502993, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1314, "step": 19330 }, { "epoch": 28.938622754491018, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1305, "step": 19331 }, { "epoch": 28.940119760479043, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 19332 }, { "epoch": 28.941616766467067, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1321, "step": 19333 }, { "epoch": 28.94311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1269, "step": 19334 }, { "epoch": 28.944610778443113, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1212, "step": 19335 }, { "epoch": 28.94610778443114, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1203, "step": 19336 }, { "epoch": 28.947604790419163, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 19337 }, { "epoch": 28.949101796407184, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1342, "step": 19338 }, { "epoch": 28.95059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1271, "step": 19339 }, { "epoch": 28.952095808383234, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1269, "step": 19340 }, { "epoch": 28.95359281437126, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1294, "step": 19341 }, { "epoch": 28.95508982035928, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 19342 }, { "epoch": 28.956586826347305, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19343 }, { "epoch": 28.95808383233533, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.129, "step": 19344 }, { "epoch": 28.959580838323355, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1285, "step": 19345 }, { "epoch": 28.961077844311376, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1272, "step": 19346 }, { "epoch": 28.9625748502994, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19347 }, { "epoch": 28.964071856287426, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1252, "step": 19348 }, { "epoch": 28.96556886227545, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1295, "step": 19349 }, { "epoch": 28.96706586826347, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1304, "step": 19350 }, { "epoch": 28.968562874251496, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1277, "step": 19351 }, { "epoch": 28.97005988023952, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1236, "step": 19352 }, { "epoch": 28.971556886227546, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19353 }, { "epoch": 28.973053892215567, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.128, "step": 19354 }, { "epoch": 28.974550898203592, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1285, "step": 19355 }, { "epoch": 28.976047904191617, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1301, "step": 19356 }, { "epoch": 28.977544910179642, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.127, "step": 19357 }, { "epoch": 28.979041916167663, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1313, "step": 19358 }, { "epoch": 28.980538922155688, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 19359 }, { "epoch": 28.982035928143713, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.135, "step": 19360 }, { "epoch": 28.983532934131738, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1277, "step": 19361 }, { "epoch": 28.98502994011976, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1316, "step": 19362 }, { "epoch": 28.986526946107784, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 19363 }, { "epoch": 28.98802395209581, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1299, "step": 19364 }, { "epoch": 28.989520958083833, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1323, "step": 19365 }, { "epoch": 28.991017964071855, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.129, "step": 19366 }, { "epoch": 28.99251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1285, "step": 19367 }, { "epoch": 28.994011976047904, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1327, "step": 19368 }, { "epoch": 28.99550898203593, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1288, "step": 19369 }, { "epoch": 28.99700598802395, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1301, "step": 19370 }, { "epoch": 28.998502994011975, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1278, "step": 19371 }, { "epoch": 29.0, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1191, "step": 19372 }, { "epoch": 29.001497005988025, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1194, "step": 19373 }, { "epoch": 29.00299401197605, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 19374 }, { "epoch": 29.00449101796407, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1301, "step": 19375 }, { "epoch": 29.005988023952096, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 19376 }, { "epoch": 29.00748502994012, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1247, "step": 19377 }, { "epoch": 29.008982035928145, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1278, "step": 19378 }, { "epoch": 29.010479041916167, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.129, "step": 19379 }, { "epoch": 29.01197604790419, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1357, "step": 19380 }, { "epoch": 29.013473053892216, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1163, "step": 19381 }, { "epoch": 29.01497005988024, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.122, "step": 19382 }, { "epoch": 29.016467065868262, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1323, "step": 19383 }, { "epoch": 29.017964071856287, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1263, "step": 19384 }, { "epoch": 29.019461077844312, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1298, "step": 19385 }, { "epoch": 29.020958083832337, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19386 }, { "epoch": 29.022455089820358, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1216, "step": 19387 }, { "epoch": 29.023952095808383, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1295, "step": 19388 }, { "epoch": 29.025449101796408, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1231, "step": 19389 }, { "epoch": 29.026946107784433, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19390 }, { "epoch": 29.028443113772454, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1263, "step": 19391 }, { "epoch": 29.02994011976048, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1252, "step": 19392 }, { "epoch": 29.031437125748504, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1288, "step": 19393 }, { "epoch": 29.03293413173653, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1246, "step": 19394 }, { "epoch": 29.03443113772455, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1271, "step": 19395 }, { "epoch": 29.035928143712574, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1314, "step": 19396 }, { "epoch": 29.0374251497006, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.131, "step": 19397 }, { "epoch": 29.038922155688624, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1229, "step": 19398 }, { "epoch": 29.040419161676645, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.13, "step": 19399 }, { "epoch": 29.04191616766467, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1244, "step": 19400 }, { "epoch": 29.043413173652695, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1265, "step": 19401 }, { "epoch": 29.04491017964072, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 19402 }, { "epoch": 29.04640718562874, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.124, "step": 19403 }, { "epoch": 29.047904191616766, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1319, "step": 19404 }, { "epoch": 29.04940119760479, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1324, "step": 19405 }, { "epoch": 29.050898203592816, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1269, "step": 19406 }, { "epoch": 29.052395209580837, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1304, "step": 19407 }, { "epoch": 29.05389221556886, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1295, "step": 19408 }, { "epoch": 29.055389221556887, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.132, "step": 19409 }, { "epoch": 29.05688622754491, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1311, "step": 19410 }, { "epoch": 29.058383233532933, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1281, "step": 19411 }, { "epoch": 29.059880239520957, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1237, "step": 19412 }, { "epoch": 29.061377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1314, "step": 19413 }, { "epoch": 29.062874251497007, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1288, "step": 19414 }, { "epoch": 29.06437125748503, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1308, "step": 19415 }, { "epoch": 29.065868263473053, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1355, "step": 19416 }, { "epoch": 29.067365269461078, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1298, "step": 19417 }, { "epoch": 29.068862275449103, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1229, "step": 19418 }, { "epoch": 29.070359281437124, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.126, "step": 19419 }, { "epoch": 29.07185628742515, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.129, "step": 19420 }, { "epoch": 29.073353293413174, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1309, "step": 19421 }, { "epoch": 29.0748502994012, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 19422 }, { "epoch": 29.07634730538922, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1303, "step": 19423 }, { "epoch": 29.077844311377245, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 19424 }, { "epoch": 29.07934131736527, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.126, "step": 19425 }, { "epoch": 29.080838323353294, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1264, "step": 19426 }, { "epoch": 29.082335329341316, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1241, "step": 19427 }, { "epoch": 29.08383233532934, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1221, "step": 19428 }, { "epoch": 29.085329341317365, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1244, "step": 19429 }, { "epoch": 29.08682634730539, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1319, "step": 19430 }, { "epoch": 29.088323353293415, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.122, "step": 19431 }, { "epoch": 29.089820359281436, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1322, "step": 19432 }, { "epoch": 29.09131736526946, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1263, "step": 19433 }, { "epoch": 29.092814371257486, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1258, "step": 19434 }, { "epoch": 29.09431137724551, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1268, "step": 19435 }, { "epoch": 29.095808383233532, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1306, "step": 19436 }, { "epoch": 29.097305389221557, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1283, "step": 19437 }, { "epoch": 29.09880239520958, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.129, "step": 19438 }, { "epoch": 29.100299401197606, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1325, "step": 19439 }, { "epoch": 29.101796407185628, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1284, "step": 19440 }, { "epoch": 29.103293413173652, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1303, "step": 19441 }, { "epoch": 29.104790419161677, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1278, "step": 19442 }, { "epoch": 29.106287425149702, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1325, "step": 19443 }, { "epoch": 29.107784431137723, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.127, "step": 19444 }, { "epoch": 29.10928143712575, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19445 }, { "epoch": 29.110778443113773, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1172, "step": 19446 }, { "epoch": 29.112275449101798, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1305, "step": 19447 }, { "epoch": 29.11377245508982, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1273, "step": 19448 }, { "epoch": 29.115269461077844, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1376, "step": 19449 }, { "epoch": 29.11676646706587, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1315, "step": 19450 }, { "epoch": 29.118263473053894, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1263, "step": 19451 }, { "epoch": 29.119760479041915, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1259, "step": 19452 }, { "epoch": 29.12125748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1243, "step": 19453 }, { "epoch": 29.122754491017965, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1323, "step": 19454 }, { "epoch": 29.12425149700599, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1329, "step": 19455 }, { "epoch": 29.12574850299401, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1338, "step": 19456 }, { "epoch": 29.127245508982035, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1307, "step": 19457 }, { "epoch": 29.12874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1248, "step": 19458 }, { "epoch": 29.130239520958085, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1294, "step": 19459 }, { "epoch": 29.131736526946106, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1315, "step": 19460 }, { "epoch": 29.13323353293413, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1283, "step": 19461 }, { "epoch": 29.134730538922156, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1266, "step": 19462 }, { "epoch": 29.13622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1319, "step": 19463 }, { "epoch": 29.137724550898202, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1264, "step": 19464 }, { "epoch": 29.139221556886227, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1205, "step": 19465 }, { "epoch": 29.14071856287425, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1314, "step": 19466 }, { "epoch": 29.142215568862277, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1294, "step": 19467 }, { "epoch": 29.143712574850298, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1251, "step": 19468 }, { "epoch": 29.145209580838323, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1302, "step": 19469 }, { "epoch": 29.146706586826348, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1338, "step": 19470 }, { "epoch": 29.148203592814372, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19471 }, { "epoch": 29.149700598802394, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1311, "step": 19472 }, { "epoch": 29.15119760479042, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1246, "step": 19473 }, { "epoch": 29.152694610778443, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1253, "step": 19474 }, { "epoch": 29.154191616766468, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.131, "step": 19475 }, { "epoch": 29.15568862275449, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1279, "step": 19476 }, { "epoch": 29.157185628742514, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1298, "step": 19477 }, { "epoch": 29.15868263473054, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1256, "step": 19478 }, { "epoch": 29.160179640718564, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1264, "step": 19479 }, { "epoch": 29.161676646706585, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19480 }, { "epoch": 29.16317365269461, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1241, "step": 19481 }, { "epoch": 29.164670658682635, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1282, "step": 19482 }, { "epoch": 29.16616766467066, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19483 }, { "epoch": 29.16766467065868, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1299, "step": 19484 }, { "epoch": 29.169161676646706, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1269, "step": 19485 }, { "epoch": 29.17065868263473, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1265, "step": 19486 }, { "epoch": 29.172155688622755, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1298, "step": 19487 }, { "epoch": 29.17365269461078, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1375, "step": 19488 }, { "epoch": 29.1751497005988, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.126, "step": 19489 }, { "epoch": 29.176646706586826, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 19490 }, { "epoch": 29.17814371257485, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1333, "step": 19491 }, { "epoch": 29.179640718562876, "grad_norm": 0.047119140625, "learning_rate": 0.0008, "loss": 1.1303, "step": 19492 }, { "epoch": 29.181137724550897, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 19493 }, { "epoch": 29.182634730538922, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1253, "step": 19494 }, { "epoch": 29.184131736526947, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1321, "step": 19495 }, { "epoch": 29.18562874251497, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1326, "step": 19496 }, { "epoch": 29.187125748502993, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1247, "step": 19497 }, { "epoch": 29.188622754491018, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1302, "step": 19498 }, { "epoch": 29.190119760479043, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1326, "step": 19499 }, { "epoch": 29.191616766467067, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.123, "step": 19500 }, { "epoch": 29.19311377245509, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1221, "step": 19501 }, { "epoch": 29.194610778443113, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.125, "step": 19502 }, { "epoch": 29.19610778443114, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1234, "step": 19503 }, { "epoch": 29.197604790419163, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1337, "step": 19504 }, { "epoch": 29.199101796407184, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1196, "step": 19505 }, { "epoch": 29.20059880239521, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.124, "step": 19506 }, { "epoch": 29.202095808383234, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 19507 }, { "epoch": 29.20359281437126, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.129, "step": 19508 }, { "epoch": 29.20508982035928, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1288, "step": 19509 }, { "epoch": 29.206586826347305, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19510 }, { "epoch": 29.20808383233533, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.136, "step": 19511 }, { "epoch": 29.209580838323355, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1297, "step": 19512 }, { "epoch": 29.211077844311376, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1245, "step": 19513 }, { "epoch": 29.2125748502994, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1245, "step": 19514 }, { "epoch": 29.214071856287426, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1264, "step": 19515 }, { "epoch": 29.21556886227545, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1268, "step": 19516 }, { "epoch": 29.21706586826347, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1278, "step": 19517 }, { "epoch": 29.218562874251496, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.132, "step": 19518 }, { "epoch": 29.22005988023952, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1294, "step": 19519 }, { "epoch": 29.221556886227546, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1235, "step": 19520 }, { "epoch": 29.223053892215567, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1351, "step": 19521 }, { "epoch": 29.224550898203592, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.13, "step": 19522 }, { "epoch": 29.226047904191617, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1286, "step": 19523 }, { "epoch": 29.227544910179642, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1212, "step": 19524 }, { "epoch": 29.229041916167663, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1282, "step": 19525 }, { "epoch": 29.230538922155688, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1347, "step": 19526 }, { "epoch": 29.232035928143713, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1286, "step": 19527 }, { "epoch": 29.233532934131738, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1316, "step": 19528 }, { "epoch": 29.23502994011976, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19529 }, { "epoch": 29.236526946107784, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1245, "step": 19530 }, { "epoch": 29.23802395209581, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1276, "step": 19531 }, { "epoch": 29.239520958083833, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.123, "step": 19532 }, { "epoch": 29.241017964071855, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.127, "step": 19533 }, { "epoch": 29.24251497005988, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 19534 }, { "epoch": 29.244011976047904, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1261, "step": 19535 }, { "epoch": 29.24550898203593, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1335, "step": 19536 }, { "epoch": 29.24700598802395, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1288, "step": 19537 }, { "epoch": 29.248502994011975, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1295, "step": 19538 }, { "epoch": 29.25, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1228, "step": 19539 }, { "epoch": 29.251497005988025, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1256, "step": 19540 }, { "epoch": 29.25299401197605, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1298, "step": 19541 }, { "epoch": 29.25449101796407, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1235, "step": 19542 }, { "epoch": 29.255988023952096, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19543 }, { "epoch": 29.25748502994012, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1237, "step": 19544 }, { "epoch": 29.258982035928145, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1284, "step": 19545 }, { "epoch": 29.260479041916167, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1236, "step": 19546 }, { "epoch": 29.26197604790419, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1262, "step": 19547 }, { "epoch": 29.263473053892216, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1191, "step": 19548 }, { "epoch": 29.26497005988024, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1243, "step": 19549 }, { "epoch": 29.266467065868262, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1195, "step": 19550 }, { "epoch": 29.267964071856287, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1244, "step": 19551 }, { "epoch": 29.269461077844312, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1293, "step": 19552 }, { "epoch": 29.270958083832337, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1233, "step": 19553 }, { "epoch": 29.272455089820358, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1258, "step": 19554 }, { "epoch": 29.273952095808383, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1223, "step": 19555 }, { "epoch": 29.275449101796408, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1198, "step": 19556 }, { "epoch": 29.276946107784433, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1342, "step": 19557 }, { "epoch": 29.278443113772454, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1302, "step": 19558 }, { "epoch": 29.27994011976048, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1292, "step": 19559 }, { "epoch": 29.281437125748504, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1291, "step": 19560 }, { "epoch": 29.28293413173653, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1321, "step": 19561 }, { "epoch": 29.28443113772455, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1296, "step": 19562 }, { "epoch": 29.285928143712574, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1248, "step": 19563 }, { "epoch": 29.2874251497006, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1229, "step": 19564 }, { "epoch": 29.288922155688624, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1272, "step": 19565 }, { "epoch": 29.290419161676645, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.134, "step": 19566 }, { "epoch": 29.29191616766467, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 19567 }, { "epoch": 29.293413173652695, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 19568 }, { "epoch": 29.29491017964072, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1246, "step": 19569 }, { "epoch": 29.29640718562874, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1261, "step": 19570 }, { "epoch": 29.297904191616766, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1316, "step": 19571 }, { "epoch": 29.29940119760479, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1214, "step": 19572 }, { "epoch": 29.300898203592816, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1271, "step": 19573 }, { "epoch": 29.302395209580837, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1209, "step": 19574 }, { "epoch": 29.30389221556886, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1259, "step": 19575 }, { "epoch": 29.305389221556887, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1324, "step": 19576 }, { "epoch": 29.30688622754491, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1273, "step": 19577 }, { "epoch": 29.308383233532933, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1297, "step": 19578 }, { "epoch": 29.309880239520957, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1346, "step": 19579 }, { "epoch": 29.311377245508982, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.125, "step": 19580 }, { "epoch": 29.312874251497007, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.129, "step": 19581 }, { "epoch": 29.31437125748503, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1317, "step": 19582 }, { "epoch": 29.315868263473053, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1351, "step": 19583 }, { "epoch": 29.317365269461078, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 19584 }, { "epoch": 29.318862275449103, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.128, "step": 19585 }, { "epoch": 29.320359281437124, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.125, "step": 19586 }, { "epoch": 29.32185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1335, "step": 19587 }, { "epoch": 29.323353293413174, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1313, "step": 19588 }, { "epoch": 29.3248502994012, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 19589 }, { "epoch": 29.32634730538922, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.129, "step": 19590 }, { "epoch": 29.327844311377245, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1316, "step": 19591 }, { "epoch": 29.32934131736527, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 19592 }, { "epoch": 29.330838323353294, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1368, "step": 19593 }, { "epoch": 29.33233532934132, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1342, "step": 19594 }, { "epoch": 29.33383233532934, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1276, "step": 19595 }, { "epoch": 29.335329341317365, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1299, "step": 19596 }, { "epoch": 29.33682634730539, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1316, "step": 19597 }, { "epoch": 29.338323353293415, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.132, "step": 19598 }, { "epoch": 29.339820359281436, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1307, "step": 19599 }, { "epoch": 29.34131736526946, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1251, "step": 19600 }, { "epoch": 29.342814371257486, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1298, "step": 19601 }, { "epoch": 29.34431137724551, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1248, "step": 19602 }, { "epoch": 29.345808383233532, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 19603 }, { "epoch": 29.347305389221557, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1206, "step": 19604 }, { "epoch": 29.34880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1268, "step": 19605 }, { "epoch": 29.350299401197606, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1228, "step": 19606 }, { "epoch": 29.351796407185628, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19607 }, { "epoch": 29.353293413173652, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 19608 }, { "epoch": 29.354790419161677, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1332, "step": 19609 }, { "epoch": 29.356287425149702, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1287, "step": 19610 }, { "epoch": 29.357784431137723, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1259, "step": 19611 }, { "epoch": 29.35928143712575, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1266, "step": 19612 }, { "epoch": 29.360778443113773, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1345, "step": 19613 }, { "epoch": 29.362275449101798, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1302, "step": 19614 }, { "epoch": 29.36377245508982, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1276, "step": 19615 }, { "epoch": 29.365269461077844, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1274, "step": 19616 }, { "epoch": 29.36676646706587, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1243, "step": 19617 }, { "epoch": 29.368263473053894, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 19618 }, { "epoch": 29.369760479041915, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1262, "step": 19619 }, { "epoch": 29.37125748502994, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1305, "step": 19620 }, { "epoch": 29.372754491017965, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.128, "step": 19621 }, { "epoch": 29.37425149700599, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1262, "step": 19622 }, { "epoch": 29.37574850299401, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.127, "step": 19623 }, { "epoch": 29.377245508982035, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1261, "step": 19624 }, { "epoch": 29.37874251497006, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1271, "step": 19625 }, { "epoch": 29.380239520958085, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1371, "step": 19626 }, { "epoch": 29.381736526946106, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1264, "step": 19627 }, { "epoch": 29.38323353293413, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1313, "step": 19628 }, { "epoch": 29.384730538922156, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1306, "step": 19629 }, { "epoch": 29.38622754491018, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1288, "step": 19630 }, { "epoch": 29.387724550898202, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1293, "step": 19631 }, { "epoch": 29.389221556886227, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1249, "step": 19632 }, { "epoch": 29.39071856287425, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1325, "step": 19633 }, { "epoch": 29.392215568862277, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.132, "step": 19634 }, { "epoch": 29.393712574850298, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.133, "step": 19635 }, { "epoch": 29.395209580838323, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 19636 }, { "epoch": 29.396706586826348, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1295, "step": 19637 }, { "epoch": 29.398203592814372, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19638 }, { "epoch": 29.399700598802394, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19639 }, { "epoch": 29.40119760479042, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1301, "step": 19640 }, { "epoch": 29.402694610778443, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1321, "step": 19641 }, { "epoch": 29.404191616766468, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1248, "step": 19642 }, { "epoch": 29.40568862275449, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1228, "step": 19643 }, { "epoch": 29.407185628742514, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1333, "step": 19644 }, { "epoch": 29.40868263473054, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1337, "step": 19645 }, { "epoch": 29.410179640718564, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1271, "step": 19646 }, { "epoch": 29.411676646706585, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.125, "step": 19647 }, { "epoch": 29.41317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1317, "step": 19648 }, { "epoch": 29.414670658682635, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1287, "step": 19649 }, { "epoch": 29.41616766467066, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1277, "step": 19650 }, { "epoch": 29.41766467065868, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1316, "step": 19651 }, { "epoch": 29.419161676646706, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1273, "step": 19652 }, { "epoch": 29.42065868263473, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1299, "step": 19653 }, { "epoch": 29.422155688622755, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1199, "step": 19654 }, { "epoch": 29.42365269461078, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1276, "step": 19655 }, { "epoch": 29.4251497005988, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1321, "step": 19656 }, { "epoch": 29.426646706586826, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 19657 }, { "epoch": 29.42814371257485, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1284, "step": 19658 }, { "epoch": 29.429640718562876, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.124, "step": 19659 }, { "epoch": 29.431137724550897, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1279, "step": 19660 }, { "epoch": 29.432634730538922, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 19661 }, { "epoch": 29.434131736526947, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1258, "step": 19662 }, { "epoch": 29.43562874251497, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1285, "step": 19663 }, { "epoch": 29.437125748502993, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1305, "step": 19664 }, { "epoch": 29.438622754491018, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1261, "step": 19665 }, { "epoch": 29.440119760479043, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1303, "step": 19666 }, { "epoch": 29.441616766467067, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1349, "step": 19667 }, { "epoch": 29.44311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1265, "step": 19668 }, { "epoch": 29.444610778443113, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1342, "step": 19669 }, { "epoch": 29.44610778443114, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1223, "step": 19670 }, { "epoch": 29.447604790419163, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1333, "step": 19671 }, { "epoch": 29.449101796407184, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1337, "step": 19672 }, { "epoch": 29.45059880239521, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1291, "step": 19673 }, { "epoch": 29.452095808383234, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1224, "step": 19674 }, { "epoch": 29.45359281437126, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1286, "step": 19675 }, { "epoch": 29.45508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.122, "step": 19676 }, { "epoch": 29.456586826347305, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19677 }, { "epoch": 29.45808383233533, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.131, "step": 19678 }, { "epoch": 29.459580838323355, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19679 }, { "epoch": 29.461077844311376, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 19680 }, { "epoch": 29.4625748502994, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1292, "step": 19681 }, { "epoch": 29.464071856287426, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1255, "step": 19682 }, { "epoch": 29.46556886227545, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1247, "step": 19683 }, { "epoch": 29.46706586826347, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.124, "step": 19684 }, { "epoch": 29.468562874251496, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1259, "step": 19685 }, { "epoch": 29.47005988023952, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1269, "step": 19686 }, { "epoch": 29.471556886227546, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1292, "step": 19687 }, { "epoch": 29.473053892215567, "grad_norm": 0.050537109375, "learning_rate": 0.0008, "loss": 1.1275, "step": 19688 }, { "epoch": 29.474550898203592, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.13, "step": 19689 }, { "epoch": 29.476047904191617, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1279, "step": 19690 }, { "epoch": 29.477544910179642, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1331, "step": 19691 }, { "epoch": 29.479041916167663, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.1321, "step": 19692 }, { "epoch": 29.480538922155688, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1327, "step": 19693 }, { "epoch": 29.482035928143713, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1332, "step": 19694 }, { "epoch": 29.483532934131738, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1274, "step": 19695 }, { "epoch": 29.48502994011976, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1277, "step": 19696 }, { "epoch": 29.486526946107784, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1262, "step": 19697 }, { "epoch": 29.48802395209581, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1212, "step": 19698 }, { "epoch": 29.489520958083833, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1218, "step": 19699 }, { "epoch": 29.491017964071855, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1297, "step": 19700 }, { "epoch": 29.49251497005988, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1264, "step": 19701 }, { "epoch": 29.494011976047904, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 19702 }, { "epoch": 29.49550898203593, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1273, "step": 19703 }, { "epoch": 29.49700598802395, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1344, "step": 19704 }, { "epoch": 29.498502994011975, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.13, "step": 19705 }, { "epoch": 29.5, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1282, "step": 19706 }, { "epoch": 29.501497005988025, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1251, "step": 19707 }, { "epoch": 29.50299401197605, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1284, "step": 19708 }, { "epoch": 29.50449101796407, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1346, "step": 19709 }, { "epoch": 29.505988023952096, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1274, "step": 19710 }, { "epoch": 29.50748502994012, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1218, "step": 19711 }, { "epoch": 29.508982035928145, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1309, "step": 19712 }, { "epoch": 29.510479041916167, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1274, "step": 19713 }, { "epoch": 29.51197604790419, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1263, "step": 19714 }, { "epoch": 29.513473053892216, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1291, "step": 19715 }, { "epoch": 29.51497005988024, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1258, "step": 19716 }, { "epoch": 29.516467065868262, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1365, "step": 19717 }, { "epoch": 29.517964071856287, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1212, "step": 19718 }, { "epoch": 29.519461077844312, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1323, "step": 19719 }, { "epoch": 29.520958083832337, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1299, "step": 19720 }, { "epoch": 29.522455089820358, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 19721 }, { "epoch": 29.523952095808383, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1268, "step": 19722 }, { "epoch": 29.525449101796408, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1248, "step": 19723 }, { "epoch": 29.526946107784433, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1261, "step": 19724 }, { "epoch": 29.528443113772454, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1277, "step": 19725 }, { "epoch": 29.52994011976048, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1308, "step": 19726 }, { "epoch": 29.531437125748504, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1209, "step": 19727 }, { "epoch": 29.53293413173653, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1266, "step": 19728 }, { "epoch": 29.53443113772455, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1344, "step": 19729 }, { "epoch": 29.535928143712574, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1327, "step": 19730 }, { "epoch": 29.5374251497006, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 19731 }, { "epoch": 29.538922155688624, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1294, "step": 19732 }, { "epoch": 29.540419161676645, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1304, "step": 19733 }, { "epoch": 29.54191616766467, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19734 }, { "epoch": 29.543413173652695, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 19735 }, { "epoch": 29.54491017964072, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 19736 }, { "epoch": 29.54640718562874, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1304, "step": 19737 }, { "epoch": 29.547904191616766, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1252, "step": 19738 }, { "epoch": 29.54940119760479, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1291, "step": 19739 }, { "epoch": 29.550898203592816, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1274, "step": 19740 }, { "epoch": 29.552395209580837, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 19741 }, { "epoch": 29.55389221556886, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1344, "step": 19742 }, { "epoch": 29.555389221556887, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1281, "step": 19743 }, { "epoch": 29.55688622754491, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1217, "step": 19744 }, { "epoch": 29.558383233532933, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1237, "step": 19745 }, { "epoch": 29.559880239520957, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1258, "step": 19746 }, { "epoch": 29.561377245508982, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1242, "step": 19747 }, { "epoch": 29.562874251497007, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1276, "step": 19748 }, { "epoch": 29.56437125748503, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1271, "step": 19749 }, { "epoch": 29.565868263473053, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19750 }, { "epoch": 29.567365269461078, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1231, "step": 19751 }, { "epoch": 29.568862275449103, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1242, "step": 19752 }, { "epoch": 29.570359281437124, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19753 }, { "epoch": 29.57185628742515, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 19754 }, { "epoch": 29.573353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1278, "step": 19755 }, { "epoch": 29.5748502994012, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1287, "step": 19756 }, { "epoch": 29.57634730538922, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1234, "step": 19757 }, { "epoch": 29.577844311377245, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1331, "step": 19758 }, { "epoch": 29.57934131736527, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 19759 }, { "epoch": 29.580838323353294, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1332, "step": 19760 }, { "epoch": 29.58233532934132, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1252, "step": 19761 }, { "epoch": 29.58383233532934, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1289, "step": 19762 }, { "epoch": 29.585329341317365, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.125, "step": 19763 }, { "epoch": 29.58682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1285, "step": 19764 }, { "epoch": 29.58832335329341, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1306, "step": 19765 }, { "epoch": 29.589820359281436, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1314, "step": 19766 }, { "epoch": 29.59131736526946, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1267, "step": 19767 }, { "epoch": 29.592814371257486, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.129, "step": 19768 }, { "epoch": 29.59431137724551, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1217, "step": 19769 }, { "epoch": 29.595808383233532, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1223, "step": 19770 }, { "epoch": 29.597305389221557, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1214, "step": 19771 }, { "epoch": 29.59880239520958, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1267, "step": 19772 }, { "epoch": 29.600299401197606, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1265, "step": 19773 }, { "epoch": 29.601796407185628, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1308, "step": 19774 }, { "epoch": 29.603293413173652, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1287, "step": 19775 }, { "epoch": 29.604790419161677, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1311, "step": 19776 }, { "epoch": 29.606287425149702, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1251, "step": 19777 }, { "epoch": 29.607784431137723, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1244, "step": 19778 }, { "epoch": 29.60928143712575, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 19779 }, { "epoch": 29.610778443113773, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19780 }, { "epoch": 29.612275449101798, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1255, "step": 19781 }, { "epoch": 29.61377245508982, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1279, "step": 19782 }, { "epoch": 29.615269461077844, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1247, "step": 19783 }, { "epoch": 29.61676646706587, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1266, "step": 19784 }, { "epoch": 29.618263473053894, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1319, "step": 19785 }, { "epoch": 29.619760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 19786 }, { "epoch": 29.62125748502994, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1245, "step": 19787 }, { "epoch": 29.622754491017965, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1294, "step": 19788 }, { "epoch": 29.62425149700599, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1213, "step": 19789 }, { "epoch": 29.62574850299401, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1308, "step": 19790 }, { "epoch": 29.627245508982035, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1243, "step": 19791 }, { "epoch": 29.62874251497006, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1343, "step": 19792 }, { "epoch": 29.630239520958085, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.135, "step": 19793 }, { "epoch": 29.631736526946106, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1283, "step": 19794 }, { "epoch": 29.63323353293413, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.127, "step": 19795 }, { "epoch": 29.634730538922156, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19796 }, { "epoch": 29.63622754491018, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1314, "step": 19797 }, { "epoch": 29.637724550898202, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1345, "step": 19798 }, { "epoch": 29.639221556886227, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1344, "step": 19799 }, { "epoch": 29.64071856287425, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19800 }, { "epoch": 29.642215568862277, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1247, "step": 19801 }, { "epoch": 29.643712574850298, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1242, "step": 19802 }, { "epoch": 29.645209580838323, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1237, "step": 19803 }, { "epoch": 29.646706586826348, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19804 }, { "epoch": 29.648203592814372, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1257, "step": 19805 }, { "epoch": 29.649700598802394, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1311, "step": 19806 }, { "epoch": 29.65119760479042, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.128, "step": 19807 }, { "epoch": 29.652694610778443, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1211, "step": 19808 }, { "epoch": 29.654191616766468, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1301, "step": 19809 }, { "epoch": 29.65568862275449, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 19810 }, { "epoch": 29.657185628742514, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 19811 }, { "epoch": 29.65868263473054, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1249, "step": 19812 }, { "epoch": 29.660179640718564, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1208, "step": 19813 }, { "epoch": 29.66167664670659, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.128, "step": 19814 }, { "epoch": 29.66317365269461, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1195, "step": 19815 }, { "epoch": 29.664670658682635, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1284, "step": 19816 }, { "epoch": 29.66616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1239, "step": 19817 }, { "epoch": 29.66766467065868, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1248, "step": 19818 }, { "epoch": 29.669161676646706, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1177, "step": 19819 }, { "epoch": 29.67065868263473, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1227, "step": 19820 }, { "epoch": 29.672155688622755, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1241, "step": 19821 }, { "epoch": 29.67365269461078, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1235, "step": 19822 }, { "epoch": 29.6751497005988, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1257, "step": 19823 }, { "epoch": 29.676646706586826, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1306, "step": 19824 }, { "epoch": 29.67814371257485, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1319, "step": 19825 }, { "epoch": 29.679640718562876, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.125, "step": 19826 }, { "epoch": 29.681137724550897, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1272, "step": 19827 }, { "epoch": 29.682634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1286, "step": 19828 }, { "epoch": 29.684131736526947, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1292, "step": 19829 }, { "epoch": 29.68562874251497, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1326, "step": 19830 }, { "epoch": 29.687125748502993, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.123, "step": 19831 }, { "epoch": 29.688622754491018, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1313, "step": 19832 }, { "epoch": 29.690119760479043, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1336, "step": 19833 }, { "epoch": 29.691616766467067, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1245, "step": 19834 }, { "epoch": 29.69311377245509, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19835 }, { "epoch": 29.694610778443113, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1207, "step": 19836 }, { "epoch": 29.69610778443114, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1281, "step": 19837 }, { "epoch": 29.697604790419163, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1262, "step": 19838 }, { "epoch": 29.699101796407184, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19839 }, { "epoch": 29.70059880239521, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1305, "step": 19840 }, { "epoch": 29.702095808383234, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1273, "step": 19841 }, { "epoch": 29.70359281437126, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1272, "step": 19842 }, { "epoch": 29.70508982035928, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1273, "step": 19843 }, { "epoch": 29.706586826347305, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1274, "step": 19844 }, { "epoch": 29.70808383233533, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 19845 }, { "epoch": 29.709580838323355, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1229, "step": 19846 }, { "epoch": 29.711077844311376, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.125, "step": 19847 }, { "epoch": 29.7125748502994, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1315, "step": 19848 }, { "epoch": 29.714071856287426, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1275, "step": 19849 }, { "epoch": 29.71556886227545, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1303, "step": 19850 }, { "epoch": 29.71706586826347, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.124, "step": 19851 }, { "epoch": 29.718562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1255, "step": 19852 }, { "epoch": 29.72005988023952, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1246, "step": 19853 }, { "epoch": 29.721556886227546, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1287, "step": 19854 }, { "epoch": 29.723053892215567, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.126, "step": 19855 }, { "epoch": 29.724550898203592, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1256, "step": 19856 }, { "epoch": 29.726047904191617, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1262, "step": 19857 }, { "epoch": 29.727544910179642, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1292, "step": 19858 }, { "epoch": 29.729041916167663, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1288, "step": 19859 }, { "epoch": 29.730538922155688, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1319, "step": 19860 }, { "epoch": 29.732035928143713, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1259, "step": 19861 }, { "epoch": 29.733532934131738, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1282, "step": 19862 }, { "epoch": 29.73502994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1213, "step": 19863 }, { "epoch": 29.736526946107784, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1284, "step": 19864 }, { "epoch": 29.73802395209581, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1314, "step": 19865 }, { "epoch": 29.739520958083833, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1268, "step": 19866 }, { "epoch": 29.741017964071855, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1306, "step": 19867 }, { "epoch": 29.74251497005988, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1317, "step": 19868 }, { "epoch": 29.744011976047904, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1292, "step": 19869 }, { "epoch": 29.74550898203593, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1266, "step": 19870 }, { "epoch": 29.74700598802395, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1228, "step": 19871 }, { "epoch": 29.748502994011975, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1259, "step": 19872 }, { "epoch": 29.75, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1354, "step": 19873 }, { "epoch": 29.751497005988025, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19874 }, { "epoch": 29.75299401197605, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.125, "step": 19875 }, { "epoch": 29.75449101796407, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1222, "step": 19876 }, { "epoch": 29.755988023952096, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1317, "step": 19877 }, { "epoch": 29.75748502994012, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1282, "step": 19878 }, { "epoch": 29.758982035928145, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1359, "step": 19879 }, { "epoch": 29.760479041916167, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1285, "step": 19880 }, { "epoch": 29.76197604790419, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1186, "step": 19881 }, { "epoch": 29.763473053892216, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1325, "step": 19882 }, { "epoch": 29.76497005988024, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1233, "step": 19883 }, { "epoch": 29.766467065868262, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1227, "step": 19884 }, { "epoch": 29.767964071856287, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1275, "step": 19885 }, { "epoch": 29.769461077844312, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1286, "step": 19886 }, { "epoch": 29.770958083832337, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19887 }, { "epoch": 29.772455089820358, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1267, "step": 19888 }, { "epoch": 29.773952095808383, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1281, "step": 19889 }, { "epoch": 29.775449101796408, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1334, "step": 19890 }, { "epoch": 29.776946107784433, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1288, "step": 19891 }, { "epoch": 29.778443113772454, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 19892 }, { "epoch": 29.77994011976048, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1239, "step": 19893 }, { "epoch": 29.781437125748504, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1318, "step": 19894 }, { "epoch": 29.78293413173653, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1235, "step": 19895 }, { "epoch": 29.78443113772455, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1319, "step": 19896 }, { "epoch": 29.785928143712574, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.129, "step": 19897 }, { "epoch": 29.7874251497006, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19898 }, { "epoch": 29.788922155688624, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1238, "step": 19899 }, { "epoch": 29.790419161676645, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1175, "step": 19900 }, { "epoch": 29.79191616766467, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1254, "step": 19901 }, { "epoch": 29.793413173652695, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1228, "step": 19902 }, { "epoch": 29.79491017964072, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1266, "step": 19903 }, { "epoch": 29.79640718562874, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1321, "step": 19904 }, { "epoch": 29.797904191616766, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1355, "step": 19905 }, { "epoch": 29.79940119760479, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1287, "step": 19906 }, { "epoch": 29.800898203592816, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1321, "step": 19907 }, { "epoch": 29.802395209580837, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1269, "step": 19908 }, { "epoch": 29.80389221556886, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1214, "step": 19909 }, { "epoch": 29.805389221556887, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.134, "step": 19910 }, { "epoch": 29.80688622754491, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1258, "step": 19911 }, { "epoch": 29.808383233532933, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1297, "step": 19912 }, { "epoch": 29.809880239520957, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1243, "step": 19913 }, { "epoch": 29.811377245508982, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1256, "step": 19914 }, { "epoch": 29.812874251497007, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1259, "step": 19915 }, { "epoch": 29.81437125748503, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1325, "step": 19916 }, { "epoch": 29.815868263473053, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1353, "step": 19917 }, { "epoch": 29.817365269461078, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1346, "step": 19918 }, { "epoch": 29.818862275449103, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1297, "step": 19919 }, { "epoch": 29.820359281437124, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1301, "step": 19920 }, { "epoch": 29.82185628742515, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1349, "step": 19921 }, { "epoch": 29.823353293413174, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.12, "step": 19922 }, { "epoch": 29.8248502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1273, "step": 19923 }, { "epoch": 29.82634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1337, "step": 19924 }, { "epoch": 29.827844311377245, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1231, "step": 19925 }, { "epoch": 29.82934131736527, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1231, "step": 19926 }, { "epoch": 29.830838323353294, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1181, "step": 19927 }, { "epoch": 29.83233532934132, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1276, "step": 19928 }, { "epoch": 29.83383233532934, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1297, "step": 19929 }, { "epoch": 29.835329341317365, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1329, "step": 19930 }, { "epoch": 29.83682634730539, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1298, "step": 19931 }, { "epoch": 29.83832335329341, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1238, "step": 19932 }, { "epoch": 29.839820359281436, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1259, "step": 19933 }, { "epoch": 29.84131736526946, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1225, "step": 19934 }, { "epoch": 29.842814371257486, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1273, "step": 19935 }, { "epoch": 29.84431137724551, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1321, "step": 19936 }, { "epoch": 29.845808383233532, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.127, "step": 19937 }, { "epoch": 29.847305389221557, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1269, "step": 19938 }, { "epoch": 29.84880239520958, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 19939 }, { "epoch": 29.850299401197606, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1336, "step": 19940 }, { "epoch": 29.851796407185628, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1236, "step": 19941 }, { "epoch": 29.853293413173652, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1287, "step": 19942 }, { "epoch": 29.854790419161677, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 19943 }, { "epoch": 29.856287425149702, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1219, "step": 19944 }, { "epoch": 29.857784431137723, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1339, "step": 19945 }, { "epoch": 29.85928143712575, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1268, "step": 19946 }, { "epoch": 29.860778443113773, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.123, "step": 19947 }, { "epoch": 29.862275449101798, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1252, "step": 19948 }, { "epoch": 29.86377245508982, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1319, "step": 19949 }, { "epoch": 29.865269461077844, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1243, "step": 19950 }, { "epoch": 29.86676646706587, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1233, "step": 19951 }, { "epoch": 29.868263473053894, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1281, "step": 19952 }, { "epoch": 29.869760479041915, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1258, "step": 19953 }, { "epoch": 29.87125748502994, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1252, "step": 19954 }, { "epoch": 29.872754491017965, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1247, "step": 19955 }, { "epoch": 29.87425149700599, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1231, "step": 19956 }, { "epoch": 29.87574850299401, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1257, "step": 19957 }, { "epoch": 29.877245508982035, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 19958 }, { "epoch": 29.87874251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.125, "step": 19959 }, { "epoch": 29.880239520958085, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1238, "step": 19960 }, { "epoch": 29.881736526946106, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1246, "step": 19961 }, { "epoch": 29.88323353293413, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1268, "step": 19962 }, { "epoch": 29.884730538922156, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1262, "step": 19963 }, { "epoch": 29.88622754491018, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1219, "step": 19964 }, { "epoch": 29.887724550898202, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1267, "step": 19965 }, { "epoch": 29.889221556886227, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1279, "step": 19966 }, { "epoch": 29.89071856287425, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1217, "step": 19967 }, { "epoch": 29.892215568862277, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1284, "step": 19968 }, { "epoch": 29.893712574850298, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1299, "step": 19969 }, { "epoch": 29.895209580838323, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1277, "step": 19970 }, { "epoch": 29.896706586826348, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1261, "step": 19971 }, { "epoch": 29.898203592814372, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1258, "step": 19972 }, { "epoch": 29.899700598802394, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1279, "step": 19973 }, { "epoch": 29.90119760479042, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1249, "step": 19974 }, { "epoch": 29.902694610778443, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1298, "step": 19975 }, { "epoch": 29.904191616766468, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1255, "step": 19976 }, { "epoch": 29.90568862275449, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19977 }, { "epoch": 29.907185628742514, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19978 }, { "epoch": 29.90868263473054, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 19979 }, { "epoch": 29.910179640718564, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.132, "step": 19980 }, { "epoch": 29.91167664670659, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1325, "step": 19981 }, { "epoch": 29.91317365269461, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1234, "step": 19982 }, { "epoch": 29.914670658682635, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19983 }, { "epoch": 29.91616766467066, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1237, "step": 19984 }, { "epoch": 29.91766467065868, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1259, "step": 19985 }, { "epoch": 29.919161676646706, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 19986 }, { "epoch": 29.92065868263473, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1237, "step": 19987 }, { "epoch": 29.922155688622755, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 19988 }, { "epoch": 29.92365269461078, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1272, "step": 19989 }, { "epoch": 29.9251497005988, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.127, "step": 19990 }, { "epoch": 29.926646706586826, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1278, "step": 19991 }, { "epoch": 29.92814371257485, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19992 }, { "epoch": 29.929640718562876, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1305, "step": 19993 }, { "epoch": 29.931137724550897, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1208, "step": 19994 }, { "epoch": 29.932634730538922, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 19995 }, { "epoch": 29.934131736526947, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1308, "step": 19996 }, { "epoch": 29.93562874251497, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 19997 }, { "epoch": 29.937125748502993, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1325, "step": 19998 }, { "epoch": 29.938622754491018, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19999 }, { "epoch": 29.940119760479043, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1267, "step": 20000 }, { "epoch": 29.941616766467067, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1288, "step": 20001 }, { "epoch": 29.94311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1288, "step": 20002 }, { "epoch": 29.944610778443113, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1292, "step": 20003 }, { "epoch": 29.94610778443114, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1348, "step": 20004 }, { "epoch": 29.947604790419163, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1295, "step": 20005 }, { "epoch": 29.949101796407184, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 20006 }, { "epoch": 29.95059880239521, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1237, "step": 20007 }, { "epoch": 29.952095808383234, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1296, "step": 20008 }, { "epoch": 29.95359281437126, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1319, "step": 20009 }, { "epoch": 29.95508982035928, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1247, "step": 20010 }, { "epoch": 29.956586826347305, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1276, "step": 20011 }, { "epoch": 29.95808383233533, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1279, "step": 20012 }, { "epoch": 29.959580838323355, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1227, "step": 20013 }, { "epoch": 29.961077844311376, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1305, "step": 20014 }, { "epoch": 29.9625748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1332, "step": 20015 }, { "epoch": 29.964071856287426, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 20016 }, { "epoch": 29.96556886227545, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1241, "step": 20017 }, { "epoch": 29.96706586826347, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.125, "step": 20018 }, { "epoch": 29.968562874251496, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1242, "step": 20019 }, { "epoch": 29.97005988023952, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1248, "step": 20020 }, { "epoch": 29.971556886227546, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1215, "step": 20021 }, { "epoch": 29.973053892215567, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1312, "step": 20022 }, { "epoch": 29.974550898203592, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.13, "step": 20023 }, { "epoch": 29.976047904191617, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1251, "step": 20024 }, { "epoch": 29.977544910179642, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1221, "step": 20025 }, { "epoch": 29.979041916167663, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1292, "step": 20026 }, { "epoch": 29.980538922155688, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1304, "step": 20027 }, { "epoch": 29.982035928143713, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1268, "step": 20028 }, { "epoch": 29.983532934131738, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1265, "step": 20029 }, { "epoch": 29.98502994011976, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 20030 }, { "epoch": 29.986526946107784, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1243, "step": 20031 }, { "epoch": 29.98802395209581, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1345, "step": 20032 }, { "epoch": 29.989520958083833, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1264, "step": 20033 }, { "epoch": 29.991017964071855, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1285, "step": 20034 }, { "epoch": 29.99251497005988, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1256, "step": 20035 }, { "epoch": 29.994011976047904, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1333, "step": 20036 }, { "epoch": 29.99550898203593, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1273, "step": 20037 }, { "epoch": 29.99700598802395, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1312, "step": 20038 }, { "epoch": 29.998502994011975, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.127, "step": 20039 }, { "epoch": 30.0, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1287, "step": 20040 }, { "epoch": 30.001497005988025, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.129, "step": 20041 }, { "epoch": 30.00299401197605, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1221, "step": 20042 }, { "epoch": 30.00449101796407, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1312, "step": 20043 }, { "epoch": 30.005988023952096, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1238, "step": 20044 }, { "epoch": 30.00748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1317, "step": 20045 }, { "epoch": 30.008982035928145, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.124, "step": 20046 }, { "epoch": 30.010479041916167, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1251, "step": 20047 }, { "epoch": 30.01197604790419, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.129, "step": 20048 }, { "epoch": 30.013473053892216, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1304, "step": 20049 }, { "epoch": 30.01497005988024, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1272, "step": 20050 }, { "epoch": 30.016467065868262, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1176, "step": 20051 }, { "epoch": 30.017964071856287, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1248, "step": 20052 }, { "epoch": 30.019461077844312, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 20053 }, { "epoch": 30.020958083832337, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1265, "step": 20054 }, { "epoch": 30.022455089820358, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1293, "step": 20055 }, { "epoch": 30.023952095808383, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1209, "step": 20056 }, { "epoch": 30.025449101796408, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1301, "step": 20057 }, { "epoch": 30.026946107784433, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1314, "step": 20058 }, { "epoch": 30.028443113772454, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1217, "step": 20059 }, { "epoch": 30.02994011976048, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1221, "step": 20060 }, { "epoch": 30.031437125748504, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1224, "step": 20061 }, { "epoch": 30.03293413173653, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1241, "step": 20062 }, { "epoch": 30.03443113772455, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.125, "step": 20063 }, { "epoch": 30.035928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1286, "step": 20064 }, { "epoch": 30.0374251497006, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.124, "step": 20065 }, { "epoch": 30.038922155688624, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1221, "step": 20066 }, { "epoch": 30.040419161676645, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1294, "step": 20067 }, { "epoch": 30.04191616766467, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1281, "step": 20068 }, { "epoch": 30.043413173652695, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1274, "step": 20069 }, { "epoch": 30.04491017964072, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1282, "step": 20070 }, { "epoch": 30.04640718562874, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1345, "step": 20071 }, { "epoch": 30.047904191616766, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1241, "step": 20072 }, { "epoch": 30.04940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1247, "step": 20073 }, { "epoch": 30.050898203592816, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1318, "step": 20074 }, { "epoch": 30.052395209580837, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1245, "step": 20075 }, { "epoch": 30.05389221556886, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1251, "step": 20076 }, { "epoch": 30.055389221556887, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1286, "step": 20077 }, { "epoch": 30.05688622754491, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1355, "step": 20078 }, { "epoch": 30.058383233532933, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1271, "step": 20079 }, { "epoch": 30.059880239520957, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1259, "step": 20080 }, { "epoch": 30.061377245508982, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.127, "step": 20081 }, { "epoch": 30.062874251497007, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1232, "step": 20082 }, { "epoch": 30.06437125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1269, "step": 20083 }, { "epoch": 30.065868263473053, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1367, "step": 20084 }, { "epoch": 30.067365269461078, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1242, "step": 20085 }, { "epoch": 30.068862275449103, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1301, "step": 20086 }, { "epoch": 30.070359281437124, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1239, "step": 20087 }, { "epoch": 30.07185628742515, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1329, "step": 20088 }, { "epoch": 30.073353293413174, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1296, "step": 20089 }, { "epoch": 30.0748502994012, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1309, "step": 20090 }, { "epoch": 30.07634730538922, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1287, "step": 20091 }, { "epoch": 30.077844311377245, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1238, "step": 20092 }, { "epoch": 30.07934131736527, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1274, "step": 20093 }, { "epoch": 30.080838323353294, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1207, "step": 20094 }, { "epoch": 30.082335329341316, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1261, "step": 20095 }, { "epoch": 30.08383233532934, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1264, "step": 20096 }, { "epoch": 30.085329341317365, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1229, "step": 20097 }, { "epoch": 30.08682634730539, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1292, "step": 20098 }, { "epoch": 30.088323353293415, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1266, "step": 20099 }, { "epoch": 30.089820359281436, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1249, "step": 20100 }, { "epoch": 30.09131736526946, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1338, "step": 20101 }, { "epoch": 30.092814371257486, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1275, "step": 20102 }, { "epoch": 30.09431137724551, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1245, "step": 20103 }, { "epoch": 30.095808383233532, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1198, "step": 20104 }, { "epoch": 30.097305389221557, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1232, "step": 20105 }, { "epoch": 30.09880239520958, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1316, "step": 20106 }, { "epoch": 30.100299401197606, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1284, "step": 20107 }, { "epoch": 30.101796407185628, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 20108 }, { "epoch": 30.103293413173652, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 20109 }, { "epoch": 30.104790419161677, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 20110 }, { "epoch": 30.106287425149702, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.126, "step": 20111 }, { "epoch": 30.107784431137723, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1266, "step": 20112 }, { "epoch": 30.10928143712575, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1295, "step": 20113 }, { "epoch": 30.110778443113773, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1362, "step": 20114 }, { "epoch": 30.112275449101798, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.131, "step": 20115 }, { "epoch": 30.11377245508982, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.127, "step": 20116 }, { "epoch": 30.115269461077844, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1196, "step": 20117 }, { "epoch": 30.11676646706587, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1322, "step": 20118 }, { "epoch": 30.118263473053894, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1269, "step": 20119 }, { "epoch": 30.119760479041915, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 20120 }, { "epoch": 30.12125748502994, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1283, "step": 20121 }, { "epoch": 30.122754491017965, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1278, "step": 20122 }, { "epoch": 30.12425149700599, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1276, "step": 20123 }, { "epoch": 30.12574850299401, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1237, "step": 20124 }, { "epoch": 30.127245508982035, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 20125 }, { "epoch": 30.12874251497006, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.121, "step": 20126 }, { "epoch": 30.130239520958085, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1244, "step": 20127 }, { "epoch": 30.131736526946106, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1343, "step": 20128 }, { "epoch": 30.13323353293413, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1279, "step": 20129 }, { "epoch": 30.134730538922156, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1203, "step": 20130 }, { "epoch": 30.13622754491018, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.131, "step": 20131 }, { "epoch": 30.137724550898202, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1279, "step": 20132 }, { "epoch": 30.139221556886227, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1283, "step": 20133 }, { "epoch": 30.14071856287425, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1302, "step": 20134 }, { "epoch": 30.142215568862277, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1283, "step": 20135 }, { "epoch": 30.143712574850298, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.126, "step": 20136 }, { "epoch": 30.145209580838323, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1329, "step": 20137 }, { "epoch": 30.146706586826348, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1276, "step": 20138 }, { "epoch": 30.148203592814372, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1266, "step": 20139 }, { "epoch": 30.149700598802394, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1221, "step": 20140 }, { "epoch": 30.15119760479042, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1294, "step": 20141 }, { "epoch": 30.152694610778443, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1212, "step": 20142 }, { "epoch": 30.154191616766468, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1251, "step": 20143 }, { "epoch": 30.15568862275449, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1294, "step": 20144 }, { "epoch": 30.157185628742514, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1265, "step": 20145 }, { "epoch": 30.15868263473054, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1291, "step": 20146 }, { "epoch": 30.160179640718564, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1288, "step": 20147 }, { "epoch": 30.161676646706585, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.132, "step": 20148 }, { "epoch": 30.16317365269461, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1304, "step": 20149 }, { "epoch": 30.164670658682635, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1258, "step": 20150 }, { "epoch": 30.16616766467066, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1252, "step": 20151 }, { "epoch": 30.16766467065868, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1307, "step": 20152 }, { "epoch": 30.169161676646706, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1246, "step": 20153 }, { "epoch": 30.17065868263473, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1273, "step": 20154 }, { "epoch": 30.172155688622755, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.123, "step": 20155 }, { "epoch": 30.17365269461078, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1266, "step": 20156 }, { "epoch": 30.1751497005988, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1244, "step": 20157 }, { "epoch": 30.176646706586826, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 20158 }, { "epoch": 30.17814371257485, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1304, "step": 20159 }, { "epoch": 30.179640718562876, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1242, "step": 20160 }, { "epoch": 30.181137724550897, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1217, "step": 20161 }, { "epoch": 30.182634730538922, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1211, "step": 20162 }, { "epoch": 30.184131736526947, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1304, "step": 20163 }, { "epoch": 30.18562874251497, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1276, "step": 20164 }, { "epoch": 30.187125748502993, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1285, "step": 20165 }, { "epoch": 30.188622754491018, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.127, "step": 20166 }, { "epoch": 30.190119760479043, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1291, "step": 20167 }, { "epoch": 30.191616766467067, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1223, "step": 20168 }, { "epoch": 30.19311377245509, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1213, "step": 20169 }, { "epoch": 30.194610778443113, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1281, "step": 20170 }, { "epoch": 30.19610778443114, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.121, "step": 20171 }, { "epoch": 30.197604790419163, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1242, "step": 20172 }, { "epoch": 30.199101796407184, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1259, "step": 20173 }, { "epoch": 30.20059880239521, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1262, "step": 20174 }, { "epoch": 30.202095808383234, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1282, "step": 20175 }, { "epoch": 30.20359281437126, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1282, "step": 20176 }, { "epoch": 30.20508982035928, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1293, "step": 20177 }, { "epoch": 30.206586826347305, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1259, "step": 20178 }, { "epoch": 30.20808383233533, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1256, "step": 20179 }, { "epoch": 30.209580838323355, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1275, "step": 20180 }, { "epoch": 30.211077844311376, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1309, "step": 20181 }, { "epoch": 30.2125748502994, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1281, "step": 20182 }, { "epoch": 30.214071856287426, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1319, "step": 20183 }, { "epoch": 30.21556886227545, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1264, "step": 20184 }, { "epoch": 30.21706586826347, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1229, "step": 20185 }, { "epoch": 30.218562874251496, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1308, "step": 20186 }, { "epoch": 30.22005988023952, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.125, "step": 20187 }, { "epoch": 30.221556886227546, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 20188 }, { "epoch": 30.223053892215567, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1312, "step": 20189 }, { "epoch": 30.224550898203592, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.126, "step": 20190 }, { "epoch": 30.226047904191617, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1348, "step": 20191 }, { "epoch": 30.227544910179642, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.125, "step": 20192 }, { "epoch": 30.229041916167663, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1256, "step": 20193 }, { "epoch": 30.230538922155688, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1266, "step": 20194 }, { "epoch": 30.232035928143713, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1208, "step": 20195 }, { "epoch": 30.233532934131738, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1307, "step": 20196 }, { "epoch": 30.23502994011976, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1302, "step": 20197 }, { "epoch": 30.236526946107784, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 20198 }, { "epoch": 30.23802395209581, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 20199 }, { "epoch": 30.239520958083833, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 20200 }, { "epoch": 30.241017964071855, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1329, "step": 20201 }, { "epoch": 30.24251497005988, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1345, "step": 20202 }, { "epoch": 30.244011976047904, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1293, "step": 20203 }, { "epoch": 30.24550898203593, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1305, "step": 20204 }, { "epoch": 30.24700598802395, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1337, "step": 20205 }, { "epoch": 30.248502994011975, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1222, "step": 20206 }, { "epoch": 30.25, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1221, "step": 20207 }, { "epoch": 30.251497005988025, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.128, "step": 20208 }, { "epoch": 30.25299401197605, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1197, "step": 20209 }, { "epoch": 30.25449101796407, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.126, "step": 20210 }, { "epoch": 30.255988023952096, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1256, "step": 20211 }, { "epoch": 30.25748502994012, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.127, "step": 20212 }, { "epoch": 30.258982035928145, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1304, "step": 20213 }, { "epoch": 30.260479041916167, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1299, "step": 20214 }, { "epoch": 30.26197604790419, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1235, "step": 20215 }, { "epoch": 30.263473053892216, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1269, "step": 20216 }, { "epoch": 30.26497005988024, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1241, "step": 20217 }, { "epoch": 30.266467065868262, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1276, "step": 20218 }, { "epoch": 30.267964071856287, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1225, "step": 20219 }, { "epoch": 30.269461077844312, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1229, "step": 20220 }, { "epoch": 30.270958083832337, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1261, "step": 20221 }, { "epoch": 30.272455089820358, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.127, "step": 20222 }, { "epoch": 30.273952095808383, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.126, "step": 20223 }, { "epoch": 30.275449101796408, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1323, "step": 20224 }, { "epoch": 30.276946107784433, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1355, "step": 20225 }, { "epoch": 30.278443113772454, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.123, "step": 20226 }, { "epoch": 30.27994011976048, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1411, "step": 20227 }, { "epoch": 30.281437125748504, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1252, "step": 20228 }, { "epoch": 30.28293413173653, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 20229 }, { "epoch": 30.28443113772455, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1255, "step": 20230 }, { "epoch": 30.285928143712574, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1273, "step": 20231 }, { "epoch": 30.2874251497006, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1323, "step": 20232 }, { "epoch": 30.288922155688624, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1164, "step": 20233 }, { "epoch": 30.290419161676645, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1281, "step": 20234 }, { "epoch": 30.29191616766467, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1308, "step": 20235 }, { "epoch": 30.293413173652695, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1251, "step": 20236 }, { "epoch": 30.29491017964072, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1203, "step": 20237 }, { "epoch": 30.29640718562874, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1244, "step": 20238 }, { "epoch": 30.297904191616766, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1233, "step": 20239 }, { "epoch": 30.29940119760479, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1241, "step": 20240 }, { "epoch": 30.300898203592816, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 20241 }, { "epoch": 30.302395209580837, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1275, "step": 20242 }, { "epoch": 30.30389221556886, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1335, "step": 20243 }, { "epoch": 30.305389221556887, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1345, "step": 20244 }, { "epoch": 30.30688622754491, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1255, "step": 20245 }, { "epoch": 30.308383233532933, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1305, "step": 20246 }, { "epoch": 30.309880239520957, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1227, "step": 20247 }, { "epoch": 30.311377245508982, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.128, "step": 20248 }, { "epoch": 30.312874251497007, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 20249 }, { "epoch": 30.31437125748503, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1232, "step": 20250 }, { "epoch": 30.315868263473053, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1325, "step": 20251 }, { "epoch": 30.317365269461078, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1311, "step": 20252 }, { "epoch": 30.318862275449103, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1367, "step": 20253 }, { "epoch": 30.320359281437124, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1233, "step": 20254 }, { "epoch": 30.32185628742515, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 20255 }, { "epoch": 30.323353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.125, "step": 20256 }, { "epoch": 30.3248502994012, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1377, "step": 20257 }, { "epoch": 30.32634730538922, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1223, "step": 20258 }, { "epoch": 30.327844311377245, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 20259 }, { "epoch": 30.32934131736527, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1269, "step": 20260 }, { "epoch": 30.330838323353294, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1225, "step": 20261 }, { "epoch": 30.33233532934132, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1309, "step": 20262 }, { "epoch": 30.33383233532934, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1223, "step": 20263 }, { "epoch": 30.335329341317365, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1289, "step": 20264 }, { "epoch": 30.33682634730539, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1244, "step": 20265 }, { "epoch": 30.338323353293415, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1267, "step": 20266 }, { "epoch": 30.339820359281436, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1239, "step": 20267 }, { "epoch": 30.34131736526946, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1272, "step": 20268 }, { "epoch": 30.342814371257486, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.132, "step": 20269 }, { "epoch": 30.34431137724551, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.132, "step": 20270 }, { "epoch": 30.345808383233532, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1269, "step": 20271 }, { "epoch": 30.347305389221557, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1272, "step": 20272 }, { "epoch": 30.34880239520958, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1301, "step": 20273 }, { "epoch": 30.350299401197606, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.129, "step": 20274 }, { "epoch": 30.351796407185628, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1272, "step": 20275 }, { "epoch": 30.353293413173652, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 20276 }, { "epoch": 30.354790419161677, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.123, "step": 20277 }, { "epoch": 30.356287425149702, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1257, "step": 20278 }, { "epoch": 30.357784431137723, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1308, "step": 20279 }, { "epoch": 30.35928143712575, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1242, "step": 20280 }, { "epoch": 30.360778443113773, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1325, "step": 20281 }, { "epoch": 30.362275449101798, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1282, "step": 20282 }, { "epoch": 30.36377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1249, "step": 20283 }, { "epoch": 30.365269461077844, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1284, "step": 20284 }, { "epoch": 30.36676646706587, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1287, "step": 20285 }, { "epoch": 30.368263473053894, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1315, "step": 20286 }, { "epoch": 30.369760479041915, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1317, "step": 20287 }, { "epoch": 30.37125748502994, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1241, "step": 20288 }, { "epoch": 30.372754491017965, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1334, "step": 20289 }, { "epoch": 30.37425149700599, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.129, "step": 20290 }, { "epoch": 30.37574850299401, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1243, "step": 20291 }, { "epoch": 30.377245508982035, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1303, "step": 20292 }, { "epoch": 30.37874251497006, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1311, "step": 20293 }, { "epoch": 30.380239520958085, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1315, "step": 20294 }, { "epoch": 30.381736526946106, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1263, "step": 20295 }, { "epoch": 30.38323353293413, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1359, "step": 20296 }, { "epoch": 30.384730538922156, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1276, "step": 20297 }, { "epoch": 30.38622754491018, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1177, "step": 20298 }, { "epoch": 30.387724550898202, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1227, "step": 20299 }, { "epoch": 30.389221556886227, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1349, "step": 20300 }, { "epoch": 30.39071856287425, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 20301 }, { "epoch": 30.392215568862277, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1299, "step": 20302 }, { "epoch": 30.393712574850298, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1224, "step": 20303 }, { "epoch": 30.395209580838323, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1258, "step": 20304 }, { "epoch": 30.396706586826348, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1235, "step": 20305 }, { "epoch": 30.398203592814372, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1252, "step": 20306 }, { "epoch": 30.399700598802394, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1279, "step": 20307 }, { "epoch": 30.40119760479042, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1276, "step": 20308 }, { "epoch": 30.402694610778443, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1263, "step": 20309 }, { "epoch": 30.404191616766468, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1326, "step": 20310 }, { "epoch": 30.40568862275449, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1314, "step": 20311 }, { "epoch": 30.407185628742514, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1286, "step": 20312 }, { "epoch": 30.40868263473054, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1321, "step": 20313 }, { "epoch": 30.410179640718564, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 20314 }, { "epoch": 30.411676646706585, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.129, "step": 20315 }, { "epoch": 30.41317365269461, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1305, "step": 20316 }, { "epoch": 30.414670658682635, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1304, "step": 20317 }, { "epoch": 30.41616766467066, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1289, "step": 20318 }, { "epoch": 30.41766467065868, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1313, "step": 20319 }, { "epoch": 30.419161676646706, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1261, "step": 20320 }, { "epoch": 30.42065868263473, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1229, "step": 20321 }, { "epoch": 30.422155688622755, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1227, "step": 20322 }, { "epoch": 30.42365269461078, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1296, "step": 20323 }, { "epoch": 30.4251497005988, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1226, "step": 20324 }, { "epoch": 30.426646706586826, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1247, "step": 20325 }, { "epoch": 30.42814371257485, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1217, "step": 20326 }, { "epoch": 30.429640718562876, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 20327 }, { "epoch": 30.431137724550897, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1299, "step": 20328 }, { "epoch": 30.432634730538922, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1254, "step": 20329 }, { "epoch": 30.434131736526947, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1246, "step": 20330 }, { "epoch": 30.43562874251497, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1323, "step": 20331 }, { "epoch": 30.437125748502993, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1206, "step": 20332 }, { "epoch": 30.438622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1309, "step": 20333 }, { "epoch": 30.440119760479043, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1237, "step": 20334 }, { "epoch": 30.441616766467067, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.124, "step": 20335 }, { "epoch": 30.44311377245509, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1196, "step": 20336 }, { "epoch": 30.444610778443113, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 20337 }, { "epoch": 30.44610778443114, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1224, "step": 20338 }, { "epoch": 30.447604790419163, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 20339 }, { "epoch": 30.449101796407184, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1266, "step": 20340 }, { "epoch": 30.45059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1265, "step": 20341 }, { "epoch": 30.452095808383234, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1191, "step": 20342 }, { "epoch": 30.45359281437126, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1219, "step": 20343 }, { "epoch": 30.45508982035928, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1259, "step": 20344 }, { "epoch": 30.456586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1282, "step": 20345 }, { "epoch": 30.45808383233533, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1229, "step": 20346 }, { "epoch": 30.459580838323355, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1282, "step": 20347 }, { "epoch": 30.461077844311376, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1222, "step": 20348 }, { "epoch": 30.4625748502994, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 20349 }, { "epoch": 30.464071856287426, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1265, "step": 20350 }, { "epoch": 30.46556886227545, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1246, "step": 20351 }, { "epoch": 30.46706586826347, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1305, "step": 20352 }, { "epoch": 30.468562874251496, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1257, "step": 20353 }, { "epoch": 30.47005988023952, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 20354 }, { "epoch": 30.471556886227546, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.128, "step": 20355 }, { "epoch": 30.473053892215567, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.13, "step": 20356 }, { "epoch": 30.474550898203592, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1215, "step": 20357 }, { "epoch": 30.476047904191617, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1252, "step": 20358 }, { "epoch": 30.477544910179642, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 20359 }, { "epoch": 30.479041916167663, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1294, "step": 20360 }, { "epoch": 30.480538922155688, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1324, "step": 20361 }, { "epoch": 30.482035928143713, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1284, "step": 20362 }, { "epoch": 30.483532934131738, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1307, "step": 20363 }, { "epoch": 30.48502994011976, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1257, "step": 20364 }, { "epoch": 30.486526946107784, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1244, "step": 20365 }, { "epoch": 30.48802395209581, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1255, "step": 20366 }, { "epoch": 30.489520958083833, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1227, "step": 20367 }, { "epoch": 30.491017964071855, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.13, "step": 20368 }, { "epoch": 30.49251497005988, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1222, "step": 20369 }, { "epoch": 30.494011976047904, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1296, "step": 20370 }, { "epoch": 30.49550898203593, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1235, "step": 20371 }, { "epoch": 30.49700598802395, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1151, "step": 20372 }, { "epoch": 30.498502994011975, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1251, "step": 20373 }, { "epoch": 30.5, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1253, "step": 20374 }, { "epoch": 30.501497005988025, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1209, "step": 20375 }, { "epoch": 30.50299401197605, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.123, "step": 20376 }, { "epoch": 30.50449101796407, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1266, "step": 20377 }, { "epoch": 30.505988023952096, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 20378 }, { "epoch": 30.50748502994012, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1256, "step": 20379 }, { "epoch": 30.508982035928145, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1295, "step": 20380 }, { "epoch": 30.510479041916167, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1309, "step": 20381 }, { "epoch": 30.51197604790419, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1244, "step": 20382 }, { "epoch": 30.513473053892216, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1268, "step": 20383 }, { "epoch": 30.51497005988024, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 20384 }, { "epoch": 30.516467065868262, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1325, "step": 20385 }, { "epoch": 30.517964071856287, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 20386 }, { "epoch": 30.519461077844312, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1246, "step": 20387 }, { "epoch": 30.520958083832337, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1228, "step": 20388 }, { "epoch": 30.522455089820358, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1346, "step": 20389 }, { "epoch": 30.523952095808383, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1275, "step": 20390 }, { "epoch": 30.525449101796408, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1315, "step": 20391 }, { "epoch": 30.526946107784433, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1208, "step": 20392 }, { "epoch": 30.528443113772454, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1248, "step": 20393 }, { "epoch": 30.52994011976048, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1273, "step": 20394 }, { "epoch": 30.531437125748504, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1284, "step": 20395 }, { "epoch": 30.53293413173653, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1266, "step": 20396 }, { "epoch": 30.53443113772455, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1239, "step": 20397 }, { "epoch": 30.535928143712574, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1271, "step": 20398 }, { "epoch": 30.5374251497006, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1273, "step": 20399 }, { "epoch": 30.538922155688624, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1244, "step": 20400 }, { "epoch": 30.540419161676645, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1239, "step": 20401 }, { "epoch": 30.54191616766467, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1267, "step": 20402 }, { "epoch": 30.543413173652695, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.121, "step": 20403 }, { "epoch": 30.54491017964072, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1292, "step": 20404 }, { "epoch": 30.54640718562874, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1249, "step": 20405 }, { "epoch": 30.547904191616766, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.128, "step": 20406 }, { "epoch": 30.54940119760479, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1272, "step": 20407 }, { "epoch": 30.550898203592816, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1325, "step": 20408 }, { "epoch": 30.552395209580837, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1228, "step": 20409 }, { "epoch": 30.55389221556886, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1211, "step": 20410 }, { "epoch": 30.555389221556887, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1241, "step": 20411 }, { "epoch": 30.55688622754491, "grad_norm": 0.048095703125, "learning_rate": 0.0008, "loss": 1.1251, "step": 20412 }, { "epoch": 30.558383233532933, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1264, "step": 20413 }, { "epoch": 30.559880239520957, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1291, "step": 20414 }, { "epoch": 30.561377245508982, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1336, "step": 20415 }, { "epoch": 30.562874251497007, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1254, "step": 20416 }, { "epoch": 30.56437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1235, "step": 20417 }, { "epoch": 30.565868263473053, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1281, "step": 20418 }, { "epoch": 30.567365269461078, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1273, "step": 20419 }, { "epoch": 30.568862275449103, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1223, "step": 20420 }, { "epoch": 30.570359281437124, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.121, "step": 20421 }, { "epoch": 30.57185628742515, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1307, "step": 20422 }, { "epoch": 30.573353293413174, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1254, "step": 20423 }, { "epoch": 30.5748502994012, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1261, "step": 20424 }, { "epoch": 30.57634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1265, "step": 20425 }, { "epoch": 30.577844311377245, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1257, "step": 20426 }, { "epoch": 30.57934131736527, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1313, "step": 20427 }, { "epoch": 30.580838323353294, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1208, "step": 20428 }, { "epoch": 30.58233532934132, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1251, "step": 20429 }, { "epoch": 30.58383233532934, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1261, "step": 20430 }, { "epoch": 30.585329341317365, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1279, "step": 20431 }, { "epoch": 30.58682634730539, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1292, "step": 20432 }, { "epoch": 30.58832335329341, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1251, "step": 20433 }, { "epoch": 30.589820359281436, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 20434 }, { "epoch": 30.59131736526946, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1314, "step": 20435 }, { "epoch": 30.592814371257486, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1353, "step": 20436 }, { "epoch": 30.59431137724551, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.124, "step": 20437 }, { "epoch": 30.595808383233532, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1273, "step": 20438 }, { "epoch": 30.597305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1306, "step": 20439 }, { "epoch": 30.59880239520958, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1364, "step": 20440 }, { "epoch": 30.600299401197606, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1278, "step": 20441 }, { "epoch": 30.601796407185628, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1312, "step": 20442 }, { "epoch": 30.603293413173652, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1306, "step": 20443 }, { "epoch": 30.604790419161677, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1292, "step": 20444 }, { "epoch": 30.606287425149702, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1265, "step": 20445 }, { "epoch": 30.607784431137723, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1266, "step": 20446 }, { "epoch": 30.60928143712575, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1295, "step": 20447 }, { "epoch": 30.610778443113773, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1305, "step": 20448 }, { "epoch": 30.612275449101798, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.128, "step": 20449 }, { "epoch": 30.61377245508982, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1259, "step": 20450 }, { "epoch": 30.615269461077844, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1272, "step": 20451 }, { "epoch": 30.61676646706587, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.131, "step": 20452 }, { "epoch": 30.618263473053894, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 20453 }, { "epoch": 30.619760479041915, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1256, "step": 20454 }, { "epoch": 30.62125748502994, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1275, "step": 20455 }, { "epoch": 30.622754491017965, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1309, "step": 20456 }, { "epoch": 30.62425149700599, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1272, "step": 20457 }, { "epoch": 30.62574850299401, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1303, "step": 20458 }, { "epoch": 30.627245508982035, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.126, "step": 20459 }, { "epoch": 30.62874251497006, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 20460 }, { "epoch": 30.630239520958085, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1243, "step": 20461 }, { "epoch": 30.631736526946106, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1274, "step": 20462 }, { "epoch": 30.63323353293413, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1346, "step": 20463 }, { "epoch": 30.634730538922156, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1338, "step": 20464 }, { "epoch": 30.63622754491018, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1285, "step": 20465 }, { "epoch": 30.637724550898202, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.126, "step": 20466 }, { "epoch": 30.639221556886227, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1275, "step": 20467 }, { "epoch": 30.64071856287425, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1282, "step": 20468 }, { "epoch": 30.642215568862277, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1225, "step": 20469 }, { "epoch": 30.643712574850298, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1315, "step": 20470 }, { "epoch": 30.645209580838323, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1231, "step": 20471 }, { "epoch": 30.646706586826348, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1272, "step": 20472 }, { "epoch": 30.648203592814372, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1242, "step": 20473 }, { "epoch": 30.649700598802394, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.126, "step": 20474 }, { "epoch": 30.65119760479042, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1286, "step": 20475 }, { "epoch": 30.652694610778443, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1256, "step": 20476 }, { "epoch": 30.654191616766468, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1254, "step": 20477 }, { "epoch": 30.65568862275449, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1301, "step": 20478 }, { "epoch": 30.657185628742514, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1266, "step": 20479 }, { "epoch": 30.65868263473054, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1286, "step": 20480 }, { "epoch": 30.660179640718564, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1339, "step": 20481 }, { "epoch": 30.66167664670659, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 20482 }, { "epoch": 30.66317365269461, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 20483 }, { "epoch": 30.664670658682635, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1248, "step": 20484 }, { "epoch": 30.66616766467066, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1311, "step": 20485 }, { "epoch": 30.66766467065868, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1314, "step": 20486 }, { "epoch": 30.669161676646706, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1287, "step": 20487 }, { "epoch": 30.67065868263473, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1274, "step": 20488 }, { "epoch": 30.672155688622755, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 20489 }, { "epoch": 30.67365269461078, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1229, "step": 20490 }, { "epoch": 30.6751497005988, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1262, "step": 20491 }, { "epoch": 30.676646706586826, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1263, "step": 20492 }, { "epoch": 30.67814371257485, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 20493 }, { "epoch": 30.679640718562876, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1321, "step": 20494 }, { "epoch": 30.681137724550897, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1237, "step": 20495 }, { "epoch": 30.682634730538922, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1308, "step": 20496 }, { "epoch": 30.684131736526947, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1221, "step": 20497 }, { "epoch": 30.68562874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1198, "step": 20498 }, { "epoch": 30.687125748502993, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1292, "step": 20499 }, { "epoch": 30.688622754491018, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1249, "step": 20500 }, { "epoch": 30.690119760479043, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1235, "step": 20501 }, { "epoch": 30.691616766467067, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1241, "step": 20502 }, { "epoch": 30.69311377245509, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1262, "step": 20503 }, { "epoch": 30.694610778443113, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1306, "step": 20504 }, { "epoch": 30.69610778443114, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.121, "step": 20505 }, { "epoch": 30.697604790419163, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1225, "step": 20506 }, { "epoch": 30.699101796407184, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1213, "step": 20507 }, { "epoch": 30.70059880239521, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1296, "step": 20508 }, { "epoch": 30.702095808383234, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1266, "step": 20509 }, { "epoch": 30.70359281437126, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1269, "step": 20510 }, { "epoch": 30.70508982035928, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1334, "step": 20511 }, { "epoch": 30.706586826347305, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1275, "step": 20512 }, { "epoch": 30.70808383233533, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1217, "step": 20513 }, { "epoch": 30.709580838323355, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1208, "step": 20514 }, { "epoch": 30.711077844311376, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1314, "step": 20515 }, { "epoch": 30.7125748502994, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1294, "step": 20516 }, { "epoch": 30.714071856287426, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1205, "step": 20517 }, { "epoch": 30.71556886227545, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 20518 }, { "epoch": 30.71706586826347, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1232, "step": 20519 }, { "epoch": 30.718562874251496, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1238, "step": 20520 }, { "epoch": 30.72005988023952, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1245, "step": 20521 }, { "epoch": 30.721556886227546, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1325, "step": 20522 }, { "epoch": 30.723053892215567, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1266, "step": 20523 }, { "epoch": 30.724550898203592, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1254, "step": 20524 }, { "epoch": 30.726047904191617, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1291, "step": 20525 }, { "epoch": 30.727544910179642, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1232, "step": 20526 }, { "epoch": 30.729041916167663, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 20527 }, { "epoch": 30.730538922155688, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1274, "step": 20528 }, { "epoch": 30.732035928143713, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 20529 }, { "epoch": 30.733532934131738, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.125, "step": 20530 }, { "epoch": 30.73502994011976, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1331, "step": 20531 }, { "epoch": 30.736526946107784, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1421, "step": 20532 }, { "epoch": 30.73802395209581, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1352, "step": 20533 }, { "epoch": 30.739520958083833, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1223, "step": 20534 }, { "epoch": 30.741017964071855, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1253, "step": 20535 }, { "epoch": 30.74251497005988, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1282, "step": 20536 }, { "epoch": 30.744011976047904, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.122, "step": 20537 }, { "epoch": 30.74550898203593, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1321, "step": 20538 }, { "epoch": 30.74700598802395, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1273, "step": 20539 }, { "epoch": 30.748502994011975, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1266, "step": 20540 }, { "epoch": 30.75, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.134, "step": 20541 }, { "epoch": 30.751497005988025, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1294, "step": 20542 }, { "epoch": 30.75299401197605, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1311, "step": 20543 }, { "epoch": 30.75449101796407, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1227, "step": 20544 }, { "epoch": 30.755988023952096, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.123, "step": 20545 }, { "epoch": 30.75748502994012, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1285, "step": 20546 }, { "epoch": 30.758982035928145, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1286, "step": 20547 }, { "epoch": 30.760479041916167, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1326, "step": 20548 }, { "epoch": 30.76197604790419, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1289, "step": 20549 }, { "epoch": 30.763473053892216, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.128, "step": 20550 }, { "epoch": 30.76497005988024, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1222, "step": 20551 }, { "epoch": 30.766467065868262, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1332, "step": 20552 }, { "epoch": 30.767964071856287, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1289, "step": 20553 }, { "epoch": 30.769461077844312, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1376, "step": 20554 }, { "epoch": 30.770958083832337, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1319, "step": 20555 }, { "epoch": 30.772455089820358, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1249, "step": 20556 }, { "epoch": 30.773952095808383, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.127, "step": 20557 }, { "epoch": 30.775449101796408, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1246, "step": 20558 }, { "epoch": 30.776946107784433, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.13, "step": 20559 }, { "epoch": 30.778443113772454, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1267, "step": 20560 }, { "epoch": 30.77994011976048, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1237, "step": 20561 }, { "epoch": 30.781437125748504, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1342, "step": 20562 }, { "epoch": 30.78293413173653, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1244, "step": 20563 }, { "epoch": 30.78443113772455, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1145, "step": 20564 }, { "epoch": 30.785928143712574, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1322, "step": 20565 }, { "epoch": 30.7874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1258, "step": 20566 }, { "epoch": 30.788922155688624, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1243, "step": 20567 }, { "epoch": 30.790419161676645, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1196, "step": 20568 }, { "epoch": 30.79191616766467, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1288, "step": 20569 }, { "epoch": 30.793413173652695, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 20570 }, { "epoch": 30.79491017964072, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.126, "step": 20571 }, { "epoch": 30.79640718562874, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1312, "step": 20572 }, { "epoch": 30.797904191616766, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1259, "step": 20573 }, { "epoch": 30.79940119760479, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 20574 }, { "epoch": 30.800898203592816, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1293, "step": 20575 }, { "epoch": 30.802395209580837, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1266, "step": 20576 }, { "epoch": 30.80389221556886, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1274, "step": 20577 }, { "epoch": 30.805389221556887, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1304, "step": 20578 }, { "epoch": 30.80688622754491, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1309, "step": 20579 }, { "epoch": 30.808383233532933, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1282, "step": 20580 }, { "epoch": 30.809880239520957, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 20581 }, { "epoch": 30.811377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1295, "step": 20582 }, { "epoch": 30.812874251497007, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1238, "step": 20583 }, { "epoch": 30.81437125748503, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1324, "step": 20584 }, { "epoch": 30.815868263473053, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1284, "step": 20585 }, { "epoch": 30.817365269461078, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1287, "step": 20586 }, { "epoch": 30.818862275449103, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1259, "step": 20587 }, { "epoch": 30.820359281437124, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1356, "step": 20588 }, { "epoch": 30.82185628742515, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1262, "step": 20589 }, { "epoch": 30.823353293413174, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1292, "step": 20590 }, { "epoch": 30.8248502994012, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1286, "step": 20591 }, { "epoch": 30.82634730538922, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.128, "step": 20592 }, { "epoch": 30.827844311377245, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1286, "step": 20593 }, { "epoch": 30.82934131736527, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 20594 }, { "epoch": 30.830838323353294, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1257, "step": 20595 }, { "epoch": 30.83233532934132, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1312, "step": 20596 }, { "epoch": 30.83383233532934, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1295, "step": 20597 }, { "epoch": 30.835329341317365, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 20598 }, { "epoch": 30.83682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1289, "step": 20599 }, { "epoch": 30.83832335329341, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1336, "step": 20600 }, { "epoch": 30.839820359281436, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1327, "step": 20601 }, { "epoch": 30.84131736526946, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.127, "step": 20602 }, { "epoch": 30.842814371257486, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1353, "step": 20603 }, { "epoch": 30.84431137724551, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1238, "step": 20604 }, { "epoch": 30.845808383233532, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.13, "step": 20605 }, { "epoch": 30.847305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1277, "step": 20606 }, { "epoch": 30.84880239520958, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1288, "step": 20607 }, { "epoch": 30.850299401197606, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.131, "step": 20608 }, { "epoch": 30.851796407185628, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 20609 }, { "epoch": 30.853293413173652, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.13, "step": 20610 }, { "epoch": 30.854790419161677, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1263, "step": 20611 }, { "epoch": 30.856287425149702, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1328, "step": 20612 }, { "epoch": 30.857784431137723, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1254, "step": 20613 }, { "epoch": 30.85928143712575, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1305, "step": 20614 }, { "epoch": 30.860778443113773, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1316, "step": 20615 }, { "epoch": 30.862275449101798, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1256, "step": 20616 }, { "epoch": 30.86377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1251, "step": 20617 }, { "epoch": 30.865269461077844, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1244, "step": 20618 }, { "epoch": 30.86676646706587, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1285, "step": 20619 }, { "epoch": 30.868263473053894, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1298, "step": 20620 }, { "epoch": 30.869760479041915, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1322, "step": 20621 }, { "epoch": 30.87125748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1282, "step": 20622 }, { "epoch": 30.872754491017965, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1229, "step": 20623 }, { "epoch": 30.87425149700599, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1203, "step": 20624 }, { "epoch": 30.87574850299401, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1358, "step": 20625 }, { "epoch": 30.877245508982035, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1257, "step": 20626 }, { "epoch": 30.87874251497006, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1264, "step": 20627 }, { "epoch": 30.880239520958085, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1269, "step": 20628 }, { "epoch": 30.881736526946106, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.13, "step": 20629 }, { "epoch": 30.88323353293413, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 20630 }, { "epoch": 30.884730538922156, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1273, "step": 20631 }, { "epoch": 30.88622754491018, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.12, "step": 20632 }, { "epoch": 30.887724550898202, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1268, "step": 20633 }, { "epoch": 30.889221556886227, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1324, "step": 20634 }, { "epoch": 30.89071856287425, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1276, "step": 20635 }, { "epoch": 30.892215568862277, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.136, "step": 20636 }, { "epoch": 30.893712574850298, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 20637 }, { "epoch": 30.895209580838323, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1318, "step": 20638 }, { "epoch": 30.896706586826348, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1307, "step": 20639 }, { "epoch": 30.898203592814372, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1271, "step": 20640 }, { "epoch": 30.899700598802394, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.128, "step": 20641 }, { "epoch": 30.90119760479042, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.129, "step": 20642 }, { "epoch": 30.902694610778443, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1314, "step": 20643 }, { "epoch": 30.904191616766468, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1225, "step": 20644 }, { "epoch": 30.90568862275449, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1242, "step": 20645 }, { "epoch": 30.907185628742514, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1214, "step": 20646 }, { "epoch": 30.90868263473054, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1203, "step": 20647 }, { "epoch": 30.910179640718564, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1327, "step": 20648 }, { "epoch": 30.91167664670659, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1263, "step": 20649 }, { "epoch": 30.91317365269461, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1252, "step": 20650 }, { "epoch": 30.914670658682635, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1365, "step": 20651 }, { "epoch": 30.91616766467066, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1256, "step": 20652 }, { "epoch": 30.91766467065868, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1325, "step": 20653 }, { "epoch": 30.919161676646706, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1284, "step": 20654 }, { "epoch": 30.92065868263473, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1314, "step": 20655 }, { "epoch": 30.922155688622755, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1306, "step": 20656 }, { "epoch": 30.92365269461078, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1276, "step": 20657 }, { "epoch": 30.9251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1202, "step": 20658 }, { "epoch": 30.926646706586826, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.131, "step": 20659 }, { "epoch": 30.92814371257485, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1158, "step": 20660 }, { "epoch": 30.929640718562876, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1227, "step": 20661 }, { "epoch": 30.931137724550897, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1296, "step": 20662 }, { "epoch": 30.932634730538922, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1277, "step": 20663 }, { "epoch": 30.934131736526947, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1256, "step": 20664 }, { "epoch": 30.93562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1324, "step": 20665 }, { "epoch": 30.937125748502993, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1281, "step": 20666 }, { "epoch": 30.938622754491018, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1305, "step": 20667 }, { "epoch": 30.940119760479043, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.126, "step": 20668 }, { "epoch": 30.941616766467067, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1237, "step": 20669 }, { "epoch": 30.94311377245509, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1265, "step": 20670 }, { "epoch": 30.944610778443113, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1346, "step": 20671 }, { "epoch": 30.94610778443114, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1295, "step": 20672 }, { "epoch": 30.947604790419163, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1299, "step": 20673 }, { "epoch": 30.949101796407184, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 20674 }, { "epoch": 30.95059880239521, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1223, "step": 20675 }, { "epoch": 30.952095808383234, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.133, "step": 20676 }, { "epoch": 30.95359281437126, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1212, "step": 20677 }, { "epoch": 30.95508982035928, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1294, "step": 20678 }, { "epoch": 30.956586826347305, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1258, "step": 20679 }, { "epoch": 30.95808383233533, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1302, "step": 20680 }, { "epoch": 30.959580838323355, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.129, "step": 20681 }, { "epoch": 30.961077844311376, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1277, "step": 20682 }, { "epoch": 30.9625748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1243, "step": 20683 }, { "epoch": 30.964071856287426, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1253, "step": 20684 }, { "epoch": 30.96556886227545, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1279, "step": 20685 }, { "epoch": 30.96706586826347, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1203, "step": 20686 }, { "epoch": 30.968562874251496, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1349, "step": 20687 }, { "epoch": 30.97005988023952, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1227, "step": 20688 }, { "epoch": 30.971556886227546, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1268, "step": 20689 }, { "epoch": 30.973053892215567, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1355, "step": 20690 }, { "epoch": 30.974550898203592, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1258, "step": 20691 }, { "epoch": 30.976047904191617, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1336, "step": 20692 }, { "epoch": 30.977544910179642, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1268, "step": 20693 }, { "epoch": 30.979041916167663, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1261, "step": 20694 }, { "epoch": 30.980538922155688, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.129, "step": 20695 }, { "epoch": 30.982035928143713, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1348, "step": 20696 }, { "epoch": 30.983532934131738, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1241, "step": 20697 }, { "epoch": 30.98502994011976, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1287, "step": 20698 }, { "epoch": 30.986526946107784, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1286, "step": 20699 }, { "epoch": 30.98802395209581, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1306, "step": 20700 }, { "epoch": 30.989520958083833, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1318, "step": 20701 }, { "epoch": 30.991017964071855, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1249, "step": 20702 }, { "epoch": 30.99251497005988, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1253, "step": 20703 }, { "epoch": 30.994011976047904, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1335, "step": 20704 }, { "epoch": 30.99550898203593, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1249, "step": 20705 }, { "epoch": 30.99700598802395, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1244, "step": 20706 }, { "epoch": 30.998502994011975, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1282, "step": 20707 }, { "epoch": 31.0, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 20708 }, { "epoch": 31.001497005988025, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1284, "step": 20709 }, { "epoch": 31.00299401197605, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1253, "step": 20710 }, { "epoch": 31.00449101796407, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1177, "step": 20711 }, { "epoch": 31.005988023952096, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1221, "step": 20712 }, { "epoch": 31.00748502994012, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1196, "step": 20713 }, { "epoch": 31.008982035928145, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1333, "step": 20714 }, { "epoch": 31.010479041916167, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1251, "step": 20715 }, { "epoch": 31.01197604790419, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1269, "step": 20716 }, { "epoch": 31.013473053892216, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1225, "step": 20717 }, { "epoch": 31.01497005988024, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1294, "step": 20718 }, { "epoch": 31.016467065868262, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1272, "step": 20719 }, { "epoch": 31.017964071856287, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1264, "step": 20720 }, { "epoch": 31.019461077844312, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1307, "step": 20721 }, { "epoch": 31.020958083832337, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1238, "step": 20722 }, { "epoch": 31.022455089820358, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.129, "step": 20723 }, { "epoch": 31.023952095808383, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1276, "step": 20724 }, { "epoch": 31.025449101796408, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.12, "step": 20725 }, { "epoch": 31.026946107784433, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1279, "step": 20726 }, { "epoch": 31.028443113772454, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.121, "step": 20727 }, { "epoch": 31.02994011976048, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.126, "step": 20728 }, { "epoch": 31.031437125748504, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1254, "step": 20729 }, { "epoch": 31.03293413173653, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1285, "step": 20730 }, { "epoch": 31.03443113772455, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1242, "step": 20731 }, { "epoch": 31.035928143712574, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1229, "step": 20732 }, { "epoch": 31.0374251497006, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1277, "step": 20733 }, { "epoch": 31.038922155688624, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1328, "step": 20734 }, { "epoch": 31.040419161676645, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1251, "step": 20735 }, { "epoch": 31.04191616766467, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1186, "step": 20736 }, { "epoch": 31.043413173652695, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1252, "step": 20737 }, { "epoch": 31.04491017964072, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1232, "step": 20738 }, { "epoch": 31.04640718562874, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1275, "step": 20739 }, { "epoch": 31.047904191616766, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1256, "step": 20740 }, { "epoch": 31.04940119760479, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.128, "step": 20741 }, { "epoch": 31.050898203592816, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.126, "step": 20742 }, { "epoch": 31.052395209580837, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1329, "step": 20743 }, { "epoch": 31.05389221556886, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1259, "step": 20744 }, { "epoch": 31.055389221556887, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1292, "step": 20745 }, { "epoch": 31.05688622754491, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 20746 }, { "epoch": 31.058383233532933, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 20747 }, { "epoch": 31.059880239520957, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1285, "step": 20748 }, { "epoch": 31.061377245508982, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1241, "step": 20749 }, { "epoch": 31.062874251497007, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1289, "step": 20750 }, { "epoch": 31.06437125748503, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1234, "step": 20751 }, { "epoch": 31.065868263473053, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1354, "step": 20752 }, { "epoch": 31.067365269461078, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1301, "step": 20753 }, { "epoch": 31.068862275449103, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.126, "step": 20754 }, { "epoch": 31.070359281437124, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1234, "step": 20755 }, { "epoch": 31.07185628742515, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 20756 }, { "epoch": 31.073353293413174, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1306, "step": 20757 }, { "epoch": 31.0748502994012, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1274, "step": 20758 }, { "epoch": 31.07634730538922, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1192, "step": 20759 }, { "epoch": 31.077844311377245, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1254, "step": 20760 }, { "epoch": 31.07934131736527, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1276, "step": 20761 }, { "epoch": 31.080838323353294, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1303, "step": 20762 }, { "epoch": 31.082335329341316, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1287, "step": 20763 }, { "epoch": 31.08383233532934, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1227, "step": 20764 }, { "epoch": 31.085329341317365, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1271, "step": 20765 }, { "epoch": 31.08682634730539, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1242, "step": 20766 }, { "epoch": 31.088323353293415, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1226, "step": 20767 }, { "epoch": 31.089820359281436, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1252, "step": 20768 }, { "epoch": 31.09131736526946, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1201, "step": 20769 }, { "epoch": 31.092814371257486, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1223, "step": 20770 }, { "epoch": 31.09431137724551, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1242, "step": 20771 }, { "epoch": 31.095808383233532, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1288, "step": 20772 }, { "epoch": 31.097305389221557, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1287, "step": 20773 }, { "epoch": 31.09880239520958, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1231, "step": 20774 }, { "epoch": 31.100299401197606, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1315, "step": 20775 }, { "epoch": 31.101796407185628, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1154, "step": 20776 }, { "epoch": 31.103293413173652, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1324, "step": 20777 }, { "epoch": 31.104790419161677, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1284, "step": 20778 }, { "epoch": 31.106287425149702, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1223, "step": 20779 }, { "epoch": 31.107784431137723, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1211, "step": 20780 }, { "epoch": 31.10928143712575, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1262, "step": 20781 }, { "epoch": 31.110778443113773, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1205, "step": 20782 }, { "epoch": 31.112275449101798, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1298, "step": 20783 }, { "epoch": 31.11377245508982, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1217, "step": 20784 }, { "epoch": 31.115269461077844, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1321, "step": 20785 }, { "epoch": 31.11676646706587, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1238, "step": 20786 }, { "epoch": 31.118263473053894, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1323, "step": 20787 }, { "epoch": 31.119760479041915, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1226, "step": 20788 }, { "epoch": 31.12125748502994, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1282, "step": 20789 }, { "epoch": 31.122754491017965, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.129, "step": 20790 }, { "epoch": 31.12425149700599, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1311, "step": 20791 }, { "epoch": 31.12574850299401, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1291, "step": 20792 }, { "epoch": 31.127245508982035, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.127, "step": 20793 }, { "epoch": 31.12874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.122, "step": 20794 }, { "epoch": 31.130239520958085, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1216, "step": 20795 }, { "epoch": 31.131736526946106, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1288, "step": 20796 }, { "epoch": 31.13323353293413, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1252, "step": 20797 }, { "epoch": 31.134730538922156, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1304, "step": 20798 }, { "epoch": 31.13622754491018, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 20799 }, { "epoch": 31.137724550898202, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1304, "step": 20800 }, { "epoch": 31.139221556886227, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1326, "step": 20801 }, { "epoch": 31.14071856287425, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1249, "step": 20802 }, { "epoch": 31.142215568862277, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1208, "step": 20803 }, { "epoch": 31.143712574850298, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1282, "step": 20804 }, { "epoch": 31.145209580838323, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1217, "step": 20805 }, { "epoch": 31.146706586826348, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1238, "step": 20806 }, { "epoch": 31.148203592814372, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1284, "step": 20807 }, { "epoch": 31.149700598802394, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1216, "step": 20808 }, { "epoch": 31.15119760479042, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1296, "step": 20809 }, { "epoch": 31.152694610778443, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1299, "step": 20810 }, { "epoch": 31.154191616766468, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1234, "step": 20811 }, { "epoch": 31.15568862275449, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 20812 }, { "epoch": 31.157185628742514, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1305, "step": 20813 }, { "epoch": 31.15868263473054, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1244, "step": 20814 }, { "epoch": 31.160179640718564, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1281, "step": 20815 }, { "epoch": 31.161676646706585, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1312, "step": 20816 }, { "epoch": 31.16317365269461, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1273, "step": 20817 }, { "epoch": 31.164670658682635, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1198, "step": 20818 }, { "epoch": 31.16616766467066, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1256, "step": 20819 }, { "epoch": 31.16766467065868, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1281, "step": 20820 }, { "epoch": 31.169161676646706, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1201, "step": 20821 }, { "epoch": 31.17065868263473, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1316, "step": 20822 }, { "epoch": 31.172155688622755, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1235, "step": 20823 }, { "epoch": 31.17365269461078, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1251, "step": 20824 }, { "epoch": 31.1751497005988, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1251, "step": 20825 }, { "epoch": 31.176646706586826, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 20826 }, { "epoch": 31.17814371257485, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1377, "step": 20827 }, { "epoch": 31.179640718562876, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1248, "step": 20828 }, { "epoch": 31.181137724550897, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1266, "step": 20829 }, { "epoch": 31.182634730538922, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.124, "step": 20830 }, { "epoch": 31.184131736526947, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1306, "step": 20831 }, { "epoch": 31.18562874251497, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1234, "step": 20832 }, { "epoch": 31.187125748502993, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1311, "step": 20833 }, { "epoch": 31.188622754491018, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1241, "step": 20834 }, { "epoch": 31.190119760479043, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.122, "step": 20835 }, { "epoch": 31.191616766467067, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1268, "step": 20836 }, { "epoch": 31.19311377245509, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1291, "step": 20837 }, { "epoch": 31.194610778443113, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1266, "step": 20838 }, { "epoch": 31.19610778443114, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1357, "step": 20839 }, { "epoch": 31.197604790419163, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1259, "step": 20840 }, { "epoch": 31.199101796407184, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1199, "step": 20841 }, { "epoch": 31.20059880239521, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1276, "step": 20842 }, { "epoch": 31.202095808383234, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1321, "step": 20843 }, { "epoch": 31.20359281437126, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1282, "step": 20844 }, { "epoch": 31.20508982035928, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1258, "step": 20845 }, { "epoch": 31.206586826347305, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1262, "step": 20846 }, { "epoch": 31.20808383233533, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1233, "step": 20847 }, { "epoch": 31.209580838323355, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1269, "step": 20848 }, { "epoch": 31.211077844311376, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.122, "step": 20849 }, { "epoch": 31.2125748502994, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1252, "step": 20850 }, { "epoch": 31.214071856287426, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 20851 }, { "epoch": 31.21556886227545, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1302, "step": 20852 }, { "epoch": 31.21706586826347, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1316, "step": 20853 }, { "epoch": 31.218562874251496, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1269, "step": 20854 }, { "epoch": 31.22005988023952, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1257, "step": 20855 }, { "epoch": 31.221556886227546, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1211, "step": 20856 }, { "epoch": 31.223053892215567, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1268, "step": 20857 }, { "epoch": 31.224550898203592, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1244, "step": 20858 }, { "epoch": 31.226047904191617, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1275, "step": 20859 }, { "epoch": 31.227544910179642, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1185, "step": 20860 }, { "epoch": 31.229041916167663, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1238, "step": 20861 }, { "epoch": 31.230538922155688, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1298, "step": 20862 }, { "epoch": 31.232035928143713, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.133, "step": 20863 }, { "epoch": 31.233532934131738, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1247, "step": 20864 }, { "epoch": 31.23502994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1314, "step": 20865 }, { "epoch": 31.236526946107784, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.128, "step": 20866 }, { "epoch": 31.23802395209581, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1277, "step": 20867 }, { "epoch": 31.239520958083833, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1307, "step": 20868 }, { "epoch": 31.241017964071855, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1214, "step": 20869 }, { "epoch": 31.24251497005988, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.124, "step": 20870 }, { "epoch": 31.244011976047904, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1275, "step": 20871 }, { "epoch": 31.24550898203593, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1304, "step": 20872 }, { "epoch": 31.24700598802395, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1265, "step": 20873 }, { "epoch": 31.248502994011975, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1285, "step": 20874 }, { "epoch": 31.25, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1274, "step": 20875 }, { "epoch": 31.251497005988025, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1211, "step": 20876 }, { "epoch": 31.25299401197605, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1259, "step": 20877 }, { "epoch": 31.25449101796407, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1236, "step": 20878 }, { "epoch": 31.255988023952096, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1253, "step": 20879 }, { "epoch": 31.25748502994012, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1305, "step": 20880 }, { "epoch": 31.258982035928145, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1313, "step": 20881 }, { "epoch": 31.260479041916167, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1249, "step": 20882 }, { "epoch": 31.26197604790419, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1226, "step": 20883 }, { "epoch": 31.263473053892216, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1207, "step": 20884 }, { "epoch": 31.26497005988024, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1233, "step": 20885 }, { "epoch": 31.266467065868262, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1264, "step": 20886 }, { "epoch": 31.267964071856287, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1216, "step": 20887 }, { "epoch": 31.269461077844312, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1321, "step": 20888 }, { "epoch": 31.270958083832337, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1253, "step": 20889 }, { "epoch": 31.272455089820358, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1328, "step": 20890 }, { "epoch": 31.273952095808383, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.128, "step": 20891 }, { "epoch": 31.275449101796408, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1247, "step": 20892 }, { "epoch": 31.276946107784433, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1327, "step": 20893 }, { "epoch": 31.278443113772454, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1325, "step": 20894 }, { "epoch": 31.27994011976048, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1228, "step": 20895 }, { "epoch": 31.281437125748504, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 20896 }, { "epoch": 31.28293413173653, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1302, "step": 20897 }, { "epoch": 31.28443113772455, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1237, "step": 20898 }, { "epoch": 31.285928143712574, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1194, "step": 20899 }, { "epoch": 31.2874251497006, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.125, "step": 20900 }, { "epoch": 31.288922155688624, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1282, "step": 20901 }, { "epoch": 31.290419161676645, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1222, "step": 20902 }, { "epoch": 31.29191616766467, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 20903 }, { "epoch": 31.293413173652695, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 20904 }, { "epoch": 31.29491017964072, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1306, "step": 20905 }, { "epoch": 31.29640718562874, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1289, "step": 20906 }, { "epoch": 31.297904191616766, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1258, "step": 20907 }, { "epoch": 31.29940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 20908 }, { "epoch": 31.300898203592816, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.122, "step": 20909 }, { "epoch": 31.302395209580837, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.125, "step": 20910 }, { "epoch": 31.30389221556886, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.126, "step": 20911 }, { "epoch": 31.305389221556887, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1211, "step": 20912 }, { "epoch": 31.30688622754491, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1287, "step": 20913 }, { "epoch": 31.308383233532933, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1362, "step": 20914 }, { "epoch": 31.309880239520957, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1276, "step": 20915 }, { "epoch": 31.311377245508982, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1248, "step": 20916 }, { "epoch": 31.312874251497007, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1281, "step": 20917 }, { "epoch": 31.31437125748503, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1284, "step": 20918 }, { "epoch": 31.315868263473053, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1249, "step": 20919 }, { "epoch": 31.317365269461078, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1238, "step": 20920 }, { "epoch": 31.318862275449103, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1222, "step": 20921 }, { "epoch": 31.320359281437124, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1207, "step": 20922 }, { "epoch": 31.32185628742515, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1248, "step": 20923 }, { "epoch": 31.323353293413174, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1275, "step": 20924 }, { "epoch": 31.3248502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1301, "step": 20925 }, { "epoch": 31.32634730538922, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.13, "step": 20926 }, { "epoch": 31.327844311377245, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1249, "step": 20927 }, { "epoch": 31.32934131736527, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1268, "step": 20928 }, { "epoch": 31.330838323353294, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1256, "step": 20929 }, { "epoch": 31.33233532934132, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1245, "step": 20930 }, { "epoch": 31.33383233532934, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1263, "step": 20931 }, { "epoch": 31.335329341317365, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1279, "step": 20932 }, { "epoch": 31.33682634730539, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1278, "step": 20933 }, { "epoch": 31.338323353293415, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1276, "step": 20934 }, { "epoch": 31.339820359281436, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1305, "step": 20935 }, { "epoch": 31.34131736526946, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1277, "step": 20936 }, { "epoch": 31.342814371257486, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1251, "step": 20937 }, { "epoch": 31.34431137724551, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 20938 }, { "epoch": 31.345808383233532, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1258, "step": 20939 }, { "epoch": 31.347305389221557, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1291, "step": 20940 }, { "epoch": 31.34880239520958, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1239, "step": 20941 }, { "epoch": 31.350299401197606, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1294, "step": 20942 }, { "epoch": 31.351796407185628, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1209, "step": 20943 }, { "epoch": 31.353293413173652, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1271, "step": 20944 }, { "epoch": 31.354790419161677, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 20945 }, { "epoch": 31.356287425149702, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1325, "step": 20946 }, { "epoch": 31.357784431137723, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1281, "step": 20947 }, { "epoch": 31.35928143712575, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1273, "step": 20948 }, { "epoch": 31.360778443113773, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1264, "step": 20949 }, { "epoch": 31.362275449101798, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1271, "step": 20950 }, { "epoch": 31.36377245508982, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1375, "step": 20951 }, { "epoch": 31.365269461077844, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.125, "step": 20952 }, { "epoch": 31.36676646706587, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1306, "step": 20953 }, { "epoch": 31.368263473053894, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1246, "step": 20954 }, { "epoch": 31.369760479041915, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1291, "step": 20955 }, { "epoch": 31.37125748502994, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1357, "step": 20956 }, { "epoch": 31.372754491017965, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1227, "step": 20957 }, { "epoch": 31.37425149700599, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1256, "step": 20958 }, { "epoch": 31.37574850299401, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1268, "step": 20959 }, { "epoch": 31.377245508982035, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1259, "step": 20960 }, { "epoch": 31.37874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1289, "step": 20961 }, { "epoch": 31.380239520958085, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1257, "step": 20962 }, { "epoch": 31.381736526946106, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1286, "step": 20963 }, { "epoch": 31.38323353293413, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1272, "step": 20964 }, { "epoch": 31.384730538922156, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1297, "step": 20965 }, { "epoch": 31.38622754491018, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1279, "step": 20966 }, { "epoch": 31.387724550898202, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.13, "step": 20967 }, { "epoch": 31.389221556886227, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1234, "step": 20968 }, { "epoch": 31.39071856287425, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1281, "step": 20969 }, { "epoch": 31.392215568862277, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1178, "step": 20970 }, { "epoch": 31.393712574850298, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1229, "step": 20971 }, { "epoch": 31.395209580838323, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1301, "step": 20972 }, { "epoch": 31.396706586826348, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1234, "step": 20973 }, { "epoch": 31.398203592814372, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1286, "step": 20974 }, { "epoch": 31.399700598802394, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1241, "step": 20975 }, { "epoch": 31.40119760479042, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1251, "step": 20976 }, { "epoch": 31.402694610778443, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1234, "step": 20977 }, { "epoch": 31.404191616766468, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1264, "step": 20978 }, { "epoch": 31.40568862275449, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1242, "step": 20979 }, { "epoch": 31.407185628742514, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1235, "step": 20980 }, { "epoch": 31.40868263473054, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1226, "step": 20981 }, { "epoch": 31.410179640718564, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1214, "step": 20982 }, { "epoch": 31.411676646706585, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1234, "step": 20983 }, { "epoch": 31.41317365269461, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1263, "step": 20984 }, { "epoch": 31.414670658682635, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 20985 }, { "epoch": 31.41616766467066, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1276, "step": 20986 }, { "epoch": 31.41766467065868, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1226, "step": 20987 }, { "epoch": 31.419161676646706, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.133, "step": 20988 }, { "epoch": 31.42065868263473, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1264, "step": 20989 }, { "epoch": 31.422155688622755, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1215, "step": 20990 }, { "epoch": 31.42365269461078, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 20991 }, { "epoch": 31.4251497005988, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1266, "step": 20992 }, { "epoch": 31.426646706586826, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1238, "step": 20993 }, { "epoch": 31.42814371257485, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1254, "step": 20994 }, { "epoch": 31.429640718562876, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1321, "step": 20995 }, { "epoch": 31.431137724550897, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1188, "step": 20996 }, { "epoch": 31.432634730538922, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1251, "step": 20997 }, { "epoch": 31.434131736526947, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 20998 }, { "epoch": 31.43562874251497, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1305, "step": 20999 }, { "epoch": 31.437125748502993, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1229, "step": 21000 }, { "epoch": 31.438622754491018, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 21001 }, { "epoch": 31.440119760479043, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1253, "step": 21002 }, { "epoch": 31.441616766467067, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1275, "step": 21003 }, { "epoch": 31.44311377245509, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 21004 }, { "epoch": 31.444610778443113, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1247, "step": 21005 }, { "epoch": 31.44610778443114, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1253, "step": 21006 }, { "epoch": 31.447604790419163, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 21007 }, { "epoch": 31.449101796407184, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 21008 }, { "epoch": 31.45059880239521, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1327, "step": 21009 }, { "epoch": 31.452095808383234, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.134, "step": 21010 }, { "epoch": 31.45359281437126, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1275, "step": 21011 }, { "epoch": 31.45508982035928, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1221, "step": 21012 }, { "epoch": 31.456586826347305, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1253, "step": 21013 }, { "epoch": 31.45808383233533, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 21014 }, { "epoch": 31.459580838323355, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1297, "step": 21015 }, { "epoch": 31.461077844311376, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1291, "step": 21016 }, { "epoch": 31.4625748502994, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 21017 }, { "epoch": 31.464071856287426, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 21018 }, { "epoch": 31.46556886227545, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1299, "step": 21019 }, { "epoch": 31.46706586826347, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1262, "step": 21020 }, { "epoch": 31.468562874251496, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1248, "step": 21021 }, { "epoch": 31.47005988023952, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1337, "step": 21022 }, { "epoch": 31.471556886227546, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1268, "step": 21023 }, { "epoch": 31.473053892215567, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.125, "step": 21024 }, { "epoch": 31.474550898203592, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1338, "step": 21025 }, { "epoch": 31.476047904191617, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1259, "step": 21026 }, { "epoch": 31.477544910179642, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1306, "step": 21027 }, { "epoch": 31.479041916167663, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 21028 }, { "epoch": 31.480538922155688, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1265, "step": 21029 }, { "epoch": 31.482035928143713, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1269, "step": 21030 }, { "epoch": 31.483532934131738, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1267, "step": 21031 }, { "epoch": 31.48502994011976, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1301, "step": 21032 }, { "epoch": 31.486526946107784, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1258, "step": 21033 }, { "epoch": 31.48802395209581, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.123, "step": 21034 }, { "epoch": 31.489520958083833, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1243, "step": 21035 }, { "epoch": 31.491017964071855, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1292, "step": 21036 }, { "epoch": 31.49251497005988, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1255, "step": 21037 }, { "epoch": 31.494011976047904, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1267, "step": 21038 }, { "epoch": 31.49550898203593, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1229, "step": 21039 }, { "epoch": 31.49700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1271, "step": 21040 }, { "epoch": 31.498502994011975, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1241, "step": 21041 }, { "epoch": 31.5, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1265, "step": 21042 }, { "epoch": 31.501497005988025, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1278, "step": 21043 }, { "epoch": 31.50299401197605, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 21044 }, { "epoch": 31.50449101796407, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 21045 }, { "epoch": 31.505988023952096, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 21046 }, { "epoch": 31.50748502994012, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1304, "step": 21047 }, { "epoch": 31.508982035928145, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 21048 }, { "epoch": 31.510479041916167, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1279, "step": 21049 }, { "epoch": 31.51197604790419, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1245, "step": 21050 }, { "epoch": 31.513473053892216, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 21051 }, { "epoch": 31.51497005988024, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1307, "step": 21052 }, { "epoch": 31.516467065868262, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1315, "step": 21053 }, { "epoch": 31.517964071856287, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1314, "step": 21054 }, { "epoch": 31.519461077844312, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1279, "step": 21055 }, { "epoch": 31.520958083832337, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1235, "step": 21056 }, { "epoch": 31.522455089820358, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1292, "step": 21057 }, { "epoch": 31.523952095808383, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1221, "step": 21058 }, { "epoch": 31.525449101796408, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1249, "step": 21059 }, { "epoch": 31.526946107784433, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1333, "step": 21060 }, { "epoch": 31.528443113772454, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1203, "step": 21061 }, { "epoch": 31.52994011976048, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1316, "step": 21062 }, { "epoch": 31.531437125748504, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1256, "step": 21063 }, { "epoch": 31.53293413173653, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1235, "step": 21064 }, { "epoch": 31.53443113772455, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1283, "step": 21065 }, { "epoch": 31.535928143712574, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1253, "step": 21066 }, { "epoch": 31.5374251497006, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1251, "step": 21067 }, { "epoch": 31.538922155688624, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1237, "step": 21068 }, { "epoch": 31.540419161676645, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 21069 }, { "epoch": 31.54191616766467, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1231, "step": 21070 }, { "epoch": 31.543413173652695, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1271, "step": 21071 }, { "epoch": 31.54491017964072, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1242, "step": 21072 }, { "epoch": 31.54640718562874, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1254, "step": 21073 }, { "epoch": 31.547904191616766, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1254, "step": 21074 }, { "epoch": 31.54940119760479, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1272, "step": 21075 }, { "epoch": 31.550898203592816, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1279, "step": 21076 }, { "epoch": 31.552395209580837, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1209, "step": 21077 }, { "epoch": 31.55389221556886, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1304, "step": 21078 }, { "epoch": 31.555389221556887, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1296, "step": 21079 }, { "epoch": 31.55688622754491, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 21080 }, { "epoch": 31.558383233532933, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1248, "step": 21081 }, { "epoch": 31.559880239520957, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1244, "step": 21082 }, { "epoch": 31.561377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1292, "step": 21083 }, { "epoch": 31.562874251497007, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1265, "step": 21084 }, { "epoch": 31.56437125748503, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1243, "step": 21085 }, { "epoch": 31.565868263473053, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1293, "step": 21086 }, { "epoch": 31.567365269461078, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1314, "step": 21087 }, { "epoch": 31.568862275449103, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1324, "step": 21088 }, { "epoch": 31.570359281437124, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1188, "step": 21089 }, { "epoch": 31.57185628742515, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1266, "step": 21090 }, { "epoch": 31.573353293413174, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1249, "step": 21091 }, { "epoch": 31.5748502994012, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1343, "step": 21092 }, { "epoch": 31.57634730538922, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.126, "step": 21093 }, { "epoch": 31.577844311377245, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1243, "step": 21094 }, { "epoch": 31.57934131736527, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1291, "step": 21095 }, { "epoch": 31.580838323353294, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1256, "step": 21096 }, { "epoch": 31.58233532934132, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 21097 }, { "epoch": 31.58383233532934, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.134, "step": 21098 }, { "epoch": 31.585329341317365, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1277, "step": 21099 }, { "epoch": 31.58682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.122, "step": 21100 }, { "epoch": 31.58832335329341, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1271, "step": 21101 }, { "epoch": 31.589820359281436, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1306, "step": 21102 }, { "epoch": 31.59131736526946, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 21103 }, { "epoch": 31.592814371257486, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1257, "step": 21104 }, { "epoch": 31.59431137724551, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1272, "step": 21105 }, { "epoch": 31.595808383233532, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.123, "step": 21106 }, { "epoch": 31.597305389221557, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1256, "step": 21107 }, { "epoch": 31.59880239520958, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1229, "step": 21108 }, { "epoch": 31.600299401197606, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1172, "step": 21109 }, { "epoch": 31.601796407185628, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1223, "step": 21110 }, { "epoch": 31.603293413173652, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1337, "step": 21111 }, { "epoch": 31.604790419161677, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.13, "step": 21112 }, { "epoch": 31.606287425149702, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1247, "step": 21113 }, { "epoch": 31.607784431137723, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1251, "step": 21114 }, { "epoch": 31.60928143712575, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1244, "step": 21115 }, { "epoch": 31.610778443113773, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1263, "step": 21116 }, { "epoch": 31.612275449101798, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1217, "step": 21117 }, { "epoch": 31.61377245508982, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1273, "step": 21118 }, { "epoch": 31.615269461077844, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1239, "step": 21119 }, { "epoch": 31.61676646706587, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1302, "step": 21120 }, { "epoch": 31.618263473053894, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1329, "step": 21121 }, { "epoch": 31.619760479041915, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1238, "step": 21122 }, { "epoch": 31.62125748502994, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1281, "step": 21123 }, { "epoch": 31.622754491017965, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1252, "step": 21124 }, { "epoch": 31.62425149700599, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1308, "step": 21125 }, { "epoch": 31.62574850299401, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1274, "step": 21126 }, { "epoch": 31.627245508982035, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1252, "step": 21127 }, { "epoch": 31.62874251497006, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1311, "step": 21128 }, { "epoch": 31.630239520958085, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1196, "step": 21129 }, { "epoch": 31.631736526946106, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1232, "step": 21130 }, { "epoch": 31.63323353293413, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1246, "step": 21131 }, { "epoch": 31.634730538922156, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1262, "step": 21132 }, { "epoch": 31.63622754491018, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.128, "step": 21133 }, { "epoch": 31.637724550898202, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1226, "step": 21134 }, { "epoch": 31.639221556886227, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.128, "step": 21135 }, { "epoch": 31.64071856287425, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1295, "step": 21136 }, { "epoch": 31.642215568862277, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1228, "step": 21137 }, { "epoch": 31.643712574850298, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1208, "step": 21138 }, { "epoch": 31.645209580838323, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1311, "step": 21139 }, { "epoch": 31.646706586826348, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1243, "step": 21140 }, { "epoch": 31.648203592814372, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1212, "step": 21141 }, { "epoch": 31.649700598802394, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1189, "step": 21142 }, { "epoch": 31.65119760479042, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.127, "step": 21143 }, { "epoch": 31.652694610778443, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1242, "step": 21144 }, { "epoch": 31.654191616766468, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 21145 }, { "epoch": 31.65568862275449, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1254, "step": 21146 }, { "epoch": 31.657185628742514, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1276, "step": 21147 }, { "epoch": 31.65868263473054, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1284, "step": 21148 }, { "epoch": 31.660179640718564, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1246, "step": 21149 }, { "epoch": 31.66167664670659, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1248, "step": 21150 }, { "epoch": 31.66317365269461, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1377, "step": 21151 }, { "epoch": 31.664670658682635, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1186, "step": 21152 }, { "epoch": 31.66616766467066, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1303, "step": 21153 }, { "epoch": 31.66766467065868, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1237, "step": 21154 }, { "epoch": 31.669161676646706, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1223, "step": 21155 }, { "epoch": 31.67065868263473, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1229, "step": 21156 }, { "epoch": 31.672155688622755, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1261, "step": 21157 }, { "epoch": 31.67365269461078, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1236, "step": 21158 }, { "epoch": 31.6751497005988, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1301, "step": 21159 }, { "epoch": 31.676646706586826, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1232, "step": 21160 }, { "epoch": 31.67814371257485, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1256, "step": 21161 }, { "epoch": 31.679640718562876, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.125, "step": 21162 }, { "epoch": 31.681137724550897, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1186, "step": 21163 }, { "epoch": 31.682634730538922, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1325, "step": 21164 }, { "epoch": 31.684131736526947, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1198, "step": 21165 }, { "epoch": 31.68562874251497, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1257, "step": 21166 }, { "epoch": 31.687125748502993, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1233, "step": 21167 }, { "epoch": 31.688622754491018, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1253, "step": 21168 }, { "epoch": 31.690119760479043, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.13, "step": 21169 }, { "epoch": 31.691616766467067, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1234, "step": 21170 }, { "epoch": 31.69311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1323, "step": 21171 }, { "epoch": 31.694610778443113, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1299, "step": 21172 }, { "epoch": 31.69610778443114, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1287, "step": 21173 }, { "epoch": 31.697604790419163, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1249, "step": 21174 }, { "epoch": 31.699101796407184, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1268, "step": 21175 }, { "epoch": 31.70059880239521, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1297, "step": 21176 }, { "epoch": 31.702095808383234, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.129, "step": 21177 }, { "epoch": 31.70359281437126, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1287, "step": 21178 }, { "epoch": 31.70508982035928, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1295, "step": 21179 }, { "epoch": 31.706586826347305, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.126, "step": 21180 }, { "epoch": 31.70808383233533, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1274, "step": 21181 }, { "epoch": 31.709580838323355, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1269, "step": 21182 }, { "epoch": 31.711077844311376, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1223, "step": 21183 }, { "epoch": 31.7125748502994, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1239, "step": 21184 }, { "epoch": 31.714071856287426, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1256, "step": 21185 }, { "epoch": 31.71556886227545, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1217, "step": 21186 }, { "epoch": 31.71706586826347, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1267, "step": 21187 }, { "epoch": 31.718562874251496, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1283, "step": 21188 }, { "epoch": 31.72005988023952, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1273, "step": 21189 }, { "epoch": 31.721556886227546, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1324, "step": 21190 }, { "epoch": 31.723053892215567, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1339, "step": 21191 }, { "epoch": 31.724550898203592, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1237, "step": 21192 }, { "epoch": 31.726047904191617, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1307, "step": 21193 }, { "epoch": 31.727544910179642, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1287, "step": 21194 }, { "epoch": 31.729041916167663, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.126, "step": 21195 }, { "epoch": 31.730538922155688, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.131, "step": 21196 }, { "epoch": 31.732035928143713, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1227, "step": 21197 }, { "epoch": 31.733532934131738, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1259, "step": 21198 }, { "epoch": 31.73502994011976, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.127, "step": 21199 }, { "epoch": 31.736526946107784, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1263, "step": 21200 }, { "epoch": 31.73802395209581, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1297, "step": 21201 }, { "epoch": 31.739520958083833, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1294, "step": 21202 }, { "epoch": 31.741017964071855, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1291, "step": 21203 }, { "epoch": 31.74251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1285, "step": 21204 }, { "epoch": 31.744011976047904, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1313, "step": 21205 }, { "epoch": 31.74550898203593, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1308, "step": 21206 }, { "epoch": 31.74700598802395, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1211, "step": 21207 }, { "epoch": 31.748502994011975, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1267, "step": 21208 }, { "epoch": 31.75, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 21209 }, { "epoch": 31.751497005988025, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1339, "step": 21210 }, { "epoch": 31.75299401197605, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1303, "step": 21211 }, { "epoch": 31.75449101796407, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 21212 }, { "epoch": 31.755988023952096, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1263, "step": 21213 }, { "epoch": 31.75748502994012, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1217, "step": 21214 }, { "epoch": 31.758982035928145, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1249, "step": 21215 }, { "epoch": 31.760479041916167, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1236, "step": 21216 }, { "epoch": 31.76197604790419, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1221, "step": 21217 }, { "epoch": 31.763473053892216, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1231, "step": 21218 }, { "epoch": 31.76497005988024, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1212, "step": 21219 }, { "epoch": 31.766467065868262, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.128, "step": 21220 }, { "epoch": 31.767964071856287, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1237, "step": 21221 }, { "epoch": 31.769461077844312, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1235, "step": 21222 }, { "epoch": 31.770958083832337, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1278, "step": 21223 }, { "epoch": 31.772455089820358, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1276, "step": 21224 }, { "epoch": 31.773952095808383, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1336, "step": 21225 }, { "epoch": 31.775449101796408, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1326, "step": 21226 }, { "epoch": 31.776946107784433, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1199, "step": 21227 }, { "epoch": 31.778443113772454, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1281, "step": 21228 }, { "epoch": 31.77994011976048, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1278, "step": 21229 }, { "epoch": 31.781437125748504, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1319, "step": 21230 }, { "epoch": 31.78293413173653, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1286, "step": 21231 }, { "epoch": 31.78443113772455, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1322, "step": 21232 }, { "epoch": 31.785928143712574, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1241, "step": 21233 }, { "epoch": 31.7874251497006, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 21234 }, { "epoch": 31.788922155688624, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1248, "step": 21235 }, { "epoch": 31.790419161676645, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1309, "step": 21236 }, { "epoch": 31.79191616766467, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1244, "step": 21237 }, { "epoch": 31.793413173652695, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1238, "step": 21238 }, { "epoch": 31.79491017964072, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1316, "step": 21239 }, { "epoch": 31.79640718562874, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1326, "step": 21240 }, { "epoch": 31.797904191616766, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1261, "step": 21241 }, { "epoch": 31.79940119760479, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1244, "step": 21242 }, { "epoch": 31.800898203592816, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1331, "step": 21243 }, { "epoch": 31.802395209580837, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1244, "step": 21244 }, { "epoch": 31.80389221556886, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1291, "step": 21245 }, { "epoch": 31.805389221556887, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1234, "step": 21246 }, { "epoch": 31.80688622754491, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1316, "step": 21247 }, { "epoch": 31.808383233532933, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1302, "step": 21248 }, { "epoch": 31.809880239520957, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1305, "step": 21249 }, { "epoch": 31.811377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1229, "step": 21250 }, { "epoch": 31.812874251497007, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1304, "step": 21251 }, { "epoch": 31.81437125748503, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.125, "step": 21252 }, { "epoch": 31.815868263473053, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1221, "step": 21253 }, { "epoch": 31.817365269461078, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1311, "step": 21254 }, { "epoch": 31.818862275449103, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1231, "step": 21255 }, { "epoch": 31.820359281437124, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1292, "step": 21256 }, { "epoch": 31.82185628742515, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.124, "step": 21257 }, { "epoch": 31.823353293413174, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1323, "step": 21258 }, { "epoch": 31.8248502994012, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1357, "step": 21259 }, { "epoch": 31.82634730538922, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1268, "step": 21260 }, { "epoch": 31.827844311377245, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1233, "step": 21261 }, { "epoch": 31.82934131736527, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1258, "step": 21262 }, { "epoch": 31.830838323353294, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1249, "step": 21263 }, { "epoch": 31.83233532934132, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1292, "step": 21264 }, { "epoch": 31.83383233532934, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1256, "step": 21265 }, { "epoch": 31.835329341317365, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1284, "step": 21266 }, { "epoch": 31.83682634730539, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.134, "step": 21267 }, { "epoch": 31.83832335329341, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1237, "step": 21268 }, { "epoch": 31.839820359281436, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1261, "step": 21269 }, { "epoch": 31.84131736526946, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 21270 }, { "epoch": 31.842814371257486, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1236, "step": 21271 }, { "epoch": 31.84431137724551, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1297, "step": 21272 }, { "epoch": 31.845808383233532, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1213, "step": 21273 }, { "epoch": 31.847305389221557, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1248, "step": 21274 }, { "epoch": 31.84880239520958, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.13, "step": 21275 }, { "epoch": 31.850299401197606, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1236, "step": 21276 }, { "epoch": 31.851796407185628, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1261, "step": 21277 }, { "epoch": 31.853293413173652, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1257, "step": 21278 }, { "epoch": 31.854790419161677, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1295, "step": 21279 }, { "epoch": 31.856287425149702, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1244, "step": 21280 }, { "epoch": 31.857784431137723, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1316, "step": 21281 }, { "epoch": 31.85928143712575, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1278, "step": 21282 }, { "epoch": 31.860778443113773, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1245, "step": 21283 }, { "epoch": 31.862275449101798, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1259, "step": 21284 }, { "epoch": 31.86377245508982, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 21285 }, { "epoch": 31.865269461077844, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.13, "step": 21286 }, { "epoch": 31.86676646706587, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1271, "step": 21287 }, { "epoch": 31.868263473053894, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1232, "step": 21288 }, { "epoch": 31.869760479041915, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1263, "step": 21289 }, { "epoch": 31.87125748502994, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1263, "step": 21290 }, { "epoch": 31.872754491017965, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1204, "step": 21291 }, { "epoch": 31.87425149700599, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1358, "step": 21292 }, { "epoch": 31.87574850299401, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 21293 }, { "epoch": 31.877245508982035, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1221, "step": 21294 }, { "epoch": 31.87874251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1275, "step": 21295 }, { "epoch": 31.880239520958085, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1271, "step": 21296 }, { "epoch": 31.881736526946106, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1267, "step": 21297 }, { "epoch": 31.88323353293413, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1269, "step": 21298 }, { "epoch": 31.884730538922156, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1238, "step": 21299 }, { "epoch": 31.88622754491018, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1203, "step": 21300 }, { "epoch": 31.887724550898202, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1261, "step": 21301 }, { "epoch": 31.889221556886227, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1237, "step": 21302 }, { "epoch": 31.89071856287425, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1223, "step": 21303 }, { "epoch": 31.892215568862277, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1251, "step": 21304 }, { "epoch": 31.893712574850298, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1299, "step": 21305 }, { "epoch": 31.895209580838323, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1231, "step": 21306 }, { "epoch": 31.896706586826348, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1226, "step": 21307 }, { "epoch": 31.898203592814372, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.125, "step": 21308 }, { "epoch": 31.899700598802394, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1317, "step": 21309 }, { "epoch": 31.90119760479042, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1308, "step": 21310 }, { "epoch": 31.902694610778443, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1279, "step": 21311 }, { "epoch": 31.904191616766468, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.132, "step": 21312 }, { "epoch": 31.90568862275449, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1189, "step": 21313 }, { "epoch": 31.907185628742514, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1247, "step": 21314 }, { "epoch": 31.90868263473054, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1347, "step": 21315 }, { "epoch": 31.910179640718564, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1228, "step": 21316 }, { "epoch": 31.91167664670659, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1297, "step": 21317 }, { "epoch": 31.91317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1219, "step": 21318 }, { "epoch": 31.914670658682635, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1255, "step": 21319 }, { "epoch": 31.91616766467066, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1268, "step": 21320 }, { "epoch": 31.91766467065868, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.126, "step": 21321 }, { "epoch": 31.919161676646706, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1254, "step": 21322 }, { "epoch": 31.92065868263473, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1206, "step": 21323 }, { "epoch": 31.922155688622755, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1325, "step": 21324 }, { "epoch": 31.92365269461078, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1308, "step": 21325 }, { "epoch": 31.9251497005988, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.13, "step": 21326 }, { "epoch": 31.926646706586826, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1282, "step": 21327 }, { "epoch": 31.92814371257485, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1245, "step": 21328 }, { "epoch": 31.929640718562876, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1218, "step": 21329 }, { "epoch": 31.931137724550897, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1264, "step": 21330 }, { "epoch": 31.932634730538922, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1292, "step": 21331 }, { "epoch": 31.934131736526947, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1196, "step": 21332 }, { "epoch": 31.93562874251497, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1257, "step": 21333 }, { "epoch": 31.937125748502993, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1292, "step": 21334 }, { "epoch": 31.938622754491018, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.123, "step": 21335 }, { "epoch": 31.940119760479043, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1289, "step": 21336 }, { "epoch": 31.941616766467067, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1267, "step": 21337 }, { "epoch": 31.94311377245509, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1293, "step": 21338 }, { "epoch": 31.944610778443113, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1273, "step": 21339 }, { "epoch": 31.94610778443114, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1294, "step": 21340 }, { "epoch": 31.947604790419163, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1304, "step": 21341 }, { "epoch": 31.949101796407184, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1274, "step": 21342 }, { "epoch": 31.95059880239521, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1314, "step": 21343 }, { "epoch": 31.952095808383234, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1285, "step": 21344 }, { "epoch": 31.95359281437126, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.133, "step": 21345 }, { "epoch": 31.95508982035928, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.13, "step": 21346 }, { "epoch": 31.956586826347305, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 21347 }, { "epoch": 31.95808383233533, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1287, "step": 21348 }, { "epoch": 31.959580838323355, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1306, "step": 21349 }, { "epoch": 31.961077844311376, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1221, "step": 21350 }, { "epoch": 31.9625748502994, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1193, "step": 21351 }, { "epoch": 31.964071856287426, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1232, "step": 21352 }, { "epoch": 31.96556886227545, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1286, "step": 21353 }, { "epoch": 31.96706586826347, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1267, "step": 21354 }, { "epoch": 31.968562874251496, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1316, "step": 21355 }, { "epoch": 31.97005988023952, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1263, "step": 21356 }, { "epoch": 31.971556886227546, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1273, "step": 21357 }, { "epoch": 31.973053892215567, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1276, "step": 21358 }, { "epoch": 31.974550898203592, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1297, "step": 21359 }, { "epoch": 31.976047904191617, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1305, "step": 21360 }, { "epoch": 31.977544910179642, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1331, "step": 21361 }, { "epoch": 31.979041916167663, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1321, "step": 21362 }, { "epoch": 31.980538922155688, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1319, "step": 21363 }, { "epoch": 31.982035928143713, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1308, "step": 21364 }, { "epoch": 31.983532934131738, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1252, "step": 21365 }, { "epoch": 31.98502994011976, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1235, "step": 21366 }, { "epoch": 31.986526946107784, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1262, "step": 21367 }, { "epoch": 31.98802395209581, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1273, "step": 21368 }, { "epoch": 31.989520958083833, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1336, "step": 21369 }, { "epoch": 31.991017964071855, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1251, "step": 21370 }, { "epoch": 31.99251497005988, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.125, "step": 21371 }, { "epoch": 31.994011976047904, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1269, "step": 21372 }, { "epoch": 31.99550898203593, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1227, "step": 21373 }, { "epoch": 31.99700598802395, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1321, "step": 21374 }, { "epoch": 31.998502994011975, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1278, "step": 21375 }, { "epoch": 32.0, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1298, "step": 21376 }, { "epoch": 32.00149700598802, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1242, "step": 21377 }, { "epoch": 32.00299401197605, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.119, "step": 21378 }, { "epoch": 32.00449101796407, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1232, "step": 21379 }, { "epoch": 32.0059880239521, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1275, "step": 21380 }, { "epoch": 32.00748502994012, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1248, "step": 21381 }, { "epoch": 32.00898203592814, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1272, "step": 21382 }, { "epoch": 32.01047904191617, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1285, "step": 21383 }, { "epoch": 32.01197604790419, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 21384 }, { "epoch": 32.01347305389221, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1256, "step": 21385 }, { "epoch": 32.01497005988024, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1251, "step": 21386 }, { "epoch": 32.01646706586826, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1268, "step": 21387 }, { "epoch": 32.01796407185629, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1218, "step": 21388 }, { "epoch": 32.01946107784431, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1246, "step": 21389 }, { "epoch": 32.02095808383233, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1313, "step": 21390 }, { "epoch": 32.02245508982036, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1237, "step": 21391 }, { "epoch": 32.02395209580838, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1254, "step": 21392 }, { "epoch": 32.025449101796404, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1265, "step": 21393 }, { "epoch": 32.02694610778443, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1286, "step": 21394 }, { "epoch": 32.028443113772454, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1271, "step": 21395 }, { "epoch": 32.02994011976048, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1214, "step": 21396 }, { "epoch": 32.0314371257485, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1275, "step": 21397 }, { "epoch": 32.032934131736525, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1339, "step": 21398 }, { "epoch": 32.03443113772455, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1283, "step": 21399 }, { "epoch": 32.035928143712574, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.128, "step": 21400 }, { "epoch": 32.037425149700596, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1295, "step": 21401 }, { "epoch": 32.038922155688624, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1245, "step": 21402 }, { "epoch": 32.040419161676645, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1259, "step": 21403 }, { "epoch": 32.041916167664674, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1225, "step": 21404 }, { "epoch": 32.043413173652695, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1291, "step": 21405 }, { "epoch": 32.044910179640716, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1294, "step": 21406 }, { "epoch": 32.046407185628745, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1246, "step": 21407 }, { "epoch": 32.047904191616766, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1232, "step": 21408 }, { "epoch": 32.04940119760479, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1234, "step": 21409 }, { "epoch": 32.050898203592816, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1241, "step": 21410 }, { "epoch": 32.05239520958084, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.129, "step": 21411 }, { "epoch": 32.053892215568865, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1205, "step": 21412 }, { "epoch": 32.05538922155689, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.126, "step": 21413 }, { "epoch": 32.05688622754491, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1374, "step": 21414 }, { "epoch": 32.058383233532936, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1264, "step": 21415 }, { "epoch": 32.05988023952096, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1292, "step": 21416 }, { "epoch": 32.06137724550898, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1234, "step": 21417 }, { "epoch": 32.06287425149701, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1287, "step": 21418 }, { "epoch": 32.06437125748503, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1255, "step": 21419 }, { "epoch": 32.06586826347306, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1289, "step": 21420 }, { "epoch": 32.06736526946108, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 21421 }, { "epoch": 32.0688622754491, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1287, "step": 21422 }, { "epoch": 32.07035928143713, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1254, "step": 21423 }, { "epoch": 32.07185628742515, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1339, "step": 21424 }, { "epoch": 32.07335329341317, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1315, "step": 21425 }, { "epoch": 32.0748502994012, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1262, "step": 21426 }, { "epoch": 32.07634730538922, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1263, "step": 21427 }, { "epoch": 32.07784431137725, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1208, "step": 21428 }, { "epoch": 32.07934131736527, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1248, "step": 21429 }, { "epoch": 32.08083832335329, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.1256, "step": 21430 }, { "epoch": 32.08233532934132, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.126, "step": 21431 }, { "epoch": 32.08383233532934, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1232, "step": 21432 }, { "epoch": 32.08532934131737, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.127, "step": 21433 }, { "epoch": 32.08682634730539, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1255, "step": 21434 }, { "epoch": 32.08832335329341, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1239, "step": 21435 }, { "epoch": 32.08982035928144, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.123, "step": 21436 }, { "epoch": 32.09131736526946, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 21437 }, { "epoch": 32.09281437125748, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 21438 }, { "epoch": 32.09431137724551, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1246, "step": 21439 }, { "epoch": 32.09580838323353, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1239, "step": 21440 }, { "epoch": 32.09730538922156, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1239, "step": 21441 }, { "epoch": 32.09880239520958, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1248, "step": 21442 }, { "epoch": 32.1002994011976, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1235, "step": 21443 }, { "epoch": 32.10179640718563, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1256, "step": 21444 }, { "epoch": 32.10329341317365, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1213, "step": 21445 }, { "epoch": 32.104790419161674, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1288, "step": 21446 }, { "epoch": 32.1062874251497, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1279, "step": 21447 }, { "epoch": 32.10778443113772, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.128, "step": 21448 }, { "epoch": 32.10928143712575, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1256, "step": 21449 }, { "epoch": 32.11077844311377, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1293, "step": 21450 }, { "epoch": 32.112275449101794, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1265, "step": 21451 }, { "epoch": 32.11377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1303, "step": 21452 }, { "epoch": 32.115269461077844, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1235, "step": 21453 }, { "epoch": 32.116766467065865, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.125, "step": 21454 }, { "epoch": 32.118263473053894, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1205, "step": 21455 }, { "epoch": 32.119760479041915, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.124, "step": 21456 }, { "epoch": 32.12125748502994, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1275, "step": 21457 }, { "epoch": 32.122754491017965, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1283, "step": 21458 }, { "epoch": 32.124251497005986, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1314, "step": 21459 }, { "epoch": 32.125748502994014, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1278, "step": 21460 }, { "epoch": 32.127245508982035, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1243, "step": 21461 }, { "epoch": 32.12874251497006, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 21462 }, { "epoch": 32.130239520958085, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1337, "step": 21463 }, { "epoch": 32.131736526946106, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1286, "step": 21464 }, { "epoch": 32.133233532934135, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1219, "step": 21465 }, { "epoch": 32.134730538922156, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1306, "step": 21466 }, { "epoch": 32.13622754491018, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.126, "step": 21467 }, { "epoch": 32.137724550898206, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.123, "step": 21468 }, { "epoch": 32.13922155688623, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1257, "step": 21469 }, { "epoch": 32.14071856287425, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1229, "step": 21470 }, { "epoch": 32.14221556886228, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.128, "step": 21471 }, { "epoch": 32.1437125748503, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1272, "step": 21472 }, { "epoch": 32.145209580838326, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1227, "step": 21473 }, { "epoch": 32.14670658682635, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.128, "step": 21474 }, { "epoch": 32.14820359281437, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1272, "step": 21475 }, { "epoch": 32.1497005988024, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1258, "step": 21476 }, { "epoch": 32.15119760479042, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1305, "step": 21477 }, { "epoch": 32.15269461077844, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1259, "step": 21478 }, { "epoch": 32.15419161676647, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1204, "step": 21479 }, { "epoch": 32.15568862275449, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.119, "step": 21480 }, { "epoch": 32.15718562874252, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1221, "step": 21481 }, { "epoch": 32.15868263473054, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1333, "step": 21482 }, { "epoch": 32.16017964071856, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1343, "step": 21483 }, { "epoch": 32.16167664670659, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1283, "step": 21484 }, { "epoch": 32.16317365269461, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1227, "step": 21485 }, { "epoch": 32.16467065868263, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1236, "step": 21486 }, { "epoch": 32.16616766467066, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1237, "step": 21487 }, { "epoch": 32.16766467065868, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1278, "step": 21488 }, { "epoch": 32.16916167664671, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1315, "step": 21489 }, { "epoch": 32.17065868263473, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1242, "step": 21490 }, { "epoch": 32.17215568862275, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1266, "step": 21491 }, { "epoch": 32.17365269461078, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1264, "step": 21492 }, { "epoch": 32.1751497005988, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1215, "step": 21493 }, { "epoch": 32.17664670658683, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1276, "step": 21494 }, { "epoch": 32.17814371257485, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.124, "step": 21495 }, { "epoch": 32.17964071856287, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1308, "step": 21496 }, { "epoch": 32.1811377245509, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1224, "step": 21497 }, { "epoch": 32.18263473053892, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1227, "step": 21498 }, { "epoch": 32.18413173652694, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1259, "step": 21499 }, { "epoch": 32.18562874251497, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1309, "step": 21500 }, { "epoch": 32.18712574850299, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1274, "step": 21501 }, { "epoch": 32.18862275449102, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1265, "step": 21502 }, { "epoch": 32.19011976047904, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1241, "step": 21503 }, { "epoch": 32.191616766467064, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1269, "step": 21504 }, { "epoch": 32.19311377245509, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1201, "step": 21505 }, { "epoch": 32.19461077844311, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1256, "step": 21506 }, { "epoch": 32.196107784431135, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 21507 }, { "epoch": 32.19760479041916, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1271, "step": 21508 }, { "epoch": 32.199101796407184, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1244, "step": 21509 }, { "epoch": 32.20059880239521, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1248, "step": 21510 }, { "epoch": 32.202095808383234, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1288, "step": 21511 }, { "epoch": 32.203592814371255, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1268, "step": 21512 }, { "epoch": 32.205089820359284, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1265, "step": 21513 }, { "epoch": 32.206586826347305, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1263, "step": 21514 }, { "epoch": 32.208083832335326, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1256, "step": 21515 }, { "epoch": 32.209580838323355, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1178, "step": 21516 }, { "epoch": 32.211077844311376, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1232, "step": 21517 }, { "epoch": 32.212574850299404, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1317, "step": 21518 }, { "epoch": 32.214071856287426, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.127, "step": 21519 }, { "epoch": 32.21556886227545, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1306, "step": 21520 }, { "epoch": 32.217065868263475, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1337, "step": 21521 }, { "epoch": 32.2185628742515, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1267, "step": 21522 }, { "epoch": 32.22005988023952, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1295, "step": 21523 }, { "epoch": 32.221556886227546, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 21524 }, { "epoch": 32.22305389221557, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1264, "step": 21525 }, { "epoch": 32.224550898203596, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1302, "step": 21526 }, { "epoch": 32.22604790419162, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1269, "step": 21527 }, { "epoch": 32.22754491017964, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1246, "step": 21528 }, { "epoch": 32.22904191616767, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1271, "step": 21529 }, { "epoch": 32.23053892215569, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1245, "step": 21530 }, { "epoch": 32.23203592814371, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.123, "step": 21531 }, { "epoch": 32.23353293413174, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1298, "step": 21532 }, { "epoch": 32.23502994011976, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1239, "step": 21533 }, { "epoch": 32.23652694610779, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1187, "step": 21534 }, { "epoch": 32.23802395209581, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1249, "step": 21535 }, { "epoch": 32.23952095808383, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 21536 }, { "epoch": 32.24101796407186, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1257, "step": 21537 }, { "epoch": 32.24251497005988, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1314, "step": 21538 }, { "epoch": 32.2440119760479, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.132, "step": 21539 }, { "epoch": 32.24550898203593, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1186, "step": 21540 }, { "epoch": 32.24700598802395, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1235, "step": 21541 }, { "epoch": 32.24850299401198, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1242, "step": 21542 }, { "epoch": 32.25, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1262, "step": 21543 }, { "epoch": 32.25149700598802, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1297, "step": 21544 }, { "epoch": 32.25299401197605, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1251, "step": 21545 }, { "epoch": 32.25449101796407, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.128, "step": 21546 }, { "epoch": 32.2559880239521, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1273, "step": 21547 }, { "epoch": 32.25748502994012, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1295, "step": 21548 }, { "epoch": 32.25898203592814, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1259, "step": 21549 }, { "epoch": 32.26047904191617, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1216, "step": 21550 }, { "epoch": 32.26197604790419, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1292, "step": 21551 }, { "epoch": 32.26347305389221, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1272, "step": 21552 }, { "epoch": 32.26497005988024, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1251, "step": 21553 }, { "epoch": 32.26646706586826, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1321, "step": 21554 }, { "epoch": 32.26796407185629, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1264, "step": 21555 }, { "epoch": 32.26946107784431, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1225, "step": 21556 }, { "epoch": 32.27095808383233, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1208, "step": 21557 }, { "epoch": 32.27245508982036, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1237, "step": 21558 }, { "epoch": 32.27395209580838, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1191, "step": 21559 }, { "epoch": 32.275449101796404, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1356, "step": 21560 }, { "epoch": 32.27694610778443, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 21561 }, { "epoch": 32.278443113772454, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1272, "step": 21562 }, { "epoch": 32.27994011976048, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1295, "step": 21563 }, { "epoch": 32.2814371257485, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1297, "step": 21564 }, { "epoch": 32.282934131736525, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1239, "step": 21565 }, { "epoch": 32.28443113772455, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1232, "step": 21566 }, { "epoch": 32.285928143712574, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1255, "step": 21567 }, { "epoch": 32.287425149700596, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1239, "step": 21568 }, { "epoch": 32.288922155688624, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1253, "step": 21569 }, { "epoch": 32.290419161676645, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.126, "step": 21570 }, { "epoch": 32.291916167664674, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1243, "step": 21571 }, { "epoch": 32.293413173652695, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1306, "step": 21572 }, { "epoch": 32.294910179640716, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1277, "step": 21573 }, { "epoch": 32.296407185628745, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1195, "step": 21574 }, { "epoch": 32.297904191616766, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1317, "step": 21575 }, { "epoch": 32.29940119760479, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1198, "step": 21576 }, { "epoch": 32.300898203592816, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1294, "step": 21577 }, { "epoch": 32.30239520958084, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1222, "step": 21578 }, { "epoch": 32.303892215568865, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1175, "step": 21579 }, { "epoch": 32.30538922155689, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1233, "step": 21580 }, { "epoch": 32.30688622754491, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1344, "step": 21581 }, { "epoch": 32.308383233532936, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1285, "step": 21582 }, { "epoch": 32.30988023952096, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1282, "step": 21583 }, { "epoch": 32.31137724550898, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1242, "step": 21584 }, { "epoch": 32.31287425149701, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1273, "step": 21585 }, { "epoch": 32.31437125748503, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1322, "step": 21586 }, { "epoch": 32.31586826347306, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1324, "step": 21587 }, { "epoch": 32.31736526946108, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1321, "step": 21588 }, { "epoch": 32.3188622754491, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1275, "step": 21589 }, { "epoch": 32.32035928143713, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1267, "step": 21590 }, { "epoch": 32.32185628742515, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1324, "step": 21591 }, { "epoch": 32.32335329341317, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1273, "step": 21592 }, { "epoch": 32.3248502994012, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1271, "step": 21593 }, { "epoch": 32.32634730538922, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1294, "step": 21594 }, { "epoch": 32.32784431137725, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 21595 }, { "epoch": 32.32934131736527, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1301, "step": 21596 }, { "epoch": 32.33083832335329, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1279, "step": 21597 }, { "epoch": 32.33233532934132, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1322, "step": 21598 }, { "epoch": 32.33383233532934, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1233, "step": 21599 }, { "epoch": 32.33532934131736, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1217, "step": 21600 }, { "epoch": 32.33682634730539, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.122, "step": 21601 }, { "epoch": 32.33832335329341, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1313, "step": 21602 }, { "epoch": 32.33982035928144, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1367, "step": 21603 }, { "epoch": 32.34131736526946, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1269, "step": 21604 }, { "epoch": 32.34281437125748, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1308, "step": 21605 }, { "epoch": 32.34431137724551, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.13, "step": 21606 }, { "epoch": 32.34580838323353, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1229, "step": 21607 }, { "epoch": 32.34730538922156, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1262, "step": 21608 }, { "epoch": 32.34880239520958, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1233, "step": 21609 }, { "epoch": 32.3502994011976, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1257, "step": 21610 }, { "epoch": 32.35179640718563, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.1269, "step": 21611 }, { "epoch": 32.35329341317365, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1211, "step": 21612 }, { "epoch": 32.354790419161674, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1251, "step": 21613 }, { "epoch": 32.3562874251497, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 21614 }, { "epoch": 32.35778443113772, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.138, "step": 21615 }, { "epoch": 32.35928143712575, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1222, "step": 21616 }, { "epoch": 32.36077844311377, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1268, "step": 21617 }, { "epoch": 32.362275449101794, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1245, "step": 21618 }, { "epoch": 32.36377245508982, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1226, "step": 21619 }, { "epoch": 32.365269461077844, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.129, "step": 21620 }, { "epoch": 32.366766467065865, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1206, "step": 21621 }, { "epoch": 32.368263473053894, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1293, "step": 21622 }, { "epoch": 32.369760479041915, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1288, "step": 21623 }, { "epoch": 32.37125748502994, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1226, "step": 21624 }, { "epoch": 32.372754491017965, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1277, "step": 21625 }, { "epoch": 32.374251497005986, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1285, "step": 21626 }, { "epoch": 32.375748502994014, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1226, "step": 21627 }, { "epoch": 32.377245508982035, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.129, "step": 21628 }, { "epoch": 32.37874251497006, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1261, "step": 21629 }, { "epoch": 32.380239520958085, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1345, "step": 21630 }, { "epoch": 32.381736526946106, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 21631 }, { "epoch": 32.383233532934135, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1221, "step": 21632 }, { "epoch": 32.384730538922156, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1287, "step": 21633 }, { "epoch": 32.38622754491018, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1283, "step": 21634 }, { "epoch": 32.387724550898206, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1252, "step": 21635 }, { "epoch": 32.38922155688623, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.125, "step": 21636 }, { "epoch": 32.39071856287425, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1309, "step": 21637 }, { "epoch": 32.39221556886228, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1266, "step": 21638 }, { "epoch": 32.3937125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1248, "step": 21639 }, { "epoch": 32.395209580838326, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.119, "step": 21640 }, { "epoch": 32.39670658682635, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1291, "step": 21641 }, { "epoch": 32.39820359281437, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1253, "step": 21642 }, { "epoch": 32.3997005988024, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1256, "step": 21643 }, { "epoch": 32.40119760479042, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1278, "step": 21644 }, { "epoch": 32.40269461077844, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1264, "step": 21645 }, { "epoch": 32.40419161676647, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.12, "step": 21646 }, { "epoch": 32.40568862275449, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1225, "step": 21647 }, { "epoch": 32.40718562874252, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 21648 }, { "epoch": 32.40868263473054, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1266, "step": 21649 }, { "epoch": 32.41017964071856, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1267, "step": 21650 }, { "epoch": 32.41167664670659, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1285, "step": 21651 }, { "epoch": 32.41317365269461, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1273, "step": 21652 }, { "epoch": 32.41467065868264, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1242, "step": 21653 }, { "epoch": 32.41616766467066, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1282, "step": 21654 }, { "epoch": 32.41766467065868, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1274, "step": 21655 }, { "epoch": 32.41916167664671, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1182, "step": 21656 }, { "epoch": 32.42065868263473, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1315, "step": 21657 }, { "epoch": 32.42215568862275, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.132, "step": 21658 }, { "epoch": 32.42365269461078, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1256, "step": 21659 }, { "epoch": 32.4251497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1163, "step": 21660 }, { "epoch": 32.42664670658683, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1259, "step": 21661 }, { "epoch": 32.42814371257485, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1267, "step": 21662 }, { "epoch": 32.42964071856287, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1286, "step": 21663 }, { "epoch": 32.4311377245509, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1282, "step": 21664 }, { "epoch": 32.43263473053892, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1297, "step": 21665 }, { "epoch": 32.43413173652694, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1259, "step": 21666 }, { "epoch": 32.43562874251497, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1278, "step": 21667 }, { "epoch": 32.43712574850299, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1276, "step": 21668 }, { "epoch": 32.43862275449102, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1206, "step": 21669 }, { "epoch": 32.44011976047904, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 21670 }, { "epoch": 32.441616766467064, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 21671 }, { "epoch": 32.44311377245509, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1303, "step": 21672 }, { "epoch": 32.44461077844311, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1259, "step": 21673 }, { "epoch": 32.446107784431135, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1219, "step": 21674 }, { "epoch": 32.44760479041916, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1308, "step": 21675 }, { "epoch": 32.449101796407184, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1252, "step": 21676 }, { "epoch": 32.45059880239521, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1285, "step": 21677 }, { "epoch": 32.452095808383234, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1278, "step": 21678 }, { "epoch": 32.453592814371255, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1216, "step": 21679 }, { "epoch": 32.455089820359284, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1197, "step": 21680 }, { "epoch": 32.456586826347305, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1234, "step": 21681 }, { "epoch": 32.458083832335326, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1328, "step": 21682 }, { "epoch": 32.459580838323355, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1295, "step": 21683 }, { "epoch": 32.461077844311376, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1319, "step": 21684 }, { "epoch": 32.462574850299404, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.126, "step": 21685 }, { "epoch": 32.464071856287426, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1281, "step": 21686 }, { "epoch": 32.46556886227545, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1221, "step": 21687 }, { "epoch": 32.467065868263475, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1264, "step": 21688 }, { "epoch": 32.4685628742515, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1215, "step": 21689 }, { "epoch": 32.47005988023952, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1259, "step": 21690 }, { "epoch": 32.471556886227546, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 21691 }, { "epoch": 32.47305389221557, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1259, "step": 21692 }, { "epoch": 32.474550898203596, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1257, "step": 21693 }, { "epoch": 32.47604790419162, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 21694 }, { "epoch": 32.47754491017964, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1278, "step": 21695 }, { "epoch": 32.47904191616767, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1263, "step": 21696 }, { "epoch": 32.48053892215569, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1186, "step": 21697 }, { "epoch": 32.48203592814371, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1252, "step": 21698 }, { "epoch": 32.48353293413174, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1282, "step": 21699 }, { "epoch": 32.48502994011976, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1235, "step": 21700 }, { "epoch": 32.48652694610779, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1214, "step": 21701 }, { "epoch": 32.48802395209581, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.131, "step": 21702 }, { "epoch": 32.48952095808383, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1262, "step": 21703 }, { "epoch": 32.49101796407186, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.131, "step": 21704 }, { "epoch": 32.49251497005988, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1276, "step": 21705 }, { "epoch": 32.4940119760479, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1262, "step": 21706 }, { "epoch": 32.49550898203593, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1254, "step": 21707 }, { "epoch": 32.49700598802395, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 21708 }, { "epoch": 32.49850299401198, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 21709 }, { "epoch": 32.5, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1325, "step": 21710 }, { "epoch": 32.50149700598802, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1219, "step": 21711 }, { "epoch": 32.50299401197605, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1282, "step": 21712 }, { "epoch": 32.50449101796407, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1206, "step": 21713 }, { "epoch": 32.5059880239521, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1285, "step": 21714 }, { "epoch": 32.50748502994012, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1231, "step": 21715 }, { "epoch": 32.50898203592814, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1263, "step": 21716 }, { "epoch": 32.51047904191617, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1294, "step": 21717 }, { "epoch": 32.51197604790419, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1332, "step": 21718 }, { "epoch": 32.51347305389221, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1233, "step": 21719 }, { "epoch": 32.51497005988024, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1272, "step": 21720 }, { "epoch": 32.51646706586826, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1283, "step": 21721 }, { "epoch": 32.51796407185629, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.131, "step": 21722 }, { "epoch": 32.51946107784431, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 21723 }, { "epoch": 32.52095808383233, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1307, "step": 21724 }, { "epoch": 32.52245508982036, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1258, "step": 21725 }, { "epoch": 32.52395209580838, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1225, "step": 21726 }, { "epoch": 32.525449101796404, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1308, "step": 21727 }, { "epoch": 32.52694610778443, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1312, "step": 21728 }, { "epoch": 32.528443113772454, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1283, "step": 21729 }, { "epoch": 32.52994011976048, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1275, "step": 21730 }, { "epoch": 32.5314371257485, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1269, "step": 21731 }, { "epoch": 32.532934131736525, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1265, "step": 21732 }, { "epoch": 32.53443113772455, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1252, "step": 21733 }, { "epoch": 32.535928143712574, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1274, "step": 21734 }, { "epoch": 32.537425149700596, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1259, "step": 21735 }, { "epoch": 32.538922155688624, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1328, "step": 21736 }, { "epoch": 32.540419161676645, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1235, "step": 21737 }, { "epoch": 32.541916167664674, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1354, "step": 21738 }, { "epoch": 32.543413173652695, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1304, "step": 21739 }, { "epoch": 32.544910179640716, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1231, "step": 21740 }, { "epoch": 32.546407185628745, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1274, "step": 21741 }, { "epoch": 32.547904191616766, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1275, "step": 21742 }, { "epoch": 32.54940119760479, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1288, "step": 21743 }, { "epoch": 32.550898203592816, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1247, "step": 21744 }, { "epoch": 32.55239520958084, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1262, "step": 21745 }, { "epoch": 32.553892215568865, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1267, "step": 21746 }, { "epoch": 32.55538922155689, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1231, "step": 21747 }, { "epoch": 32.55688622754491, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1235, "step": 21748 }, { "epoch": 32.558383233532936, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1294, "step": 21749 }, { "epoch": 32.55988023952096, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1266, "step": 21750 }, { "epoch": 32.56137724550898, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1239, "step": 21751 }, { "epoch": 32.56287425149701, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1313, "step": 21752 }, { "epoch": 32.56437125748503, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1298, "step": 21753 }, { "epoch": 32.56586826347306, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1253, "step": 21754 }, { "epoch": 32.56736526946108, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1215, "step": 21755 }, { "epoch": 32.5688622754491, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 21756 }, { "epoch": 32.57035928143713, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1312, "step": 21757 }, { "epoch": 32.57185628742515, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1307, "step": 21758 }, { "epoch": 32.57335329341317, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1307, "step": 21759 }, { "epoch": 32.5748502994012, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1267, "step": 21760 }, { "epoch": 32.57634730538922, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1318, "step": 21761 }, { "epoch": 32.57784431137725, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1226, "step": 21762 }, { "epoch": 32.57934131736527, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1279, "step": 21763 }, { "epoch": 32.58083832335329, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1225, "step": 21764 }, { "epoch": 32.58233532934132, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1339, "step": 21765 }, { "epoch": 32.58383233532934, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1272, "step": 21766 }, { "epoch": 32.58532934131736, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1277, "step": 21767 }, { "epoch": 32.58682634730539, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1199, "step": 21768 }, { "epoch": 32.58832335329341, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1316, "step": 21769 }, { "epoch": 32.58982035928144, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1287, "step": 21770 }, { "epoch": 32.59131736526946, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1291, "step": 21771 }, { "epoch": 32.59281437125748, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1288, "step": 21772 }, { "epoch": 32.59431137724551, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 21773 }, { "epoch": 32.59580838323353, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1294, "step": 21774 }, { "epoch": 32.59730538922156, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1307, "step": 21775 }, { "epoch": 32.59880239520958, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1281, "step": 21776 }, { "epoch": 32.6002994011976, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1224, "step": 21777 }, { "epoch": 32.60179640718563, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1209, "step": 21778 }, { "epoch": 32.60329341317365, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1249, "step": 21779 }, { "epoch": 32.604790419161674, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1222, "step": 21780 }, { "epoch": 32.6062874251497, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1308, "step": 21781 }, { "epoch": 32.60778443113772, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.131, "step": 21782 }, { "epoch": 32.60928143712575, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.122, "step": 21783 }, { "epoch": 32.61077844311377, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1243, "step": 21784 }, { "epoch": 32.612275449101794, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1204, "step": 21785 }, { "epoch": 32.61377245508982, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1261, "step": 21786 }, { "epoch": 32.615269461077844, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1292, "step": 21787 }, { "epoch": 32.616766467065865, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1239, "step": 21788 }, { "epoch": 32.618263473053894, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1318, "step": 21789 }, { "epoch": 32.619760479041915, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1284, "step": 21790 }, { "epoch": 32.62125748502994, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1272, "step": 21791 }, { "epoch": 32.622754491017965, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1223, "step": 21792 }, { "epoch": 32.624251497005986, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1243, "step": 21793 }, { "epoch": 32.625748502994014, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1264, "step": 21794 }, { "epoch": 32.627245508982035, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1272, "step": 21795 }, { "epoch": 32.62874251497006, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1264, "step": 21796 }, { "epoch": 32.630239520958085, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1257, "step": 21797 }, { "epoch": 32.631736526946106, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1297, "step": 21798 }, { "epoch": 32.633233532934135, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1228, "step": 21799 }, { "epoch": 32.634730538922156, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1306, "step": 21800 }, { "epoch": 32.63622754491018, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1249, "step": 21801 }, { "epoch": 32.637724550898206, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1242, "step": 21802 }, { "epoch": 32.63922155688623, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1362, "step": 21803 }, { "epoch": 32.64071856287425, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1212, "step": 21804 }, { "epoch": 32.64221556886228, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1333, "step": 21805 }, { "epoch": 32.6437125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1279, "step": 21806 }, { "epoch": 32.645209580838326, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.131, "step": 21807 }, { "epoch": 32.64670658682635, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1251, "step": 21808 }, { "epoch": 32.64820359281437, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1259, "step": 21809 }, { "epoch": 32.6497005988024, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1252, "step": 21810 }, { "epoch": 32.65119760479042, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1287, "step": 21811 }, { "epoch": 32.65269461077844, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1315, "step": 21812 }, { "epoch": 32.65419161676647, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1264, "step": 21813 }, { "epoch": 32.65568862275449, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1258, "step": 21814 }, { "epoch": 32.65718562874252, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1292, "step": 21815 }, { "epoch": 32.65868263473054, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1283, "step": 21816 }, { "epoch": 32.66017964071856, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1262, "step": 21817 }, { "epoch": 32.66167664670659, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1247, "step": 21818 }, { "epoch": 32.66317365269461, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1236, "step": 21819 }, { "epoch": 32.66467065868264, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1301, "step": 21820 }, { "epoch": 32.66616766467066, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1288, "step": 21821 }, { "epoch": 32.66766467065868, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1237, "step": 21822 }, { "epoch": 32.66916167664671, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1231, "step": 21823 }, { "epoch": 32.67065868263473, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1218, "step": 21824 }, { "epoch": 32.67215568862275, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 21825 }, { "epoch": 32.67365269461078, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1255, "step": 21826 }, { "epoch": 32.6751497005988, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1292, "step": 21827 }, { "epoch": 32.67664670658683, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1333, "step": 21828 }, { "epoch": 32.67814371257485, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1276, "step": 21829 }, { "epoch": 32.67964071856287, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1233, "step": 21830 }, { "epoch": 32.6811377245509, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1393, "step": 21831 }, { "epoch": 32.68263473053892, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1217, "step": 21832 }, { "epoch": 32.68413173652694, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1351, "step": 21833 }, { "epoch": 32.68562874251497, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1307, "step": 21834 }, { "epoch": 32.68712574850299, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1222, "step": 21835 }, { "epoch": 32.68862275449102, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1244, "step": 21836 }, { "epoch": 32.69011976047904, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.123, "step": 21837 }, { "epoch": 32.691616766467064, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 21838 }, { "epoch": 32.69311377245509, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1254, "step": 21839 }, { "epoch": 32.69461077844311, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1314, "step": 21840 }, { "epoch": 32.696107784431135, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1248, "step": 21841 }, { "epoch": 32.69760479041916, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1269, "step": 21842 }, { "epoch": 32.699101796407184, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1324, "step": 21843 }, { "epoch": 32.70059880239521, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.128, "step": 21844 }, { "epoch": 32.702095808383234, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1276, "step": 21845 }, { "epoch": 32.703592814371255, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1235, "step": 21846 }, { "epoch": 32.705089820359284, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 21847 }, { "epoch": 32.706586826347305, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1275, "step": 21848 }, { "epoch": 32.708083832335326, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.129, "step": 21849 }, { "epoch": 32.709580838323355, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1227, "step": 21850 }, { "epoch": 32.711077844311376, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1302, "step": 21851 }, { "epoch": 32.712574850299404, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1277, "step": 21852 }, { "epoch": 32.714071856287426, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1257, "step": 21853 }, { "epoch": 32.71556886227545, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1278, "step": 21854 }, { "epoch": 32.717065868263475, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1277, "step": 21855 }, { "epoch": 32.7185628742515, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1228, "step": 21856 }, { "epoch": 32.72005988023952, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1309, "step": 21857 }, { "epoch": 32.721556886227546, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1318, "step": 21858 }, { "epoch": 32.72305389221557, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1245, "step": 21859 }, { "epoch": 32.724550898203596, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1271, "step": 21860 }, { "epoch": 32.72604790419162, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1261, "step": 21861 }, { "epoch": 32.72754491017964, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1244, "step": 21862 }, { "epoch": 32.72904191616767, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1263, "step": 21863 }, { "epoch": 32.73053892215569, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.131, "step": 21864 }, { "epoch": 32.73203592814371, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1263, "step": 21865 }, { "epoch": 32.73353293413174, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1197, "step": 21866 }, { "epoch": 32.73502994011976, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1182, "step": 21867 }, { "epoch": 32.73652694610779, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.128, "step": 21868 }, { "epoch": 32.73802395209581, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1275, "step": 21869 }, { "epoch": 32.73952095808383, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1235, "step": 21870 }, { "epoch": 32.74101796407186, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1187, "step": 21871 }, { "epoch": 32.74251497005988, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1213, "step": 21872 }, { "epoch": 32.7440119760479, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1248, "step": 21873 }, { "epoch": 32.74550898203593, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1256, "step": 21874 }, { "epoch": 32.74700598802395, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1326, "step": 21875 }, { "epoch": 32.74850299401198, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1218, "step": 21876 }, { "epoch": 32.75, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1262, "step": 21877 }, { "epoch": 32.75149700598802, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1325, "step": 21878 }, { "epoch": 32.75299401197605, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1284, "step": 21879 }, { "epoch": 32.75449101796407, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1298, "step": 21880 }, { "epoch": 32.7559880239521, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1286, "step": 21881 }, { "epoch": 32.75748502994012, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1326, "step": 21882 }, { "epoch": 32.75898203592814, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1329, "step": 21883 }, { "epoch": 32.76047904191617, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1231, "step": 21884 }, { "epoch": 32.76197604790419, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1318, "step": 21885 }, { "epoch": 32.76347305389221, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1252, "step": 21886 }, { "epoch": 32.76497005988024, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1196, "step": 21887 }, { "epoch": 32.76646706586826, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1258, "step": 21888 }, { "epoch": 32.76796407185629, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 21889 }, { "epoch": 32.76946107784431, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1225, "step": 21890 }, { "epoch": 32.77095808383233, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1256, "step": 21891 }, { "epoch": 32.77245508982036, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1275, "step": 21892 }, { "epoch": 32.77395209580838, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1175, "step": 21893 }, { "epoch": 32.775449101796404, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1322, "step": 21894 }, { "epoch": 32.77694610778443, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.125, "step": 21895 }, { "epoch": 32.778443113772454, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1291, "step": 21896 }, { "epoch": 32.77994011976048, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 21897 }, { "epoch": 32.7814371257485, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1272, "step": 21898 }, { "epoch": 32.782934131736525, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1254, "step": 21899 }, { "epoch": 32.78443113772455, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1307, "step": 21900 }, { "epoch": 32.785928143712574, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.127, "step": 21901 }, { "epoch": 32.787425149700596, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1266, "step": 21902 }, { "epoch": 32.788922155688624, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1316, "step": 21903 }, { "epoch": 32.790419161676645, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1222, "step": 21904 }, { "epoch": 32.791916167664674, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1233, "step": 21905 }, { "epoch": 32.793413173652695, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1253, "step": 21906 }, { "epoch": 32.794910179640716, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 21907 }, { "epoch": 32.796407185628745, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1261, "step": 21908 }, { "epoch": 32.797904191616766, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1242, "step": 21909 }, { "epoch": 32.79940119760479, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1314, "step": 21910 }, { "epoch": 32.800898203592816, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1263, "step": 21911 }, { "epoch": 32.80239520958084, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1255, "step": 21912 }, { "epoch": 32.803892215568865, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1243, "step": 21913 }, { "epoch": 32.80538922155689, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1261, "step": 21914 }, { "epoch": 32.80688622754491, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1224, "step": 21915 }, { "epoch": 32.808383233532936, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1251, "step": 21916 }, { "epoch": 32.80988023952096, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1277, "step": 21917 }, { "epoch": 32.81137724550898, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.118, "step": 21918 }, { "epoch": 32.81287425149701, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1243, "step": 21919 }, { "epoch": 32.81437125748503, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 21920 }, { "epoch": 32.81586826347306, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1277, "step": 21921 }, { "epoch": 32.81736526946108, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1268, "step": 21922 }, { "epoch": 32.8188622754491, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.13, "step": 21923 }, { "epoch": 32.82035928143713, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1216, "step": 21924 }, { "epoch": 32.82185628742515, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1251, "step": 21925 }, { "epoch": 32.82335329341317, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1221, "step": 21926 }, { "epoch": 32.8248502994012, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1266, "step": 21927 }, { "epoch": 32.82634730538922, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.124, "step": 21928 }, { "epoch": 32.82784431137725, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1239, "step": 21929 }, { "epoch": 32.82934131736527, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1263, "step": 21930 }, { "epoch": 32.83083832335329, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1272, "step": 21931 }, { "epoch": 32.83233532934132, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.13, "step": 21932 }, { "epoch": 32.83383233532934, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.129, "step": 21933 }, { "epoch": 32.83532934131736, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1214, "step": 21934 }, { "epoch": 32.83682634730539, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1269, "step": 21935 }, { "epoch": 32.83832335329341, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.129, "step": 21936 }, { "epoch": 32.83982035928144, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1212, "step": 21937 }, { "epoch": 32.84131736526946, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1259, "step": 21938 }, { "epoch": 32.84281437125748, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1311, "step": 21939 }, { "epoch": 32.84431137724551, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 21940 }, { "epoch": 32.84580838323353, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1285, "step": 21941 }, { "epoch": 32.84730538922156, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1276, "step": 21942 }, { "epoch": 32.84880239520958, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1198, "step": 21943 }, { "epoch": 32.8502994011976, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1269, "step": 21944 }, { "epoch": 32.85179640718563, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1152, "step": 21945 }, { "epoch": 32.85329341317365, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1263, "step": 21946 }, { "epoch": 32.854790419161674, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1279, "step": 21947 }, { "epoch": 32.8562874251497, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1294, "step": 21948 }, { "epoch": 32.85778443113772, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1315, "step": 21949 }, { "epoch": 32.85928143712575, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.126, "step": 21950 }, { "epoch": 32.86077844311377, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1268, "step": 21951 }, { "epoch": 32.862275449101794, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1217, "step": 21952 }, { "epoch": 32.86377245508982, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1219, "step": 21953 }, { "epoch": 32.865269461077844, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1155, "step": 21954 }, { "epoch": 32.866766467065865, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1282, "step": 21955 }, { "epoch": 32.868263473053894, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1302, "step": 21956 }, { "epoch": 32.869760479041915, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1229, "step": 21957 }, { "epoch": 32.87125748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1239, "step": 21958 }, { "epoch": 32.872754491017965, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1342, "step": 21959 }, { "epoch": 32.874251497005986, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1219, "step": 21960 }, { "epoch": 32.875748502994014, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1203, "step": 21961 }, { "epoch": 32.877245508982035, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1292, "step": 21962 }, { "epoch": 32.87874251497006, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1276, "step": 21963 }, { "epoch": 32.880239520958085, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.124, "step": 21964 }, { "epoch": 32.881736526946106, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1276, "step": 21965 }, { "epoch": 32.883233532934135, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1306, "step": 21966 }, { "epoch": 32.884730538922156, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.126, "step": 21967 }, { "epoch": 32.88622754491018, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.122, "step": 21968 }, { "epoch": 32.887724550898206, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1313, "step": 21969 }, { "epoch": 32.88922155688623, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.127, "step": 21970 }, { "epoch": 32.89071856287425, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1239, "step": 21971 }, { "epoch": 32.89221556886228, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1228, "step": 21972 }, { "epoch": 32.8937125748503, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1278, "step": 21973 }, { "epoch": 32.895209580838326, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1243, "step": 21974 }, { "epoch": 32.89670658682635, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.126, "step": 21975 }, { "epoch": 32.89820359281437, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1198, "step": 21976 }, { "epoch": 32.8997005988024, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1265, "step": 21977 }, { "epoch": 32.90119760479042, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 21978 }, { "epoch": 32.90269461077844, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1222, "step": 21979 }, { "epoch": 32.90419161676647, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1289, "step": 21980 }, { "epoch": 32.90568862275449, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1287, "step": 21981 }, { "epoch": 32.90718562874252, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1283, "step": 21982 }, { "epoch": 32.90868263473054, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.122, "step": 21983 }, { "epoch": 32.91017964071856, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1278, "step": 21984 }, { "epoch": 32.91167664670659, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.127, "step": 21985 }, { "epoch": 32.91317365269461, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1281, "step": 21986 }, { "epoch": 32.91467065868264, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1242, "step": 21987 }, { "epoch": 32.91616766467066, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.123, "step": 21988 }, { "epoch": 32.91766467065868, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.124, "step": 21989 }, { "epoch": 32.91916167664671, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1267, "step": 21990 }, { "epoch": 32.92065868263473, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1301, "step": 21991 }, { "epoch": 32.92215568862275, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1284, "step": 21992 }, { "epoch": 32.92365269461078, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.126, "step": 21993 }, { "epoch": 32.9251497005988, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1248, "step": 21994 }, { "epoch": 32.92664670658683, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1219, "step": 21995 }, { "epoch": 32.92814371257485, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1283, "step": 21996 }, { "epoch": 32.92964071856287, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1317, "step": 21997 }, { "epoch": 32.9311377245509, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1254, "step": 21998 }, { "epoch": 32.93263473053892, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.128, "step": 21999 }, { "epoch": 32.93413173652694, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1286, "step": 22000 }, { "epoch": 32.93562874251497, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.1344, "step": 22001 }, { "epoch": 32.93712574850299, "grad_norm": 0.6484375, "learning_rate": 0.0008, "loss": 1.1871, "step": 22002 }, { "epoch": 32.93862275449102, "grad_norm": 0.65625, "learning_rate": 0.0008, "loss": 1.2136, "step": 22003 }, { "epoch": 32.94011976047904, "grad_norm": 0.921875, "learning_rate": 0.0008, "loss": 1.2172, "step": 22004 }, { "epoch": 32.941616766467064, "grad_norm": 1.9140625, "learning_rate": 0.0008, "loss": 1.3134, "step": 22005 }, { "epoch": 32.94311377245509, "grad_norm": 0.5625, "learning_rate": 0.0008, "loss": 1.2481, "step": 22006 }, { "epoch": 32.94461077844311, "grad_norm": 0.60546875, "learning_rate": 0.0008, "loss": 1.2431, "step": 22007 }, { "epoch": 32.946107784431135, "grad_norm": 0.83984375, "learning_rate": 0.0008, "loss": 1.2953, "step": 22008 }, { "epoch": 32.94760479041916, "grad_norm": 0.54296875, "learning_rate": 0.0008, "loss": 1.2535, "step": 22009 }, { "epoch": 32.949101796407184, "grad_norm": 0.5703125, "learning_rate": 0.0008, "loss": 1.2355, "step": 22010 }, { "epoch": 32.95059880239521, "grad_norm": 1.0234375, "learning_rate": 0.0008, "loss": 1.2542, "step": 22011 }, { "epoch": 32.952095808383234, "grad_norm": 0.84765625, "learning_rate": 0.0008, "loss": 1.2735, "step": 22012 }, { "epoch": 32.953592814371255, "grad_norm": 0.8203125, "learning_rate": 0.0008, "loss": 1.2561, "step": 22013 }, { "epoch": 32.955089820359284, "grad_norm": 1.09375, "learning_rate": 0.0008, "loss": 1.2592, "step": 22014 }, { "epoch": 32.956586826347305, "grad_norm": 0.9453125, "learning_rate": 0.0008, "loss": 1.2632, "step": 22015 }, { "epoch": 32.958083832335326, "grad_norm": 0.54296875, "learning_rate": 0.0008, "loss": 1.2407, "step": 22016 }, { "epoch": 32.959580838323355, "grad_norm": 0.361328125, "learning_rate": 0.0008, "loss": 1.2329, "step": 22017 }, { "epoch": 32.961077844311376, "grad_norm": 0.376953125, "learning_rate": 0.0008, "loss": 1.2215, "step": 22018 }, { "epoch": 32.962574850299404, "grad_norm": 0.52734375, "learning_rate": 0.0008, "loss": 1.2192, "step": 22019 }, { "epoch": 32.964071856287426, "grad_norm": 0.96484375, "learning_rate": 0.0008, "loss": 1.237, "step": 22020 }, { "epoch": 32.96556886227545, "grad_norm": 0.8515625, "learning_rate": 0.0008, "loss": 1.2716, "step": 22021 }, { "epoch": 32.967065868263475, "grad_norm": 0.6484375, "learning_rate": 0.0008, "loss": 1.2577, "step": 22022 }, { "epoch": 32.9685628742515, "grad_norm": 0.8984375, "learning_rate": 0.0008, "loss": 1.266, "step": 22023 }, { "epoch": 32.97005988023952, "grad_norm": 0.7109375, "learning_rate": 0.0008, "loss": 1.2545, "step": 22024 }, { "epoch": 32.971556886227546, "grad_norm": 0.61328125, "learning_rate": 0.0008, "loss": 1.2452, "step": 22025 }, { "epoch": 32.97305389221557, "grad_norm": 0.32421875, "learning_rate": 0.0008, "loss": 1.2326, "step": 22026 }, { "epoch": 32.974550898203596, "grad_norm": 0.400390625, "learning_rate": 0.0008, "loss": 1.2261, "step": 22027 }, { "epoch": 32.97604790419162, "grad_norm": 0.34375, "learning_rate": 0.0008, "loss": 1.2216, "step": 22028 }, { "epoch": 32.97754491017964, "grad_norm": 0.4765625, "learning_rate": 0.0008, "loss": 1.2279, "step": 22029 }, { "epoch": 32.97904191616767, "grad_norm": 0.5859375, "learning_rate": 0.0008, "loss": 1.2346, "step": 22030 }, { "epoch": 32.98053892215569, "grad_norm": 0.72265625, "learning_rate": 0.0008, "loss": 1.2277, "step": 22031 }, { "epoch": 32.98203592814371, "grad_norm": 0.8125, "learning_rate": 0.0008, "loss": 1.215, "step": 22032 }, { "epoch": 32.98353293413174, "grad_norm": 0.97265625, "learning_rate": 0.0008, "loss": 1.2357, "step": 22033 }, { "epoch": 32.98502994011976, "grad_norm": 0.6328125, "learning_rate": 0.0008, "loss": 1.2344, "step": 22034 }, { "epoch": 32.98652694610779, "grad_norm": 0.44921875, "learning_rate": 0.0008, "loss": 1.2146, "step": 22035 }, { "epoch": 32.98802395209581, "grad_norm": 0.6875, "learning_rate": 0.0008, "loss": 1.2187, "step": 22036 }, { "epoch": 32.98952095808383, "grad_norm": 0.28515625, "learning_rate": 0.0008, "loss": 1.2073, "step": 22037 }, { "epoch": 32.99101796407186, "grad_norm": 0.478515625, "learning_rate": 0.0008, "loss": 1.2128, "step": 22038 }, { "epoch": 32.99251497005988, "grad_norm": 0.546875, "learning_rate": 0.0008, "loss": 1.2062, "step": 22039 }, { "epoch": 32.9940119760479, "grad_norm": 0.416015625, "learning_rate": 0.0008, "loss": 1.2093, "step": 22040 }, { "epoch": 32.99550898203593, "grad_norm": 0.416015625, "learning_rate": 0.0008, "loss": 1.1887, "step": 22041 }, { "epoch": 32.99700598802395, "grad_norm": 0.296875, "learning_rate": 0.0008, "loss": 1.1968, "step": 22042 }, { "epoch": 32.99850299401198, "grad_norm": 0.330078125, "learning_rate": 0.0008, "loss": 1.1904, "step": 22043 }, { "epoch": 33.0, "grad_norm": 0.294921875, "learning_rate": 0.0008, "loss": 1.189, "step": 22044 }, { "epoch": 33.00149700598802, "grad_norm": 0.2197265625, "learning_rate": 0.0008, "loss": 1.1921, "step": 22045 }, { "epoch": 33.00299401197605, "grad_norm": 0.236328125, "learning_rate": 0.0008, "loss": 1.1826, "step": 22046 }, { "epoch": 33.00449101796407, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.1766, "step": 22047 }, { "epoch": 33.0059880239521, "grad_norm": 0.1826171875, "learning_rate": 0.0008, "loss": 1.184, "step": 22048 }, { "epoch": 33.00748502994012, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1781, "step": 22049 }, { "epoch": 33.00898203592814, "grad_norm": 0.216796875, "learning_rate": 0.0008, "loss": 1.1689, "step": 22050 }, { "epoch": 33.01047904191617, "grad_norm": 0.2109375, "learning_rate": 0.0008, "loss": 1.175, "step": 22051 }, { "epoch": 33.01197604790419, "grad_norm": 0.232421875, "learning_rate": 0.0008, "loss": 1.1776, "step": 22052 }, { "epoch": 33.01347305389221, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.1687, "step": 22053 }, { "epoch": 33.01497005988024, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.1611, "step": 22054 }, { "epoch": 33.01646706586826, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1713, "step": 22055 }, { "epoch": 33.01796407185629, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1673, "step": 22056 }, { "epoch": 33.01946107784431, "grad_norm": 0.1572265625, "learning_rate": 0.0008, "loss": 1.1685, "step": 22057 }, { "epoch": 33.02095808383233, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.158, "step": 22058 }, { "epoch": 33.02245508982036, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1592, "step": 22059 }, { "epoch": 33.02395209580838, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1602, "step": 22060 }, { "epoch": 33.025449101796404, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1611, "step": 22061 }, { "epoch": 33.02694610778443, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1536, "step": 22062 }, { "epoch": 33.028443113772454, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1528, "step": 22063 }, { "epoch": 33.02994011976048, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1522, "step": 22064 }, { "epoch": 33.0314371257485, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1548, "step": 22065 }, { "epoch": 33.032934131736525, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.152, "step": 22066 }, { "epoch": 33.03443113772455, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.16, "step": 22067 }, { "epoch": 33.035928143712574, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1532, "step": 22068 }, { "epoch": 33.037425149700596, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1545, "step": 22069 }, { "epoch": 33.038922155688624, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1554, "step": 22070 }, { "epoch": 33.040419161676645, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1455, "step": 22071 }, { "epoch": 33.041916167664674, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1631, "step": 22072 }, { "epoch": 33.043413173652695, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1526, "step": 22073 }, { "epoch": 33.044910179640716, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1504, "step": 22074 }, { "epoch": 33.046407185628745, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1521, "step": 22075 }, { "epoch": 33.047904191616766, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1465, "step": 22076 }, { "epoch": 33.04940119760479, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1589, "step": 22077 }, { "epoch": 33.050898203592816, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1499, "step": 22078 }, { "epoch": 33.05239520958084, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1463, "step": 22079 }, { "epoch": 33.053892215568865, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1461, "step": 22080 }, { "epoch": 33.05538922155689, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1539, "step": 22081 }, { "epoch": 33.05688622754491, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1466, "step": 22082 }, { "epoch": 33.058383233532936, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1508, "step": 22083 }, { "epoch": 33.05988023952096, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1411, "step": 22084 }, { "epoch": 33.06137724550898, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1442, "step": 22085 }, { "epoch": 33.06287425149701, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1416, "step": 22086 }, { "epoch": 33.06437125748503, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1495, "step": 22087 }, { "epoch": 33.06586826347306, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1387, "step": 22088 }, { "epoch": 33.06736526946108, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1462, "step": 22089 }, { "epoch": 33.0688622754491, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1416, "step": 22090 }, { "epoch": 33.07035928143713, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1457, "step": 22091 }, { "epoch": 33.07185628742515, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 22092 }, { "epoch": 33.07335329341317, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1429, "step": 22093 }, { "epoch": 33.0748502994012, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1482, "step": 22094 }, { "epoch": 33.07634730538922, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1421, "step": 22095 }, { "epoch": 33.07784431137725, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1519, "step": 22096 }, { "epoch": 33.07934131736527, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1434, "step": 22097 }, { "epoch": 33.08083832335329, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1451, "step": 22098 }, { "epoch": 33.08233532934132, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1424, "step": 22099 }, { "epoch": 33.08383233532934, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1431, "step": 22100 }, { "epoch": 33.08532934131737, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.146, "step": 22101 }, { "epoch": 33.08682634730539, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1455, "step": 22102 }, { "epoch": 33.08832335329341, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1518, "step": 22103 }, { "epoch": 33.08982035928144, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1408, "step": 22104 }, { "epoch": 33.09131736526946, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1406, "step": 22105 }, { "epoch": 33.09281437125748, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1425, "step": 22106 }, { "epoch": 33.09431137724551, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1404, "step": 22107 }, { "epoch": 33.09580838323353, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1472, "step": 22108 }, { "epoch": 33.09730538922156, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1467, "step": 22109 }, { "epoch": 33.09880239520958, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1442, "step": 22110 }, { "epoch": 33.1002994011976, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1448, "step": 22111 }, { "epoch": 33.10179640718563, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1449, "step": 22112 }, { "epoch": 33.10329341317365, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1409, "step": 22113 }, { "epoch": 33.104790419161674, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1447, "step": 22114 }, { "epoch": 33.1062874251497, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1477, "step": 22115 }, { "epoch": 33.10778443113772, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1504, "step": 22116 }, { "epoch": 33.10928143712575, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1434, "step": 22117 }, { "epoch": 33.11077844311377, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1434, "step": 22118 }, { "epoch": 33.112275449101794, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1436, "step": 22119 }, { "epoch": 33.11377245508982, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1415, "step": 22120 }, { "epoch": 33.115269461077844, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1455, "step": 22121 }, { "epoch": 33.116766467065865, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1535, "step": 22122 }, { "epoch": 33.118263473053894, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1398, "step": 22123 }, { "epoch": 33.119760479041915, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1391, "step": 22124 }, { "epoch": 33.12125748502994, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1417, "step": 22125 }, { "epoch": 33.122754491017965, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1427, "step": 22126 }, { "epoch": 33.124251497005986, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1435, "step": 22127 }, { "epoch": 33.125748502994014, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1403, "step": 22128 }, { "epoch": 33.127245508982035, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1424, "step": 22129 }, { "epoch": 33.12874251497006, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.143, "step": 22130 }, { "epoch": 33.130239520958085, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.147, "step": 22131 }, { "epoch": 33.131736526946106, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1438, "step": 22132 }, { "epoch": 33.133233532934135, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1431, "step": 22133 }, { "epoch": 33.134730538922156, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1477, "step": 22134 }, { "epoch": 33.13622754491018, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1433, "step": 22135 }, { "epoch": 33.137724550898206, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1368, "step": 22136 }, { "epoch": 33.13922155688623, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.14, "step": 22137 }, { "epoch": 33.14071856287425, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1385, "step": 22138 }, { "epoch": 33.14221556886228, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1435, "step": 22139 }, { "epoch": 33.1437125748503, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1351, "step": 22140 }, { "epoch": 33.145209580838326, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.141, "step": 22141 }, { "epoch": 33.14670658682635, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1283, "step": 22142 }, { "epoch": 33.14820359281437, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.1384, "step": 22143 }, { "epoch": 33.1497005988024, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1374, "step": 22144 }, { "epoch": 33.15119760479042, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1463, "step": 22145 }, { "epoch": 33.15269461077844, "grad_norm": 0.2109375, "learning_rate": 0.0008, "loss": 1.146, "step": 22146 }, { "epoch": 33.15419161676647, "grad_norm": 0.2490234375, "learning_rate": 0.0008, "loss": 1.1513, "step": 22147 }, { "epoch": 33.15568862275449, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.1528, "step": 22148 }, { "epoch": 33.15718562874252, "grad_norm": 0.271484375, "learning_rate": 0.0008, "loss": 1.1484, "step": 22149 }, { "epoch": 33.15868263473054, "grad_norm": 0.345703125, "learning_rate": 0.0008, "loss": 1.1435, "step": 22150 }, { "epoch": 33.16017964071856, "grad_norm": 0.498046875, "learning_rate": 0.0008, "loss": 1.1601, "step": 22151 }, { "epoch": 33.16167664670659, "grad_norm": 0.546875, "learning_rate": 0.0008, "loss": 1.1609, "step": 22152 }, { "epoch": 33.16317365269461, "grad_norm": 0.333984375, "learning_rate": 0.0008, "loss": 1.1705, "step": 22153 }, { "epoch": 33.16467065868263, "grad_norm": 0.34375, "learning_rate": 0.0008, "loss": 1.1547, "step": 22154 }, { "epoch": 33.16616766467066, "grad_norm": 0.48828125, "learning_rate": 0.0008, "loss": 1.1497, "step": 22155 }, { "epoch": 33.16766467065868, "grad_norm": 0.333984375, "learning_rate": 0.0008, "loss": 1.1524, "step": 22156 }, { "epoch": 33.16916167664671, "grad_norm": 0.388671875, "learning_rate": 0.0008, "loss": 1.1523, "step": 22157 }, { "epoch": 33.17065868263473, "grad_norm": 0.3984375, "learning_rate": 0.0008, "loss": 1.1466, "step": 22158 }, { "epoch": 33.17215568862275, "grad_norm": 0.326171875, "learning_rate": 0.0008, "loss": 1.165, "step": 22159 }, { "epoch": 33.17365269461078, "grad_norm": 0.189453125, "learning_rate": 0.0008, "loss": 1.1473, "step": 22160 }, { "epoch": 33.1751497005988, "grad_norm": 0.244140625, "learning_rate": 0.0008, "loss": 1.1517, "step": 22161 }, { "epoch": 33.17664670658683, "grad_norm": 0.1826171875, "learning_rate": 0.0008, "loss": 1.1496, "step": 22162 }, { "epoch": 33.17814371257485, "grad_norm": 0.2080078125, "learning_rate": 0.0008, "loss": 1.1516, "step": 22163 }, { "epoch": 33.17964071856287, "grad_norm": 0.185546875, "learning_rate": 0.0008, "loss": 1.1433, "step": 22164 }, { "epoch": 33.1811377245509, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.1488, "step": 22165 }, { "epoch": 33.18263473053892, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.148, "step": 22166 }, { "epoch": 33.18413173652694, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.143, "step": 22167 }, { "epoch": 33.18562874251497, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1414, "step": 22168 }, { "epoch": 33.18712574850299, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1468, "step": 22169 }, { "epoch": 33.18862275449102, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1442, "step": 22170 }, { "epoch": 33.19011976047904, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1446, "step": 22171 }, { "epoch": 33.191616766467064, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1444, "step": 22172 }, { "epoch": 33.19311377245509, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.14, "step": 22173 }, { "epoch": 33.19461077844311, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1468, "step": 22174 }, { "epoch": 33.196107784431135, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1437, "step": 22175 }, { "epoch": 33.19760479041916, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1415, "step": 22176 }, { "epoch": 33.199101796407184, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1443, "step": 22177 }, { "epoch": 33.20059880239521, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1474, "step": 22178 }, { "epoch": 33.202095808383234, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1438, "step": 22179 }, { "epoch": 33.203592814371255, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 22180 }, { "epoch": 33.205089820359284, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1364, "step": 22181 }, { "epoch": 33.206586826347305, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1435, "step": 22182 }, { "epoch": 33.208083832335326, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1411, "step": 22183 }, { "epoch": 33.209580838323355, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1378, "step": 22184 }, { "epoch": 33.211077844311376, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1437, "step": 22185 }, { "epoch": 33.212574850299404, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1399, "step": 22186 }, { "epoch": 33.214071856287426, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 22187 }, { "epoch": 33.21556886227545, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1457, "step": 22188 }, { "epoch": 33.217065868263475, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 22189 }, { "epoch": 33.2185628742515, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1281, "step": 22190 }, { "epoch": 33.22005988023952, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1476, "step": 22191 }, { "epoch": 33.221556886227546, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.149, "step": 22192 }, { "epoch": 33.22305389221557, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1412, "step": 22193 }, { "epoch": 33.224550898203596, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 22194 }, { "epoch": 33.22604790419162, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.142, "step": 22195 }, { "epoch": 33.22754491017964, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1335, "step": 22196 }, { "epoch": 33.22904191616767, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 22197 }, { "epoch": 33.23053892215569, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1359, "step": 22198 }, { "epoch": 33.23203592814371, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1355, "step": 22199 }, { "epoch": 33.23353293413174, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1382, "step": 22200 }, { "epoch": 33.23502994011976, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1374, "step": 22201 }, { "epoch": 33.23652694610779, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1386, "step": 22202 }, { "epoch": 33.23802395209581, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1368, "step": 22203 }, { "epoch": 33.23952095808383, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1458, "step": 22204 }, { "epoch": 33.24101796407186, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1342, "step": 22205 }, { "epoch": 33.24251497005988, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1318, "step": 22206 }, { "epoch": 33.2440119760479, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1374, "step": 22207 }, { "epoch": 33.24550898203593, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1516, "step": 22208 }, { "epoch": 33.24700598802395, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1363, "step": 22209 }, { "epoch": 33.24850299401198, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1336, "step": 22210 }, { "epoch": 33.25, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.139, "step": 22211 }, { "epoch": 33.25149700598802, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1343, "step": 22212 }, { "epoch": 33.25299401197605, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1365, "step": 22213 }, { "epoch": 33.25449101796407, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1416, "step": 22214 }, { "epoch": 33.2559880239521, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.136, "step": 22215 }, { "epoch": 33.25748502994012, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1428, "step": 22216 }, { "epoch": 33.25898203592814, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1318, "step": 22217 }, { "epoch": 33.26047904191617, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.137, "step": 22218 }, { "epoch": 33.26197604790419, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1358, "step": 22219 }, { "epoch": 33.26347305389221, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1392, "step": 22220 }, { "epoch": 33.26497005988024, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1371, "step": 22221 }, { "epoch": 33.26646706586826, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1418, "step": 22222 }, { "epoch": 33.26796407185629, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1383, "step": 22223 }, { "epoch": 33.26946107784431, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1343, "step": 22224 }, { "epoch": 33.27095808383233, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1356, "step": 22225 }, { "epoch": 33.27245508982036, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1398, "step": 22226 }, { "epoch": 33.27395209580838, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1339, "step": 22227 }, { "epoch": 33.275449101796404, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1399, "step": 22228 }, { "epoch": 33.27694610778443, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1474, "step": 22229 }, { "epoch": 33.278443113772454, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1377, "step": 22230 }, { "epoch": 33.27994011976048, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 22231 }, { "epoch": 33.2814371257485, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1383, "step": 22232 }, { "epoch": 33.282934131736525, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.135, "step": 22233 }, { "epoch": 33.28443113772455, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1409, "step": 22234 }, { "epoch": 33.285928143712574, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1318, "step": 22235 }, { "epoch": 33.287425149700596, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1363, "step": 22236 }, { "epoch": 33.288922155688624, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.138, "step": 22237 }, { "epoch": 33.290419161676645, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1379, "step": 22238 }, { "epoch": 33.291916167664674, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1386, "step": 22239 }, { "epoch": 33.293413173652695, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1353, "step": 22240 }, { "epoch": 33.294910179640716, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1327, "step": 22241 }, { "epoch": 33.296407185628745, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1405, "step": 22242 }, { "epoch": 33.297904191616766, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1347, "step": 22243 }, { "epoch": 33.29940119760479, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.137, "step": 22244 }, { "epoch": 33.300898203592816, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1371, "step": 22245 }, { "epoch": 33.30239520958084, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1328, "step": 22246 }, { "epoch": 33.303892215568865, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 22247 }, { "epoch": 33.30538922155689, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1437, "step": 22248 }, { "epoch": 33.30688622754491, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1382, "step": 22249 }, { "epoch": 33.308383233532936, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1366, "step": 22250 }, { "epoch": 33.30988023952096, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1451, "step": 22251 }, { "epoch": 33.31137724550898, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1271, "step": 22252 }, { "epoch": 33.31287425149701, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1376, "step": 22253 }, { "epoch": 33.31437125748503, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1367, "step": 22254 }, { "epoch": 33.31586826347306, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1348, "step": 22255 }, { "epoch": 33.31736526946108, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1397, "step": 22256 }, { "epoch": 33.3188622754491, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1302, "step": 22257 }, { "epoch": 33.32035928143713, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1447, "step": 22258 }, { "epoch": 33.32185628742515, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1358, "step": 22259 }, { "epoch": 33.32335329341317, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1352, "step": 22260 }, { "epoch": 33.3248502994012, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1384, "step": 22261 }, { "epoch": 33.32634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1399, "step": 22262 }, { "epoch": 33.32784431137725, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1411, "step": 22263 }, { "epoch": 33.32934131736527, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1342, "step": 22264 }, { "epoch": 33.33083832335329, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1404, "step": 22265 }, { "epoch": 33.33233532934132, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.137, "step": 22266 }, { "epoch": 33.33383233532934, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 22267 }, { "epoch": 33.33532934131736, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1473, "step": 22268 }, { "epoch": 33.33682634730539, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1389, "step": 22269 }, { "epoch": 33.33832335329341, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1361, "step": 22270 }, { "epoch": 33.33982035928144, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1348, "step": 22271 }, { "epoch": 33.34131736526946, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1294, "step": 22272 }, { "epoch": 33.34281437125748, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1356, "step": 22273 }, { "epoch": 33.34431137724551, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1321, "step": 22274 }, { "epoch": 33.34580838323353, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1357, "step": 22275 }, { "epoch": 33.34730538922156, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1401, "step": 22276 }, { "epoch": 33.34880239520958, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1332, "step": 22277 }, { "epoch": 33.3502994011976, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1378, "step": 22278 }, { "epoch": 33.35179640718563, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1311, "step": 22279 }, { "epoch": 33.35329341317365, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1387, "step": 22280 }, { "epoch": 33.354790419161674, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 22281 }, { "epoch": 33.3562874251497, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1439, "step": 22282 }, { "epoch": 33.35778443113772, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1307, "step": 22283 }, { "epoch": 33.35928143712575, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1369, "step": 22284 }, { "epoch": 33.36077844311377, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1344, "step": 22285 }, { "epoch": 33.362275449101794, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1332, "step": 22286 }, { "epoch": 33.36377245508982, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.134, "step": 22287 }, { "epoch": 33.365269461077844, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 22288 }, { "epoch": 33.366766467065865, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1393, "step": 22289 }, { "epoch": 33.368263473053894, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1286, "step": 22290 }, { "epoch": 33.369760479041915, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1387, "step": 22291 }, { "epoch": 33.37125748502994, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1332, "step": 22292 }, { "epoch": 33.372754491017965, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1332, "step": 22293 }, { "epoch": 33.374251497005986, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1365, "step": 22294 }, { "epoch": 33.375748502994014, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1362, "step": 22295 }, { "epoch": 33.377245508982035, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1387, "step": 22296 }, { "epoch": 33.37874251497006, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1397, "step": 22297 }, { "epoch": 33.380239520958085, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1347, "step": 22298 }, { "epoch": 33.381736526946106, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1329, "step": 22299 }, { "epoch": 33.383233532934135, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 22300 }, { "epoch": 33.384730538922156, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1306, "step": 22301 }, { "epoch": 33.38622754491018, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1387, "step": 22302 }, { "epoch": 33.387724550898206, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.144, "step": 22303 }, { "epoch": 33.38922155688623, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.127, "step": 22304 }, { "epoch": 33.39071856287425, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1309, "step": 22305 }, { "epoch": 33.39221556886228, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.132, "step": 22306 }, { "epoch": 33.3937125748503, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.135, "step": 22307 }, { "epoch": 33.395209580838326, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1348, "step": 22308 }, { "epoch": 33.39670658682635, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1384, "step": 22309 }, { "epoch": 33.39820359281437, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1341, "step": 22310 }, { "epoch": 33.3997005988024, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1399, "step": 22311 }, { "epoch": 33.40119760479042, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1382, "step": 22312 }, { "epoch": 33.40269461077844, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 22313 }, { "epoch": 33.40419161676647, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1364, "step": 22314 }, { "epoch": 33.40568862275449, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1354, "step": 22315 }, { "epoch": 33.40718562874252, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1412, "step": 22316 }, { "epoch": 33.40868263473054, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1408, "step": 22317 }, { "epoch": 33.41017964071856, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 22318 }, { "epoch": 33.41167664670659, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1341, "step": 22319 }, { "epoch": 33.41317365269461, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1344, "step": 22320 }, { "epoch": 33.41467065868264, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1321, "step": 22321 }, { "epoch": 33.41616766467066, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1346, "step": 22322 }, { "epoch": 33.41766467065868, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 22323 }, { "epoch": 33.41916167664671, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1348, "step": 22324 }, { "epoch": 33.42065868263473, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.136, "step": 22325 }, { "epoch": 33.42215568862275, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1346, "step": 22326 }, { "epoch": 33.42365269461078, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1385, "step": 22327 }, { "epoch": 33.4251497005988, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1404, "step": 22328 }, { "epoch": 33.42664670658683, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1358, "step": 22329 }, { "epoch": 33.42814371257485, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1351, "step": 22330 }, { "epoch": 33.42964071856287, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.131, "step": 22331 }, { "epoch": 33.4311377245509, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1349, "step": 22332 }, { "epoch": 33.43263473053892, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.14, "step": 22333 }, { "epoch": 33.43413173652694, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1334, "step": 22334 }, { "epoch": 33.43562874251497, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1387, "step": 22335 }, { "epoch": 33.43712574850299, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1332, "step": 22336 }, { "epoch": 33.43862275449102, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1396, "step": 22337 }, { "epoch": 33.44011976047904, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.14, "step": 22338 }, { "epoch": 33.441616766467064, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.14, "step": 22339 }, { "epoch": 33.44311377245509, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1376, "step": 22340 }, { "epoch": 33.44461077844311, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.139, "step": 22341 }, { "epoch": 33.446107784431135, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1394, "step": 22342 }, { "epoch": 33.44760479041916, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1342, "step": 22343 }, { "epoch": 33.449101796407184, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.133, "step": 22344 }, { "epoch": 33.45059880239521, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1379, "step": 22345 }, { "epoch": 33.452095808383234, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1329, "step": 22346 }, { "epoch": 33.453592814371255, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1357, "step": 22347 }, { "epoch": 33.455089820359284, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 22348 }, { "epoch": 33.456586826347305, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1446, "step": 22349 }, { "epoch": 33.458083832335326, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1422, "step": 22350 }, { "epoch": 33.459580838323355, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1359, "step": 22351 }, { "epoch": 33.461077844311376, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1341, "step": 22352 }, { "epoch": 33.462574850299404, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.139, "step": 22353 }, { "epoch": 33.464071856287426, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1369, "step": 22354 }, { "epoch": 33.46556886227545, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1329, "step": 22355 }, { "epoch": 33.467065868263475, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 22356 }, { "epoch": 33.4685628742515, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1379, "step": 22357 }, { "epoch": 33.47005988023952, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.131, "step": 22358 }, { "epoch": 33.471556886227546, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1355, "step": 22359 }, { "epoch": 33.47305389221557, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1299, "step": 22360 }, { "epoch": 33.474550898203596, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.132, "step": 22361 }, { "epoch": 33.47604790419162, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1304, "step": 22362 }, { "epoch": 33.47754491017964, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1339, "step": 22363 }, { "epoch": 33.47904191616767, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1389, "step": 22364 }, { "epoch": 33.48053892215569, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1355, "step": 22365 }, { "epoch": 33.48203592814371, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1388, "step": 22366 }, { "epoch": 33.48353293413174, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1378, "step": 22367 }, { "epoch": 33.48502994011976, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1338, "step": 22368 }, { "epoch": 33.48652694610779, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.132, "step": 22369 }, { "epoch": 33.48802395209581, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1336, "step": 22370 }, { "epoch": 33.48952095808383, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1393, "step": 22371 }, { "epoch": 33.49101796407186, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1348, "step": 22372 }, { "epoch": 33.49251497005988, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1305, "step": 22373 }, { "epoch": 33.4940119760479, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1291, "step": 22374 }, { "epoch": 33.49550898203593, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1344, "step": 22375 }, { "epoch": 33.49700598802395, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1282, "step": 22376 }, { "epoch": 33.49850299401198, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1359, "step": 22377 }, { "epoch": 33.5, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1355, "step": 22378 }, { "epoch": 33.50149700598802, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1304, "step": 22379 }, { "epoch": 33.50299401197605, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1352, "step": 22380 }, { "epoch": 33.50449101796407, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1339, "step": 22381 }, { "epoch": 33.5059880239521, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1341, "step": 22382 }, { "epoch": 33.50748502994012, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1328, "step": 22383 }, { "epoch": 33.50898203592814, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1368, "step": 22384 }, { "epoch": 33.51047904191617, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1336, "step": 22385 }, { "epoch": 33.51197604790419, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 22386 }, { "epoch": 33.51347305389221, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1318, "step": 22387 }, { "epoch": 33.51497005988024, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1375, "step": 22388 }, { "epoch": 33.51646706586826, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1384, "step": 22389 }, { "epoch": 33.51796407185629, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1321, "step": 22390 }, { "epoch": 33.51946107784431, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.133, "step": 22391 }, { "epoch": 33.52095808383233, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1287, "step": 22392 }, { "epoch": 33.52245508982036, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1388, "step": 22393 }, { "epoch": 33.52395209580838, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1293, "step": 22394 }, { "epoch": 33.525449101796404, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1429, "step": 22395 }, { "epoch": 33.52694610778443, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1388, "step": 22396 }, { "epoch": 33.528443113772454, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1331, "step": 22397 }, { "epoch": 33.52994011976048, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1378, "step": 22398 }, { "epoch": 33.5314371257485, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.137, "step": 22399 }, { "epoch": 33.532934131736525, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1277, "step": 22400 }, { "epoch": 33.53443113772455, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.135, "step": 22401 }, { "epoch": 33.535928143712574, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.135, "step": 22402 }, { "epoch": 33.537425149700596, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1331, "step": 22403 }, { "epoch": 33.538922155688624, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1347, "step": 22404 }, { "epoch": 33.540419161676645, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.134, "step": 22405 }, { "epoch": 33.541916167664674, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1396, "step": 22406 }, { "epoch": 33.543413173652695, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1407, "step": 22407 }, { "epoch": 33.544910179640716, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1321, "step": 22408 }, { "epoch": 33.546407185628745, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 22409 }, { "epoch": 33.547904191616766, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1352, "step": 22410 }, { "epoch": 33.54940119760479, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1308, "step": 22411 }, { "epoch": 33.550898203592816, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1364, "step": 22412 }, { "epoch": 33.55239520958084, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1336, "step": 22413 }, { "epoch": 33.553892215568865, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1407, "step": 22414 }, { "epoch": 33.55538922155689, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1305, "step": 22415 }, { "epoch": 33.55688622754491, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1298, "step": 22416 }, { "epoch": 33.558383233532936, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1231, "step": 22417 }, { "epoch": 33.55988023952096, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1337, "step": 22418 }, { "epoch": 33.56137724550898, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1384, "step": 22419 }, { "epoch": 33.56287425149701, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1346, "step": 22420 }, { "epoch": 33.56437125748503, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1351, "step": 22421 }, { "epoch": 33.56586826347306, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1329, "step": 22422 }, { "epoch": 33.56736526946108, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1332, "step": 22423 }, { "epoch": 33.5688622754491, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1337, "step": 22424 }, { "epoch": 33.57035928143713, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1317, "step": 22425 }, { "epoch": 33.57185628742515, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1367, "step": 22426 }, { "epoch": 33.57335329341317, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1397, "step": 22427 }, { "epoch": 33.5748502994012, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1359, "step": 22428 }, { "epoch": 33.57634730538922, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1371, "step": 22429 }, { "epoch": 33.57784431137725, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1388, "step": 22430 }, { "epoch": 33.57934131736527, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1404, "step": 22431 }, { "epoch": 33.58083832335329, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1326, "step": 22432 }, { "epoch": 33.58233532934132, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1359, "step": 22433 }, { "epoch": 33.58383233532934, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1321, "step": 22434 }, { "epoch": 33.58532934131736, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1377, "step": 22435 }, { "epoch": 33.58682634730539, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1363, "step": 22436 }, { "epoch": 33.58832335329341, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 22437 }, { "epoch": 33.58982035928144, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1339, "step": 22438 }, { "epoch": 33.59131736526946, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1231, "step": 22439 }, { "epoch": 33.59281437125748, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1366, "step": 22440 }, { "epoch": 33.59431137724551, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 22441 }, { "epoch": 33.59580838323353, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1275, "step": 22442 }, { "epoch": 33.59730538922156, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1365, "step": 22443 }, { "epoch": 33.59880239520958, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1332, "step": 22444 }, { "epoch": 33.6002994011976, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1305, "step": 22445 }, { "epoch": 33.60179640718563, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1329, "step": 22446 }, { "epoch": 33.60329341317365, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1333, "step": 22447 }, { "epoch": 33.604790419161674, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1384, "step": 22448 }, { "epoch": 33.6062874251497, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1414, "step": 22449 }, { "epoch": 33.60778443113772, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1285, "step": 22450 }, { "epoch": 33.60928143712575, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1316, "step": 22451 }, { "epoch": 33.61077844311377, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1351, "step": 22452 }, { "epoch": 33.612275449101794, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.142, "step": 22453 }, { "epoch": 33.61377245508982, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1387, "step": 22454 }, { "epoch": 33.615269461077844, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1292, "step": 22455 }, { "epoch": 33.616766467065865, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1294, "step": 22456 }, { "epoch": 33.618263473053894, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.138, "step": 22457 }, { "epoch": 33.619760479041915, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1301, "step": 22458 }, { "epoch": 33.62125748502994, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1364, "step": 22459 }, { "epoch": 33.622754491017965, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1261, "step": 22460 }, { "epoch": 33.624251497005986, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1412, "step": 22461 }, { "epoch": 33.625748502994014, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1365, "step": 22462 }, { "epoch": 33.627245508982035, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.133, "step": 22463 }, { "epoch": 33.62874251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1323, "step": 22464 }, { "epoch": 33.630239520958085, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1353, "step": 22465 }, { "epoch": 33.631736526946106, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1384, "step": 22466 }, { "epoch": 33.633233532934135, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1284, "step": 22467 }, { "epoch": 33.634730538922156, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.135, "step": 22468 }, { "epoch": 33.63622754491018, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1251, "step": 22469 }, { "epoch": 33.637724550898206, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1369, "step": 22470 }, { "epoch": 33.63922155688623, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1379, "step": 22471 }, { "epoch": 33.64071856287425, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1369, "step": 22472 }, { "epoch": 33.64221556886228, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 22473 }, { "epoch": 33.6437125748503, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1281, "step": 22474 }, { "epoch": 33.645209580838326, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1368, "step": 22475 }, { "epoch": 33.64670658682635, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.135, "step": 22476 }, { "epoch": 33.64820359281437, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1397, "step": 22477 }, { "epoch": 33.6497005988024, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1342, "step": 22478 }, { "epoch": 33.65119760479042, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1314, "step": 22479 }, { "epoch": 33.65269461077844, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1368, "step": 22480 }, { "epoch": 33.65419161676647, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1344, "step": 22481 }, { "epoch": 33.65568862275449, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1333, "step": 22482 }, { "epoch": 33.65718562874252, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.133, "step": 22483 }, { "epoch": 33.65868263473054, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1319, "step": 22484 }, { "epoch": 33.66017964071856, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1383, "step": 22485 }, { "epoch": 33.66167664670659, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1356, "step": 22486 }, { "epoch": 33.66317365269461, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1324, "step": 22487 }, { "epoch": 33.66467065868264, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1336, "step": 22488 }, { "epoch": 33.66616766467066, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1357, "step": 22489 }, { "epoch": 33.66766467065868, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1373, "step": 22490 }, { "epoch": 33.66916167664671, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.135, "step": 22491 }, { "epoch": 33.67065868263473, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1333, "step": 22492 }, { "epoch": 33.67215568862275, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1319, "step": 22493 }, { "epoch": 33.67365269461078, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1304, "step": 22494 }, { "epoch": 33.6751497005988, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1294, "step": 22495 }, { "epoch": 33.67664670658683, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 22496 }, { "epoch": 33.67814371257485, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1372, "step": 22497 }, { "epoch": 33.67964071856287, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1287, "step": 22498 }, { "epoch": 33.6811377245509, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 22499 }, { "epoch": 33.68263473053892, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1365, "step": 22500 }, { "epoch": 33.68413173652694, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1337, "step": 22501 }, { "epoch": 33.68562874251497, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1333, "step": 22502 }, { "epoch": 33.68712574850299, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 22503 }, { "epoch": 33.68862275449102, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 22504 }, { "epoch": 33.69011976047904, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1311, "step": 22505 }, { "epoch": 33.691616766467064, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1387, "step": 22506 }, { "epoch": 33.69311377245509, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1296, "step": 22507 }, { "epoch": 33.69461077844311, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.134, "step": 22508 }, { "epoch": 33.696107784431135, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.138, "step": 22509 }, { "epoch": 33.69760479041916, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1351, "step": 22510 }, { "epoch": 33.699101796407184, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1226, "step": 22511 }, { "epoch": 33.70059880239521, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.143, "step": 22512 }, { "epoch": 33.702095808383234, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1352, "step": 22513 }, { "epoch": 33.703592814371255, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1373, "step": 22514 }, { "epoch": 33.705089820359284, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 22515 }, { "epoch": 33.706586826347305, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1316, "step": 22516 }, { "epoch": 33.708083832335326, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1337, "step": 22517 }, { "epoch": 33.709580838323355, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1311, "step": 22518 }, { "epoch": 33.711077844311376, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1337, "step": 22519 }, { "epoch": 33.712574850299404, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1263, "step": 22520 }, { "epoch": 33.714071856287426, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 22521 }, { "epoch": 33.71556886227545, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1349, "step": 22522 }, { "epoch": 33.717065868263475, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1233, "step": 22523 }, { "epoch": 33.7185628742515, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1392, "step": 22524 }, { "epoch": 33.72005988023952, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1313, "step": 22525 }, { "epoch": 33.721556886227546, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.131, "step": 22526 }, { "epoch": 33.72305389221557, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1356, "step": 22527 }, { "epoch": 33.724550898203596, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1297, "step": 22528 }, { "epoch": 33.72604790419162, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1306, "step": 22529 }, { "epoch": 33.72754491017964, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1263, "step": 22530 }, { "epoch": 33.72904191616767, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.13, "step": 22531 }, { "epoch": 33.73053892215569, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1266, "step": 22532 }, { "epoch": 33.73203592814371, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1319, "step": 22533 }, { "epoch": 33.73353293413174, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1329, "step": 22534 }, { "epoch": 33.73502994011976, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1317, "step": 22535 }, { "epoch": 33.73652694610779, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1259, "step": 22536 }, { "epoch": 33.73802395209581, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.134, "step": 22537 }, { "epoch": 33.73952095808383, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1328, "step": 22538 }, { "epoch": 33.74101796407186, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1349, "step": 22539 }, { "epoch": 33.74251497005988, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1282, "step": 22540 }, { "epoch": 33.7440119760479, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1337, "step": 22541 }, { "epoch": 33.74550898203593, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1258, "step": 22542 }, { "epoch": 33.74700598802395, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1305, "step": 22543 }, { "epoch": 33.74850299401198, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1386, "step": 22544 }, { "epoch": 33.75, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1284, "step": 22545 }, { "epoch": 33.75149700598802, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1337, "step": 22546 }, { "epoch": 33.75299401197605, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1339, "step": 22547 }, { "epoch": 33.75449101796407, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1277, "step": 22548 }, { "epoch": 33.7559880239521, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1379, "step": 22549 }, { "epoch": 33.75748502994012, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1317, "step": 22550 }, { "epoch": 33.75898203592814, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1298, "step": 22551 }, { "epoch": 33.76047904191617, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1256, "step": 22552 }, { "epoch": 33.76197604790419, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1357, "step": 22553 }, { "epoch": 33.76347305389221, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 22554 }, { "epoch": 33.76497005988024, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1331, "step": 22555 }, { "epoch": 33.76646706586826, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1302, "step": 22556 }, { "epoch": 33.76796407185629, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1368, "step": 22557 }, { "epoch": 33.76946107784431, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1312, "step": 22558 }, { "epoch": 33.77095808383233, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1328, "step": 22559 }, { "epoch": 33.77245508982036, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1298, "step": 22560 }, { "epoch": 33.77395209580838, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.137, "step": 22561 }, { "epoch": 33.775449101796404, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1354, "step": 22562 }, { "epoch": 33.77694610778443, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1351, "step": 22563 }, { "epoch": 33.778443113772454, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.131, "step": 22564 }, { "epoch": 33.77994011976048, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1328, "step": 22565 }, { "epoch": 33.7814371257485, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1247, "step": 22566 }, { "epoch": 33.782934131736525, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1239, "step": 22567 }, { "epoch": 33.78443113772455, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1354, "step": 22568 }, { "epoch": 33.785928143712574, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1285, "step": 22569 }, { "epoch": 33.787425149700596, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1286, "step": 22570 }, { "epoch": 33.788922155688624, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.132, "step": 22571 }, { "epoch": 33.790419161676645, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1272, "step": 22572 }, { "epoch": 33.791916167664674, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1293, "step": 22573 }, { "epoch": 33.793413173652695, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1354, "step": 22574 }, { "epoch": 33.794910179640716, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1305, "step": 22575 }, { "epoch": 33.796407185628745, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1279, "step": 22576 }, { "epoch": 33.797904191616766, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1302, "step": 22577 }, { "epoch": 33.79940119760479, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1313, "step": 22578 }, { "epoch": 33.800898203592816, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1239, "step": 22579 }, { "epoch": 33.80239520958084, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1387, "step": 22580 }, { "epoch": 33.803892215568865, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1359, "step": 22581 }, { "epoch": 33.80538922155689, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1315, "step": 22582 }, { "epoch": 33.80688622754491, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1257, "step": 22583 }, { "epoch": 33.808383233532936, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1357, "step": 22584 }, { "epoch": 33.80988023952096, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1296, "step": 22585 }, { "epoch": 33.81137724550898, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1325, "step": 22586 }, { "epoch": 33.81287425149701, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1252, "step": 22587 }, { "epoch": 33.81437125748503, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1302, "step": 22588 }, { "epoch": 33.81586826347306, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1295, "step": 22589 }, { "epoch": 33.81736526946108, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1318, "step": 22590 }, { "epoch": 33.8188622754491, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 22591 }, { "epoch": 33.82035928143713, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1275, "step": 22592 }, { "epoch": 33.82185628742515, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 22593 }, { "epoch": 33.82335329341317, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 22594 }, { "epoch": 33.8248502994012, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1294, "step": 22595 }, { "epoch": 33.82634730538922, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1246, "step": 22596 }, { "epoch": 33.82784431137725, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1307, "step": 22597 }, { "epoch": 33.82934131736527, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1281, "step": 22598 }, { "epoch": 33.83083832335329, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1289, "step": 22599 }, { "epoch": 33.83233532934132, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1311, "step": 22600 }, { "epoch": 33.83383233532934, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1336, "step": 22601 }, { "epoch": 33.83532934131736, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1306, "step": 22602 }, { "epoch": 33.83682634730539, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1328, "step": 22603 }, { "epoch": 33.83832335329341, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.132, "step": 22604 }, { "epoch": 33.83982035928144, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1346, "step": 22605 }, { "epoch": 33.84131736526946, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1326, "step": 22606 }, { "epoch": 33.84281437125748, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1306, "step": 22607 }, { "epoch": 33.84431137724551, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1232, "step": 22608 }, { "epoch": 33.84580838323353, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1279, "step": 22609 }, { "epoch": 33.84730538922156, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1309, "step": 22610 }, { "epoch": 33.84880239520958, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1323, "step": 22611 }, { "epoch": 33.8502994011976, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1357, "step": 22612 }, { "epoch": 33.85179640718563, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1335, "step": 22613 }, { "epoch": 33.85329341317365, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1325, "step": 22614 }, { "epoch": 33.854790419161674, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1308, "step": 22615 }, { "epoch": 33.8562874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 22616 }, { "epoch": 33.85778443113772, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1326, "step": 22617 }, { "epoch": 33.85928143712575, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1314, "step": 22618 }, { "epoch": 33.86077844311377, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1239, "step": 22619 }, { "epoch": 33.862275449101794, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 22620 }, { "epoch": 33.86377245508982, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 22621 }, { "epoch": 33.865269461077844, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1312, "step": 22622 }, { "epoch": 33.866766467065865, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1315, "step": 22623 }, { "epoch": 33.868263473053894, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1266, "step": 22624 }, { "epoch": 33.869760479041915, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 22625 }, { "epoch": 33.87125748502994, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1257, "step": 22626 }, { "epoch": 33.872754491017965, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1307, "step": 22627 }, { "epoch": 33.874251497005986, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1387, "step": 22628 }, { "epoch": 33.875748502994014, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.133, "step": 22629 }, { "epoch": 33.877245508982035, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1323, "step": 22630 }, { "epoch": 33.87874251497006, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1325, "step": 22631 }, { "epoch": 33.880239520958085, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1296, "step": 22632 }, { "epoch": 33.881736526946106, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1251, "step": 22633 }, { "epoch": 33.883233532934135, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1288, "step": 22634 }, { "epoch": 33.884730538922156, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1296, "step": 22635 }, { "epoch": 33.88622754491018, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1398, "step": 22636 }, { "epoch": 33.887724550898206, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1259, "step": 22637 }, { "epoch": 33.88922155688623, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1299, "step": 22638 }, { "epoch": 33.89071856287425, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1278, "step": 22639 }, { "epoch": 33.89221556886228, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1268, "step": 22640 }, { "epoch": 33.8937125748503, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1303, "step": 22641 }, { "epoch": 33.895209580838326, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1371, "step": 22642 }, { "epoch": 33.89670658682635, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 22643 }, { "epoch": 33.89820359281437, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1392, "step": 22644 }, { "epoch": 33.8997005988024, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1323, "step": 22645 }, { "epoch": 33.90119760479042, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1331, "step": 22646 }, { "epoch": 33.90269461077844, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1291, "step": 22647 }, { "epoch": 33.90419161676647, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.133, "step": 22648 }, { "epoch": 33.90568862275449, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1286, "step": 22649 }, { "epoch": 33.90718562874252, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1313, "step": 22650 }, { "epoch": 33.90868263473054, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1244, "step": 22651 }, { "epoch": 33.91017964071856, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 22652 }, { "epoch": 33.91167664670659, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1289, "step": 22653 }, { "epoch": 33.91317365269461, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1361, "step": 22654 }, { "epoch": 33.91467065868264, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1283, "step": 22655 }, { "epoch": 33.91616766467066, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1257, "step": 22656 }, { "epoch": 33.91766467065868, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1325, "step": 22657 }, { "epoch": 33.91916167664671, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1271, "step": 22658 }, { "epoch": 33.92065868263473, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1257, "step": 22659 }, { "epoch": 33.92215568862275, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1332, "step": 22660 }, { "epoch": 33.92365269461078, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1325, "step": 22661 }, { "epoch": 33.9251497005988, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1322, "step": 22662 }, { "epoch": 33.92664670658683, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1336, "step": 22663 }, { "epoch": 33.92814371257485, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1292, "step": 22664 }, { "epoch": 33.92964071856287, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.135, "step": 22665 }, { "epoch": 33.9311377245509, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1284, "step": 22666 }, { "epoch": 33.93263473053892, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1291, "step": 22667 }, { "epoch": 33.93413173652694, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1326, "step": 22668 }, { "epoch": 33.93562874251497, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1307, "step": 22669 }, { "epoch": 33.93712574850299, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1338, "step": 22670 }, { "epoch": 33.93862275449102, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1293, "step": 22671 }, { "epoch": 33.94011976047904, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1244, "step": 22672 }, { "epoch": 33.941616766467064, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1308, "step": 22673 }, { "epoch": 33.94311377245509, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1305, "step": 22674 }, { "epoch": 33.94461077844311, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.132, "step": 22675 }, { "epoch": 33.946107784431135, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 22676 }, { "epoch": 33.94760479041916, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1337, "step": 22677 }, { "epoch": 33.949101796407184, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1315, "step": 22678 }, { "epoch": 33.95059880239521, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1273, "step": 22679 }, { "epoch": 33.952095808383234, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 22680 }, { "epoch": 33.953592814371255, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1321, "step": 22681 }, { "epoch": 33.955089820359284, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1354, "step": 22682 }, { "epoch": 33.956586826347305, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1323, "step": 22683 }, { "epoch": 33.958083832335326, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1309, "step": 22684 }, { "epoch": 33.959580838323355, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1326, "step": 22685 }, { "epoch": 33.961077844311376, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.129, "step": 22686 }, { "epoch": 33.962574850299404, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1242, "step": 22687 }, { "epoch": 33.964071856287426, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1335, "step": 22688 }, { "epoch": 33.96556886227545, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1311, "step": 22689 }, { "epoch": 33.967065868263475, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1265, "step": 22690 }, { "epoch": 33.9685628742515, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.126, "step": 22691 }, { "epoch": 33.97005988023952, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1268, "step": 22692 }, { "epoch": 33.971556886227546, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1256, "step": 22693 }, { "epoch": 33.97305389221557, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1374, "step": 22694 }, { "epoch": 33.974550898203596, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 22695 }, { "epoch": 33.97604790419162, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 22696 }, { "epoch": 33.97754491017964, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1435, "step": 22697 }, { "epoch": 33.97904191616767, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1293, "step": 22698 }, { "epoch": 33.98053892215569, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1318, "step": 22699 }, { "epoch": 33.98203592814371, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1241, "step": 22700 }, { "epoch": 33.98353293413174, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1346, "step": 22701 }, { "epoch": 33.98502994011976, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1262, "step": 22702 }, { "epoch": 33.98652694610779, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1324, "step": 22703 }, { "epoch": 33.98802395209581, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1311, "step": 22704 }, { "epoch": 33.98952095808383, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1325, "step": 22705 }, { "epoch": 33.99101796407186, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1262, "step": 22706 }, { "epoch": 33.99251497005988, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1282, "step": 22707 }, { "epoch": 33.9940119760479, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1278, "step": 22708 }, { "epoch": 33.99550898203593, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.132, "step": 22709 }, { "epoch": 33.99700598802395, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1314, "step": 22710 }, { "epoch": 33.99850299401198, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.136, "step": 22711 }, { "epoch": 34.0, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1333, "step": 22712 }, { "epoch": 34.00149700598802, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1297, "step": 22713 }, { "epoch": 34.00299401197605, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1255, "step": 22714 }, { "epoch": 34.00449101796407, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1325, "step": 22715 }, { "epoch": 34.0059880239521, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1331, "step": 22716 }, { "epoch": 34.00748502994012, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1307, "step": 22717 }, { "epoch": 34.00898203592814, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1281, "step": 22718 }, { "epoch": 34.01047904191617, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1293, "step": 22719 }, { "epoch": 34.01197604790419, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1306, "step": 22720 }, { "epoch": 34.01347305389221, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1274, "step": 22721 }, { "epoch": 34.01497005988024, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1252, "step": 22722 }, { "epoch": 34.01646706586826, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 22723 }, { "epoch": 34.01796407185629, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1331, "step": 22724 }, { "epoch": 34.01946107784431, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1293, "step": 22725 }, { "epoch": 34.02095808383233, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1371, "step": 22726 }, { "epoch": 34.02245508982036, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1254, "step": 22727 }, { "epoch": 34.02395209580838, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.127, "step": 22728 }, { "epoch": 34.025449101796404, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1315, "step": 22729 }, { "epoch": 34.02694610778443, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1361, "step": 22730 }, { "epoch": 34.028443113772454, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1293, "step": 22731 }, { "epoch": 34.02994011976048, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1269, "step": 22732 }, { "epoch": 34.0314371257485, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1266, "step": 22733 }, { "epoch": 34.032934131736525, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1289, "step": 22734 }, { "epoch": 34.03443113772455, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 22735 }, { "epoch": 34.035928143712574, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1244, "step": 22736 }, { "epoch": 34.037425149700596, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1305, "step": 22737 }, { "epoch": 34.038922155688624, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1263, "step": 22738 }, { "epoch": 34.040419161676645, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1241, "step": 22739 }, { "epoch": 34.041916167664674, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1319, "step": 22740 }, { "epoch": 34.043413173652695, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.131, "step": 22741 }, { "epoch": 34.044910179640716, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1304, "step": 22742 }, { "epoch": 34.046407185628745, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1218, "step": 22743 }, { "epoch": 34.047904191616766, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1293, "step": 22744 }, { "epoch": 34.04940119760479, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1241, "step": 22745 }, { "epoch": 34.050898203592816, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1316, "step": 22746 }, { "epoch": 34.05239520958084, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1286, "step": 22747 }, { "epoch": 34.053892215568865, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.128, "step": 22748 }, { "epoch": 34.05538922155689, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1253, "step": 22749 }, { "epoch": 34.05688622754491, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 22750 }, { "epoch": 34.058383233532936, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.13, "step": 22751 }, { "epoch": 34.05988023952096, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.131, "step": 22752 }, { "epoch": 34.06137724550898, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1375, "step": 22753 }, { "epoch": 34.06287425149701, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1281, "step": 22754 }, { "epoch": 34.06437125748503, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.134, "step": 22755 }, { "epoch": 34.06586826347306, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1323, "step": 22756 }, { "epoch": 34.06736526946108, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 22757 }, { "epoch": 34.0688622754491, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1305, "step": 22758 }, { "epoch": 34.07035928143713, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1355, "step": 22759 }, { "epoch": 34.07185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1302, "step": 22760 }, { "epoch": 34.07335329341317, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1247, "step": 22761 }, { "epoch": 34.0748502994012, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1216, "step": 22762 }, { "epoch": 34.07634730538922, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1308, "step": 22763 }, { "epoch": 34.07784431137725, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1304, "step": 22764 }, { "epoch": 34.07934131736527, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1274, "step": 22765 }, { "epoch": 34.08083832335329, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1362, "step": 22766 }, { "epoch": 34.08233532934132, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1352, "step": 22767 }, { "epoch": 34.08383233532934, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1316, "step": 22768 }, { "epoch": 34.08532934131737, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1331, "step": 22769 }, { "epoch": 34.08682634730539, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1313, "step": 22770 }, { "epoch": 34.08832335329341, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1297, "step": 22771 }, { "epoch": 34.08982035928144, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1313, "step": 22772 }, { "epoch": 34.09131736526946, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1247, "step": 22773 }, { "epoch": 34.09281437125748, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1255, "step": 22774 }, { "epoch": 34.09431137724551, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.133, "step": 22775 }, { "epoch": 34.09580838323353, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1297, "step": 22776 }, { "epoch": 34.09730538922156, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1372, "step": 22777 }, { "epoch": 34.09880239520958, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 22778 }, { "epoch": 34.1002994011976, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 22779 }, { "epoch": 34.10179640718563, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.13, "step": 22780 }, { "epoch": 34.10329341317365, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.126, "step": 22781 }, { "epoch": 34.104790419161674, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.134, "step": 22782 }, { "epoch": 34.1062874251497, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 22783 }, { "epoch": 34.10778443113772, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1268, "step": 22784 }, { "epoch": 34.10928143712575, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1277, "step": 22785 }, { "epoch": 34.11077844311377, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1313, "step": 22786 }, { "epoch": 34.112275449101794, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1305, "step": 22787 }, { "epoch": 34.11377245508982, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1323, "step": 22788 }, { "epoch": 34.115269461077844, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.135, "step": 22789 }, { "epoch": 34.116766467065865, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 22790 }, { "epoch": 34.118263473053894, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1278, "step": 22791 }, { "epoch": 34.119760479041915, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1313, "step": 22792 }, { "epoch": 34.12125748502994, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1226, "step": 22793 }, { "epoch": 34.122754491017965, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1338, "step": 22794 }, { "epoch": 34.124251497005986, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 22795 }, { "epoch": 34.125748502994014, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1313, "step": 22796 }, { "epoch": 34.127245508982035, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.123, "step": 22797 }, { "epoch": 34.12874251497006, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1259, "step": 22798 }, { "epoch": 34.130239520958085, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1335, "step": 22799 }, { "epoch": 34.131736526946106, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1257, "step": 22800 }, { "epoch": 34.133233532934135, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.133, "step": 22801 }, { "epoch": 34.134730538922156, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1319, "step": 22802 }, { "epoch": 34.13622754491018, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1258, "step": 22803 }, { "epoch": 34.137724550898206, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1287, "step": 22804 }, { "epoch": 34.13922155688623, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.137, "step": 22805 }, { "epoch": 34.14071856287425, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.131, "step": 22806 }, { "epoch": 34.14221556886228, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1233, "step": 22807 }, { "epoch": 34.1437125748503, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1242, "step": 22808 }, { "epoch": 34.145209580838326, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1242, "step": 22809 }, { "epoch": 34.14670658682635, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1243, "step": 22810 }, { "epoch": 34.14820359281437, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.128, "step": 22811 }, { "epoch": 34.1497005988024, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1384, "step": 22812 }, { "epoch": 34.15119760479042, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1265, "step": 22813 }, { "epoch": 34.15269461077844, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1302, "step": 22814 }, { "epoch": 34.15419161676647, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1292, "step": 22815 }, { "epoch": 34.15568862275449, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1323, "step": 22816 }, { "epoch": 34.15718562874252, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1277, "step": 22817 }, { "epoch": 34.15868263473054, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1315, "step": 22818 }, { "epoch": 34.16017964071856, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1328, "step": 22819 }, { "epoch": 34.16167664670659, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1282, "step": 22820 }, { "epoch": 34.16317365269461, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.13, "step": 22821 }, { "epoch": 34.16467065868263, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.125, "step": 22822 }, { "epoch": 34.16616766467066, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 22823 }, { "epoch": 34.16766467065868, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1353, "step": 22824 }, { "epoch": 34.16916167664671, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1291, "step": 22825 }, { "epoch": 34.17065868263473, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 22826 }, { "epoch": 34.17215568862275, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1279, "step": 22827 }, { "epoch": 34.17365269461078, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1277, "step": 22828 }, { "epoch": 34.1751497005988, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1222, "step": 22829 }, { "epoch": 34.17664670658683, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1257, "step": 22830 }, { "epoch": 34.17814371257485, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1279, "step": 22831 }, { "epoch": 34.17964071856287, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1339, "step": 22832 }, { "epoch": 34.1811377245509, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1281, "step": 22833 }, { "epoch": 34.18263473053892, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.127, "step": 22834 }, { "epoch": 34.18413173652694, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1319, "step": 22835 }, { "epoch": 34.18562874251497, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.131, "step": 22836 }, { "epoch": 34.18712574850299, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1268, "step": 22837 }, { "epoch": 34.18862275449102, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1336, "step": 22838 }, { "epoch": 34.19011976047904, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1279, "step": 22839 }, { "epoch": 34.191616766467064, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1317, "step": 22840 }, { "epoch": 34.19311377245509, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1273, "step": 22841 }, { "epoch": 34.19461077844311, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1278, "step": 22842 }, { "epoch": 34.196107784431135, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1269, "step": 22843 }, { "epoch": 34.19760479041916, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.127, "step": 22844 }, { "epoch": 34.199101796407184, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1334, "step": 22845 }, { "epoch": 34.20059880239521, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1315, "step": 22846 }, { "epoch": 34.202095808383234, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1333, "step": 22847 }, { "epoch": 34.203592814371255, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1283, "step": 22848 }, { "epoch": 34.205089820359284, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1335, "step": 22849 }, { "epoch": 34.206586826347305, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1297, "step": 22850 }, { "epoch": 34.208083832335326, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1272, "step": 22851 }, { "epoch": 34.209580838323355, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1375, "step": 22852 }, { "epoch": 34.211077844311376, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.132, "step": 22853 }, { "epoch": 34.212574850299404, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1346, "step": 22854 }, { "epoch": 34.214071856287426, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1262, "step": 22855 }, { "epoch": 34.21556886227545, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1287, "step": 22856 }, { "epoch": 34.217065868263475, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.133, "step": 22857 }, { "epoch": 34.2185628742515, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1257, "step": 22858 }, { "epoch": 34.22005988023952, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1338, "step": 22859 }, { "epoch": 34.221556886227546, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 22860 }, { "epoch": 34.22305389221557, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1291, "step": 22861 }, { "epoch": 34.224550898203596, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1274, "step": 22862 }, { "epoch": 34.22604790419162, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1339, "step": 22863 }, { "epoch": 34.22754491017964, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.134, "step": 22864 }, { "epoch": 34.22904191616767, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1294, "step": 22865 }, { "epoch": 34.23053892215569, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1298, "step": 22866 }, { "epoch": 34.23203592814371, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1327, "step": 22867 }, { "epoch": 34.23353293413174, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1262, "step": 22868 }, { "epoch": 34.23502994011976, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1297, "step": 22869 }, { "epoch": 34.23652694610779, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1376, "step": 22870 }, { "epoch": 34.23802395209581, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1278, "step": 22871 }, { "epoch": 34.23952095808383, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1276, "step": 22872 }, { "epoch": 34.24101796407186, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1238, "step": 22873 }, { "epoch": 34.24251497005988, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1315, "step": 22874 }, { "epoch": 34.2440119760479, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1336, "step": 22875 }, { "epoch": 34.24550898203593, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1313, "step": 22876 }, { "epoch": 34.24700598802395, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1345, "step": 22877 }, { "epoch": 34.24850299401198, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1353, "step": 22878 }, { "epoch": 34.25, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1259, "step": 22879 }, { "epoch": 34.25149700598802, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 22880 }, { "epoch": 34.25299401197605, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1286, "step": 22881 }, { "epoch": 34.25449101796407, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1321, "step": 22882 }, { "epoch": 34.2559880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1319, "step": 22883 }, { "epoch": 34.25748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1281, "step": 22884 }, { "epoch": 34.25898203592814, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1358, "step": 22885 }, { "epoch": 34.26047904191617, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1315, "step": 22886 }, { "epoch": 34.26197604790419, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1275, "step": 22887 }, { "epoch": 34.26347305389221, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1252, "step": 22888 }, { "epoch": 34.26497005988024, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1327, "step": 22889 }, { "epoch": 34.26646706586826, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1357, "step": 22890 }, { "epoch": 34.26796407185629, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1268, "step": 22891 }, { "epoch": 34.26946107784431, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1291, "step": 22892 }, { "epoch": 34.27095808383233, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.122, "step": 22893 }, { "epoch": 34.27245508982036, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1352, "step": 22894 }, { "epoch": 34.27395209580838, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1293, "step": 22895 }, { "epoch": 34.275449101796404, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1284, "step": 22896 }, { "epoch": 34.27694610778443, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1303, "step": 22897 }, { "epoch": 34.278443113772454, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1335, "step": 22898 }, { "epoch": 34.27994011976048, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.126, "step": 22899 }, { "epoch": 34.2814371257485, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1404, "step": 22900 }, { "epoch": 34.282934131736525, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1271, "step": 22901 }, { "epoch": 34.28443113772455, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1264, "step": 22902 }, { "epoch": 34.285928143712574, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1251, "step": 22903 }, { "epoch": 34.287425149700596, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1274, "step": 22904 }, { "epoch": 34.288922155688624, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 22905 }, { "epoch": 34.290419161676645, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1265, "step": 22906 }, { "epoch": 34.291916167664674, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1321, "step": 22907 }, { "epoch": 34.293413173652695, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1309, "step": 22908 }, { "epoch": 34.294910179640716, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.137, "step": 22909 }, { "epoch": 34.296407185628745, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1258, "step": 22910 }, { "epoch": 34.297904191616766, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1244, "step": 22911 }, { "epoch": 34.29940119760479, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1335, "step": 22912 }, { "epoch": 34.300898203592816, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1327, "step": 22913 }, { "epoch": 34.30239520958084, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.128, "step": 22914 }, { "epoch": 34.303892215568865, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1265, "step": 22915 }, { "epoch": 34.30538922155689, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1359, "step": 22916 }, { "epoch": 34.30688622754491, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1169, "step": 22917 }, { "epoch": 34.308383233532936, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1246, "step": 22918 }, { "epoch": 34.30988023952096, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1251, "step": 22919 }, { "epoch": 34.31137724550898, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1313, "step": 22920 }, { "epoch": 34.31287425149701, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1318, "step": 22921 }, { "epoch": 34.31437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1317, "step": 22922 }, { "epoch": 34.31586826347306, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1223, "step": 22923 }, { "epoch": 34.31736526946108, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.128, "step": 22924 }, { "epoch": 34.3188622754491, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 22925 }, { "epoch": 34.32035928143713, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1278, "step": 22926 }, { "epoch": 34.32185628742515, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1346, "step": 22927 }, { "epoch": 34.32335329341317, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1264, "step": 22928 }, { "epoch": 34.3248502994012, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1264, "step": 22929 }, { "epoch": 34.32634730538922, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.12, "step": 22930 }, { "epoch": 34.32784431137725, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1322, "step": 22931 }, { "epoch": 34.32934131736527, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1363, "step": 22932 }, { "epoch": 34.33083832335329, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1311, "step": 22933 }, { "epoch": 34.33233532934132, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1261, "step": 22934 }, { "epoch": 34.33383233532934, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1274, "step": 22935 }, { "epoch": 34.33532934131736, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1311, "step": 22936 }, { "epoch": 34.33682634730539, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1318, "step": 22937 }, { "epoch": 34.33832335329341, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1261, "step": 22938 }, { "epoch": 34.33982035928144, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.128, "step": 22939 }, { "epoch": 34.34131736526946, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1297, "step": 22940 }, { "epoch": 34.34281437125748, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1276, "step": 22941 }, { "epoch": 34.34431137724551, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1256, "step": 22942 }, { "epoch": 34.34580838323353, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1246, "step": 22943 }, { "epoch": 34.34730538922156, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1235, "step": 22944 }, { "epoch": 34.34880239520958, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1264, "step": 22945 }, { "epoch": 34.3502994011976, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.13, "step": 22946 }, { "epoch": 34.35179640718563, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1324, "step": 22947 }, { "epoch": 34.35329341317365, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 22948 }, { "epoch": 34.354790419161674, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1227, "step": 22949 }, { "epoch": 34.3562874251497, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1264, "step": 22950 }, { "epoch": 34.35778443113772, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1354, "step": 22951 }, { "epoch": 34.35928143712575, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1284, "step": 22952 }, { "epoch": 34.36077844311377, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1342, "step": 22953 }, { "epoch": 34.362275449101794, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1325, "step": 22954 }, { "epoch": 34.36377245508982, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1259, "step": 22955 }, { "epoch": 34.365269461077844, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1234, "step": 22956 }, { "epoch": 34.366766467065865, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1201, "step": 22957 }, { "epoch": 34.368263473053894, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1273, "step": 22958 }, { "epoch": 34.369760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1254, "step": 22959 }, { "epoch": 34.37125748502994, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1308, "step": 22960 }, { "epoch": 34.372754491017965, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1308, "step": 22961 }, { "epoch": 34.374251497005986, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1267, "step": 22962 }, { "epoch": 34.375748502994014, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.129, "step": 22963 }, { "epoch": 34.377245508982035, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1254, "step": 22964 }, { "epoch": 34.37874251497006, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.124, "step": 22965 }, { "epoch": 34.380239520958085, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.129, "step": 22966 }, { "epoch": 34.381736526946106, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1257, "step": 22967 }, { "epoch": 34.383233532934135, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1273, "step": 22968 }, { "epoch": 34.384730538922156, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1319, "step": 22969 }, { "epoch": 34.38622754491018, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1248, "step": 22970 }, { "epoch": 34.387724550898206, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1297, "step": 22971 }, { "epoch": 34.38922155688623, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1309, "step": 22972 }, { "epoch": 34.39071856287425, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1309, "step": 22973 }, { "epoch": 34.39221556886228, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.12, "step": 22974 }, { "epoch": 34.3937125748503, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1249, "step": 22975 }, { "epoch": 34.395209580838326, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1193, "step": 22976 }, { "epoch": 34.39670658682635, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1253, "step": 22977 }, { "epoch": 34.39820359281437, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1249, "step": 22978 }, { "epoch": 34.3997005988024, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1288, "step": 22979 }, { "epoch": 34.40119760479042, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1239, "step": 22980 }, { "epoch": 34.40269461077844, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1249, "step": 22981 }, { "epoch": 34.40419161676647, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1294, "step": 22982 }, { "epoch": 34.40568862275449, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1321, "step": 22983 }, { "epoch": 34.40718562874252, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1259, "step": 22984 }, { "epoch": 34.40868263473054, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1307, "step": 22985 }, { "epoch": 34.41017964071856, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1251, "step": 22986 }, { "epoch": 34.41167664670659, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1325, "step": 22987 }, { "epoch": 34.41317365269461, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 22988 }, { "epoch": 34.41467065868264, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1264, "step": 22989 }, { "epoch": 34.41616766467066, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 22990 }, { "epoch": 34.41766467065868, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1288, "step": 22991 }, { "epoch": 34.41916167664671, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1304, "step": 22992 }, { "epoch": 34.42065868263473, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1285, "step": 22993 }, { "epoch": 34.42215568862275, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1202, "step": 22994 }, { "epoch": 34.42365269461078, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1265, "step": 22995 }, { "epoch": 34.4251497005988, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 22996 }, { "epoch": 34.42664670658683, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1273, "step": 22997 }, { "epoch": 34.42814371257485, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1284, "step": 22998 }, { "epoch": 34.42964071856287, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.129, "step": 22999 }, { "epoch": 34.4311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.121, "step": 23000 }, { "epoch": 34.43263473053892, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1375, "step": 23001 }, { "epoch": 34.43413173652694, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1298, "step": 23002 }, { "epoch": 34.43562874251497, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1297, "step": 23003 }, { "epoch": 34.43712574850299, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1226, "step": 23004 }, { "epoch": 34.43862275449102, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1151, "step": 23005 }, { "epoch": 34.44011976047904, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1209, "step": 23006 }, { "epoch": 34.441616766467064, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1277, "step": 23007 }, { "epoch": 34.44311377245509, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1276, "step": 23008 }, { "epoch": 34.44461077844311, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1295, "step": 23009 }, { "epoch": 34.446107784431135, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1235, "step": 23010 }, { "epoch": 34.44760479041916, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1303, "step": 23011 }, { "epoch": 34.449101796407184, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1306, "step": 23012 }, { "epoch": 34.45059880239521, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1287, "step": 23013 }, { "epoch": 34.452095808383234, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1335, "step": 23014 }, { "epoch": 34.453592814371255, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1228, "step": 23015 }, { "epoch": 34.455089820359284, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1253, "step": 23016 }, { "epoch": 34.456586826347305, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1285, "step": 23017 }, { "epoch": 34.458083832335326, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1352, "step": 23018 }, { "epoch": 34.459580838323355, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1252, "step": 23019 }, { "epoch": 34.461077844311376, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1261, "step": 23020 }, { "epoch": 34.462574850299404, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1294, "step": 23021 }, { "epoch": 34.464071856287426, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1259, "step": 23022 }, { "epoch": 34.46556886227545, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1256, "step": 23023 }, { "epoch": 34.467065868263475, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1296, "step": 23024 }, { "epoch": 34.4685628742515, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1289, "step": 23025 }, { "epoch": 34.47005988023952, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1249, "step": 23026 }, { "epoch": 34.471556886227546, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1292, "step": 23027 }, { "epoch": 34.47305389221557, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1313, "step": 23028 }, { "epoch": 34.474550898203596, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1275, "step": 23029 }, { "epoch": 34.47604790419162, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1325, "step": 23030 }, { "epoch": 34.47754491017964, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1292, "step": 23031 }, { "epoch": 34.47904191616767, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.128, "step": 23032 }, { "epoch": 34.48053892215569, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.13, "step": 23033 }, { "epoch": 34.48203592814371, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 23034 }, { "epoch": 34.48353293413174, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1287, "step": 23035 }, { "epoch": 34.48502994011976, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1217, "step": 23036 }, { "epoch": 34.48652694610779, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1295, "step": 23037 }, { "epoch": 34.48802395209581, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.128, "step": 23038 }, { "epoch": 34.48952095808383, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1277, "step": 23039 }, { "epoch": 34.49101796407186, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1325, "step": 23040 }, { "epoch": 34.49251497005988, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1305, "step": 23041 }, { "epoch": 34.4940119760479, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1317, "step": 23042 }, { "epoch": 34.49550898203593, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.133, "step": 23043 }, { "epoch": 34.49700598802395, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1249, "step": 23044 }, { "epoch": 34.49850299401198, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.134, "step": 23045 }, { "epoch": 34.5, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1335, "step": 23046 }, { "epoch": 34.50149700598802, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1257, "step": 23047 }, { "epoch": 34.50299401197605, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1264, "step": 23048 }, { "epoch": 34.50449101796407, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 23049 }, { "epoch": 34.5059880239521, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 23050 }, { "epoch": 34.50748502994012, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1289, "step": 23051 }, { "epoch": 34.50898203592814, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1271, "step": 23052 }, { "epoch": 34.51047904191617, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1282, "step": 23053 }, { "epoch": 34.51197604790419, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1285, "step": 23054 }, { "epoch": 34.51347305389221, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.127, "step": 23055 }, { "epoch": 34.51497005988024, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 23056 }, { "epoch": 34.51646706586826, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1214, "step": 23057 }, { "epoch": 34.51796407185629, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.126, "step": 23058 }, { "epoch": 34.51946107784431, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1245, "step": 23059 }, { "epoch": 34.52095808383233, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1276, "step": 23060 }, { "epoch": 34.52245508982036, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1335, "step": 23061 }, { "epoch": 34.52395209580838, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1319, "step": 23062 }, { "epoch": 34.525449101796404, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1275, "step": 23063 }, { "epoch": 34.52694610778443, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1286, "step": 23064 }, { "epoch": 34.528443113772454, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1287, "step": 23065 }, { "epoch": 34.52994011976048, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1297, "step": 23066 }, { "epoch": 34.5314371257485, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1348, "step": 23067 }, { "epoch": 34.532934131736525, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1307, "step": 23068 }, { "epoch": 34.53443113772455, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1284, "step": 23069 }, { "epoch": 34.535928143712574, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1267, "step": 23070 }, { "epoch": 34.537425149700596, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1291, "step": 23071 }, { "epoch": 34.538922155688624, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1338, "step": 23072 }, { "epoch": 34.540419161676645, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1248, "step": 23073 }, { "epoch": 34.541916167664674, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1344, "step": 23074 }, { "epoch": 34.543413173652695, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1294, "step": 23075 }, { "epoch": 34.544910179640716, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1306, "step": 23076 }, { "epoch": 34.546407185628745, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1299, "step": 23077 }, { "epoch": 34.547904191616766, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1291, "step": 23078 }, { "epoch": 34.54940119760479, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1284, "step": 23079 }, { "epoch": 34.550898203592816, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.125, "step": 23080 }, { "epoch": 34.55239520958084, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1305, "step": 23081 }, { "epoch": 34.553892215568865, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1286, "step": 23082 }, { "epoch": 34.55538922155689, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1225, "step": 23083 }, { "epoch": 34.55688622754491, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1302, "step": 23084 }, { "epoch": 34.558383233532936, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 23085 }, { "epoch": 34.55988023952096, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.135, "step": 23086 }, { "epoch": 34.56137724550898, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1238, "step": 23087 }, { "epoch": 34.56287425149701, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1326, "step": 23088 }, { "epoch": 34.56437125748503, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1322, "step": 23089 }, { "epoch": 34.56586826347306, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1327, "step": 23090 }, { "epoch": 34.56736526946108, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 23091 }, { "epoch": 34.5688622754491, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1263, "step": 23092 }, { "epoch": 34.57035928143713, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1309, "step": 23093 }, { "epoch": 34.57185628742515, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1278, "step": 23094 }, { "epoch": 34.57335329341317, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1303, "step": 23095 }, { "epoch": 34.5748502994012, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1255, "step": 23096 }, { "epoch": 34.57634730538922, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1297, "step": 23097 }, { "epoch": 34.57784431137725, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1212, "step": 23098 }, { "epoch": 34.57934131736527, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1255, "step": 23099 }, { "epoch": 34.58083832335329, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1216, "step": 23100 }, { "epoch": 34.58233532934132, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1287, "step": 23101 }, { "epoch": 34.58383233532934, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1248, "step": 23102 }, { "epoch": 34.58532934131736, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1271, "step": 23103 }, { "epoch": 34.58682634730539, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.133, "step": 23104 }, { "epoch": 34.58832335329341, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 23105 }, { "epoch": 34.58982035928144, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1306, "step": 23106 }, { "epoch": 34.59131736526946, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1208, "step": 23107 }, { "epoch": 34.59281437125748, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1312, "step": 23108 }, { "epoch": 34.59431137724551, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1231, "step": 23109 }, { "epoch": 34.59580838323353, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1285, "step": 23110 }, { "epoch": 34.59730538922156, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1306, "step": 23111 }, { "epoch": 34.59880239520958, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1356, "step": 23112 }, { "epoch": 34.6002994011976, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 23113 }, { "epoch": 34.60179640718563, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.127, "step": 23114 }, { "epoch": 34.60329341317365, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1283, "step": 23115 }, { "epoch": 34.604790419161674, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1204, "step": 23116 }, { "epoch": 34.6062874251497, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1243, "step": 23117 }, { "epoch": 34.60778443113772, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1286, "step": 23118 }, { "epoch": 34.60928143712575, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1317, "step": 23119 }, { "epoch": 34.61077844311377, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.129, "step": 23120 }, { "epoch": 34.612275449101794, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1227, "step": 23121 }, { "epoch": 34.61377245508982, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1299, "step": 23122 }, { "epoch": 34.615269461077844, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1293, "step": 23123 }, { "epoch": 34.616766467065865, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1349, "step": 23124 }, { "epoch": 34.618263473053894, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1264, "step": 23125 }, { "epoch": 34.619760479041915, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 23126 }, { "epoch": 34.62125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1263, "step": 23127 }, { "epoch": 34.622754491017965, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1159, "step": 23128 }, { "epoch": 34.624251497005986, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1239, "step": 23129 }, { "epoch": 34.625748502994014, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1274, "step": 23130 }, { "epoch": 34.627245508982035, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1236, "step": 23131 }, { "epoch": 34.62874251497006, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 23132 }, { "epoch": 34.630239520958085, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1219, "step": 23133 }, { "epoch": 34.631736526946106, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1259, "step": 23134 }, { "epoch": 34.633233532934135, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1237, "step": 23135 }, { "epoch": 34.634730538922156, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1244, "step": 23136 }, { "epoch": 34.63622754491018, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.131, "step": 23137 }, { "epoch": 34.637724550898206, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1264, "step": 23138 }, { "epoch": 34.63922155688623, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1317, "step": 23139 }, { "epoch": 34.64071856287425, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1254, "step": 23140 }, { "epoch": 34.64221556886228, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1265, "step": 23141 }, { "epoch": 34.6437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 23142 }, { "epoch": 34.645209580838326, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1308, "step": 23143 }, { "epoch": 34.64670658682635, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1267, "step": 23144 }, { "epoch": 34.64820359281437, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1286, "step": 23145 }, { "epoch": 34.6497005988024, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1297, "step": 23146 }, { "epoch": 34.65119760479042, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1289, "step": 23147 }, { "epoch": 34.65269461077844, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 23148 }, { "epoch": 34.65419161676647, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1295, "step": 23149 }, { "epoch": 34.65568862275449, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1262, "step": 23150 }, { "epoch": 34.65718562874252, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1259, "step": 23151 }, { "epoch": 34.65868263473054, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1273, "step": 23152 }, { "epoch": 34.66017964071856, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1282, "step": 23153 }, { "epoch": 34.66167664670659, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1261, "step": 23154 }, { "epoch": 34.66317365269461, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1246, "step": 23155 }, { "epoch": 34.66467065868264, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1242, "step": 23156 }, { "epoch": 34.66616766467066, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1272, "step": 23157 }, { "epoch": 34.66766467065868, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1278, "step": 23158 }, { "epoch": 34.66916167664671, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1301, "step": 23159 }, { "epoch": 34.67065868263473, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1207, "step": 23160 }, { "epoch": 34.67215568862275, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1309, "step": 23161 }, { "epoch": 34.67365269461078, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1263, "step": 23162 }, { "epoch": 34.6751497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1277, "step": 23163 }, { "epoch": 34.67664670658683, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.13, "step": 23164 }, { "epoch": 34.67814371257485, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1299, "step": 23165 }, { "epoch": 34.67964071856287, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1317, "step": 23166 }, { "epoch": 34.6811377245509, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1262, "step": 23167 }, { "epoch": 34.68263473053892, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1312, "step": 23168 }, { "epoch": 34.68413173652694, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1328, "step": 23169 }, { "epoch": 34.68562874251497, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1281, "step": 23170 }, { "epoch": 34.68712574850299, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1219, "step": 23171 }, { "epoch": 34.68862275449102, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1265, "step": 23172 }, { "epoch": 34.69011976047904, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1236, "step": 23173 }, { "epoch": 34.691616766467064, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1246, "step": 23174 }, { "epoch": 34.69311377245509, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1297, "step": 23175 }, { "epoch": 34.69461077844311, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 23176 }, { "epoch": 34.696107784431135, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 23177 }, { "epoch": 34.69760479041916, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1273, "step": 23178 }, { "epoch": 34.699101796407184, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1273, "step": 23179 }, { "epoch": 34.70059880239521, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1274, "step": 23180 }, { "epoch": 34.702095808383234, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1276, "step": 23181 }, { "epoch": 34.703592814371255, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.127, "step": 23182 }, { "epoch": 34.705089820359284, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.126, "step": 23183 }, { "epoch": 34.706586826347305, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1239, "step": 23184 }, { "epoch": 34.708083832335326, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1234, "step": 23185 }, { "epoch": 34.709580838323355, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1228, "step": 23186 }, { "epoch": 34.711077844311376, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 23187 }, { "epoch": 34.712574850299404, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1259, "step": 23188 }, { "epoch": 34.714071856287426, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1235, "step": 23189 }, { "epoch": 34.71556886227545, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1346, "step": 23190 }, { "epoch": 34.717065868263475, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1313, "step": 23191 }, { "epoch": 34.7185628742515, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1276, "step": 23192 }, { "epoch": 34.72005988023952, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1324, "step": 23193 }, { "epoch": 34.721556886227546, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1306, "step": 23194 }, { "epoch": 34.72305389221557, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1341, "step": 23195 }, { "epoch": 34.724550898203596, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1286, "step": 23196 }, { "epoch": 34.72604790419162, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.128, "step": 23197 }, { "epoch": 34.72754491017964, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1219, "step": 23198 }, { "epoch": 34.72904191616767, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.13, "step": 23199 }, { "epoch": 34.73053892215569, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1329, "step": 23200 }, { "epoch": 34.73203592814371, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1345, "step": 23201 }, { "epoch": 34.73353293413174, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1353, "step": 23202 }, { "epoch": 34.73502994011976, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1255, "step": 23203 }, { "epoch": 34.73652694610779, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.1297, "step": 23204 }, { "epoch": 34.73802395209581, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.127, "step": 23205 }, { "epoch": 34.73952095808383, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1281, "step": 23206 }, { "epoch": 34.74101796407186, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1241, "step": 23207 }, { "epoch": 34.74251497005988, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1335, "step": 23208 }, { "epoch": 34.7440119760479, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1222, "step": 23209 }, { "epoch": 34.74550898203593, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1265, "step": 23210 }, { "epoch": 34.74700598802395, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1342, "step": 23211 }, { "epoch": 34.74850299401198, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.128, "step": 23212 }, { "epoch": 34.75, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1323, "step": 23213 }, { "epoch": 34.75149700598802, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1265, "step": 23214 }, { "epoch": 34.75299401197605, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1288, "step": 23215 }, { "epoch": 34.75449101796407, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 23216 }, { "epoch": 34.7559880239521, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1337, "step": 23217 }, { "epoch": 34.75748502994012, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1315, "step": 23218 }, { "epoch": 34.75898203592814, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1208, "step": 23219 }, { "epoch": 34.76047904191617, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1331, "step": 23220 }, { "epoch": 34.76197604790419, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1305, "step": 23221 }, { "epoch": 34.76347305389221, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1246, "step": 23222 }, { "epoch": 34.76497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1233, "step": 23223 }, { "epoch": 34.76646706586826, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1278, "step": 23224 }, { "epoch": 34.76796407185629, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 23225 }, { "epoch": 34.76946107784431, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1243, "step": 23226 }, { "epoch": 34.77095808383233, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1229, "step": 23227 }, { "epoch": 34.77245508982036, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1322, "step": 23228 }, { "epoch": 34.77395209580838, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.126, "step": 23229 }, { "epoch": 34.775449101796404, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1299, "step": 23230 }, { "epoch": 34.77694610778443, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 23231 }, { "epoch": 34.778443113772454, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.126, "step": 23232 }, { "epoch": 34.77994011976048, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1305, "step": 23233 }, { "epoch": 34.7814371257485, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1251, "step": 23234 }, { "epoch": 34.782934131736525, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1246, "step": 23235 }, { "epoch": 34.78443113772455, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1262, "step": 23236 }, { "epoch": 34.785928143712574, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1259, "step": 23237 }, { "epoch": 34.787425149700596, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1279, "step": 23238 }, { "epoch": 34.788922155688624, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1262, "step": 23239 }, { "epoch": 34.790419161676645, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1294, "step": 23240 }, { "epoch": 34.791916167664674, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1289, "step": 23241 }, { "epoch": 34.793413173652695, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1286, "step": 23242 }, { "epoch": 34.794910179640716, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1239, "step": 23243 }, { "epoch": 34.796407185628745, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1284, "step": 23244 }, { "epoch": 34.797904191616766, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1286, "step": 23245 }, { "epoch": 34.79940119760479, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1309, "step": 23246 }, { "epoch": 34.800898203592816, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1213, "step": 23247 }, { "epoch": 34.80239520958084, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1308, "step": 23248 }, { "epoch": 34.803892215568865, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1239, "step": 23249 }, { "epoch": 34.80538922155689, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1256, "step": 23250 }, { "epoch": 34.80688622754491, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1284, "step": 23251 }, { "epoch": 34.808383233532936, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1273, "step": 23252 }, { "epoch": 34.80988023952096, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1223, "step": 23253 }, { "epoch": 34.81137724550898, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1352, "step": 23254 }, { "epoch": 34.81287425149701, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1318, "step": 23255 }, { "epoch": 34.81437125748503, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1219, "step": 23256 }, { "epoch": 34.81586826347306, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 23257 }, { "epoch": 34.81736526946108, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1293, "step": 23258 }, { "epoch": 34.8188622754491, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1283, "step": 23259 }, { "epoch": 34.82035928143713, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1301, "step": 23260 }, { "epoch": 34.82185628742515, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1237, "step": 23261 }, { "epoch": 34.82335329341317, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1265, "step": 23262 }, { "epoch": 34.8248502994012, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1235, "step": 23263 }, { "epoch": 34.82634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1292, "step": 23264 }, { "epoch": 34.82784431137725, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1268, "step": 23265 }, { "epoch": 34.82934131736527, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1284, "step": 23266 }, { "epoch": 34.83083832335329, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1253, "step": 23267 }, { "epoch": 34.83233532934132, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1207, "step": 23268 }, { "epoch": 34.83383233532934, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1233, "step": 23269 }, { "epoch": 34.83532934131736, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1211, "step": 23270 }, { "epoch": 34.83682634730539, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1325, "step": 23271 }, { "epoch": 34.83832335329341, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1314, "step": 23272 }, { "epoch": 34.83982035928144, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1335, "step": 23273 }, { "epoch": 34.84131736526946, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1282, "step": 23274 }, { "epoch": 34.84281437125748, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1277, "step": 23275 }, { "epoch": 34.84431137724551, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1219, "step": 23276 }, { "epoch": 34.84580838323353, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1247, "step": 23277 }, { "epoch": 34.84730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1313, "step": 23278 }, { "epoch": 34.84880239520958, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1325, "step": 23279 }, { "epoch": 34.8502994011976, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1258, "step": 23280 }, { "epoch": 34.85179640718563, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1271, "step": 23281 }, { "epoch": 34.85329341317365, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 23282 }, { "epoch": 34.854790419161674, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 23283 }, { "epoch": 34.8562874251497, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1321, "step": 23284 }, { "epoch": 34.85778443113772, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1253, "step": 23285 }, { "epoch": 34.85928143712575, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 23286 }, { "epoch": 34.86077844311377, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1254, "step": 23287 }, { "epoch": 34.862275449101794, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.124, "step": 23288 }, { "epoch": 34.86377245508982, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1353, "step": 23289 }, { "epoch": 34.865269461077844, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.126, "step": 23290 }, { "epoch": 34.866766467065865, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1348, "step": 23291 }, { "epoch": 34.868263473053894, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 23292 }, { "epoch": 34.869760479041915, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1317, "step": 23293 }, { "epoch": 34.87125748502994, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1257, "step": 23294 }, { "epoch": 34.872754491017965, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1282, "step": 23295 }, { "epoch": 34.874251497005986, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1286, "step": 23296 }, { "epoch": 34.875748502994014, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1225, "step": 23297 }, { "epoch": 34.877245508982035, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.123, "step": 23298 }, { "epoch": 34.87874251497006, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1247, "step": 23299 }, { "epoch": 34.880239520958085, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.129, "step": 23300 }, { "epoch": 34.881736526946106, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1336, "step": 23301 }, { "epoch": 34.883233532934135, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1247, "step": 23302 }, { "epoch": 34.884730538922156, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1267, "step": 23303 }, { "epoch": 34.88622754491018, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1274, "step": 23304 }, { "epoch": 34.887724550898206, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1304, "step": 23305 }, { "epoch": 34.88922155688623, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1313, "step": 23306 }, { "epoch": 34.89071856287425, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.127, "step": 23307 }, { "epoch": 34.89221556886228, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1289, "step": 23308 }, { "epoch": 34.8937125748503, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1254, "step": 23309 }, { "epoch": 34.895209580838326, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1255, "step": 23310 }, { "epoch": 34.89670658682635, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1155, "step": 23311 }, { "epoch": 34.89820359281437, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1325, "step": 23312 }, { "epoch": 34.8997005988024, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 23313 }, { "epoch": 34.90119760479042, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1277, "step": 23314 }, { "epoch": 34.90269461077844, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1225, "step": 23315 }, { "epoch": 34.90419161676647, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1287, "step": 23316 }, { "epoch": 34.90568862275449, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1315, "step": 23317 }, { "epoch": 34.90718562874252, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1265, "step": 23318 }, { "epoch": 34.90868263473054, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1286, "step": 23319 }, { "epoch": 34.91017964071856, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.124, "step": 23320 }, { "epoch": 34.91167664670659, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1237, "step": 23321 }, { "epoch": 34.91317365269461, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.131, "step": 23322 }, { "epoch": 34.91467065868264, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.132, "step": 23323 }, { "epoch": 34.91616766467066, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1252, "step": 23324 }, { "epoch": 34.91766467065868, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1198, "step": 23325 }, { "epoch": 34.91916167664671, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1312, "step": 23326 }, { "epoch": 34.92065868263473, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.125, "step": 23327 }, { "epoch": 34.92215568862275, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1284, "step": 23328 }, { "epoch": 34.92365269461078, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 23329 }, { "epoch": 34.9251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1264, "step": 23330 }, { "epoch": 34.92664670658683, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.124, "step": 23331 }, { "epoch": 34.92814371257485, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1295, "step": 23332 }, { "epoch": 34.92964071856287, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1273, "step": 23333 }, { "epoch": 34.9311377245509, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.128, "step": 23334 }, { "epoch": 34.93263473053892, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1291, "step": 23335 }, { "epoch": 34.93413173652694, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1281, "step": 23336 }, { "epoch": 34.93562874251497, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1303, "step": 23337 }, { "epoch": 34.93712574850299, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1334, "step": 23338 }, { "epoch": 34.93862275449102, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1275, "step": 23339 }, { "epoch": 34.94011976047904, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1213, "step": 23340 }, { "epoch": 34.941616766467064, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1252, "step": 23341 }, { "epoch": 34.94311377245509, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1234, "step": 23342 }, { "epoch": 34.94461077844311, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1267, "step": 23343 }, { "epoch": 34.946107784431135, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1281, "step": 23344 }, { "epoch": 34.94760479041916, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1293, "step": 23345 }, { "epoch": 34.949101796407184, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1267, "step": 23346 }, { "epoch": 34.95059880239521, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.13, "step": 23347 }, { "epoch": 34.952095808383234, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1256, "step": 23348 }, { "epoch": 34.953592814371255, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1265, "step": 23349 }, { "epoch": 34.955089820359284, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1248, "step": 23350 }, { "epoch": 34.956586826347305, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1319, "step": 23351 }, { "epoch": 34.958083832335326, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1243, "step": 23352 }, { "epoch": 34.959580838323355, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1256, "step": 23353 }, { "epoch": 34.961077844311376, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1245, "step": 23354 }, { "epoch": 34.962574850299404, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1333, "step": 23355 }, { "epoch": 34.964071856287426, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.131, "step": 23356 }, { "epoch": 34.96556886227545, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1271, "step": 23357 }, { "epoch": 34.967065868263475, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1253, "step": 23358 }, { "epoch": 34.9685628742515, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1278, "step": 23359 }, { "epoch": 34.97005988023952, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1219, "step": 23360 }, { "epoch": 34.971556886227546, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1218, "step": 23361 }, { "epoch": 34.97305389221557, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1275, "step": 23362 }, { "epoch": 34.974550898203596, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1261, "step": 23363 }, { "epoch": 34.97604790419162, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.129, "step": 23364 }, { "epoch": 34.97754491017964, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1279, "step": 23365 }, { "epoch": 34.97904191616767, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1293, "step": 23366 }, { "epoch": 34.98053892215569, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1306, "step": 23367 }, { "epoch": 34.98203592814371, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1216, "step": 23368 }, { "epoch": 34.98353293413174, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1247, "step": 23369 }, { "epoch": 34.98502994011976, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1294, "step": 23370 }, { "epoch": 34.98652694610779, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1297, "step": 23371 }, { "epoch": 34.98802395209581, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1249, "step": 23372 }, { "epoch": 34.98952095808383, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1237, "step": 23373 }, { "epoch": 34.99101796407186, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.128, "step": 23374 }, { "epoch": 34.99251497005988, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1332, "step": 23375 }, { "epoch": 34.9940119760479, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1286, "step": 23376 }, { "epoch": 34.99550898203593, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1244, "step": 23377 }, { "epoch": 34.99700598802395, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1254, "step": 23378 }, { "epoch": 34.99850299401198, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 23379 }, { "epoch": 35.0, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 23380 }, { "epoch": 35.00149700598802, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.128, "step": 23381 }, { "epoch": 35.00299401197605, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.124, "step": 23382 }, { "epoch": 35.00449101796407, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1268, "step": 23383 }, { "epoch": 35.0059880239521, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1284, "step": 23384 }, { "epoch": 35.00748502994012, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1258, "step": 23385 }, { "epoch": 35.00898203592814, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1212, "step": 23386 }, { "epoch": 35.01047904191617, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1236, "step": 23387 }, { "epoch": 35.01197604790419, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 23388 }, { "epoch": 35.01347305389221, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1272, "step": 23389 }, { "epoch": 35.01497005988024, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1218, "step": 23390 }, { "epoch": 35.01646706586826, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1305, "step": 23391 }, { "epoch": 35.01796407185629, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1204, "step": 23392 }, { "epoch": 35.01946107784431, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1216, "step": 23393 }, { "epoch": 35.02095808383233, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1219, "step": 23394 }, { "epoch": 35.02245508982036, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1264, "step": 23395 }, { "epoch": 35.02395209580838, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1233, "step": 23396 }, { "epoch": 35.025449101796404, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 23397 }, { "epoch": 35.02694610778443, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1294, "step": 23398 }, { "epoch": 35.028443113772454, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1266, "step": 23399 }, { "epoch": 35.02994011976048, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1262, "step": 23400 }, { "epoch": 35.0314371257485, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1289, "step": 23401 }, { "epoch": 35.032934131736525, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.129, "step": 23402 }, { "epoch": 35.03443113772455, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1281, "step": 23403 }, { "epoch": 35.035928143712574, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1319, "step": 23404 }, { "epoch": 35.037425149700596, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1274, "step": 23405 }, { "epoch": 35.038922155688624, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1246, "step": 23406 }, { "epoch": 35.040419161676645, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1249, "step": 23407 }, { "epoch": 35.041916167664674, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1187, "step": 23408 }, { "epoch": 35.043413173652695, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1213, "step": 23409 }, { "epoch": 35.044910179640716, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1285, "step": 23410 }, { "epoch": 35.046407185628745, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1218, "step": 23411 }, { "epoch": 35.047904191616766, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1287, "step": 23412 }, { "epoch": 35.04940119760479, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 23413 }, { "epoch": 35.050898203592816, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.129, "step": 23414 }, { "epoch": 35.05239520958084, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1278, "step": 23415 }, { "epoch": 35.053892215568865, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.125, "step": 23416 }, { "epoch": 35.05538922155689, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1224, "step": 23417 }, { "epoch": 35.05688622754491, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1307, "step": 23418 }, { "epoch": 35.058383233532936, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.129, "step": 23419 }, { "epoch": 35.05988023952096, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 23420 }, { "epoch": 35.06137724550898, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1279, "step": 23421 }, { "epoch": 35.06287425149701, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1332, "step": 23422 }, { "epoch": 35.06437125748503, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1273, "step": 23423 }, { "epoch": 35.06586826347306, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1272, "step": 23424 }, { "epoch": 35.06736526946108, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1253, "step": 23425 }, { "epoch": 35.0688622754491, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1294, "step": 23426 }, { "epoch": 35.07035928143713, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1248, "step": 23427 }, { "epoch": 35.07185628742515, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1202, "step": 23428 }, { "epoch": 35.07335329341317, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1311, "step": 23429 }, { "epoch": 35.0748502994012, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1222, "step": 23430 }, { "epoch": 35.07634730538922, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1271, "step": 23431 }, { "epoch": 35.07784431137725, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1302, "step": 23432 }, { "epoch": 35.07934131736527, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1221, "step": 23433 }, { "epoch": 35.08083832335329, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1259, "step": 23434 }, { "epoch": 35.08233532934132, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1321, "step": 23435 }, { "epoch": 35.08383233532934, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1271, "step": 23436 }, { "epoch": 35.08532934131737, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1273, "step": 23437 }, { "epoch": 35.08682634730539, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1311, "step": 23438 }, { "epoch": 35.08832335329341, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1285, "step": 23439 }, { "epoch": 35.08982035928144, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1229, "step": 23440 }, { "epoch": 35.09131736526946, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1261, "step": 23441 }, { "epoch": 35.09281437125748, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1335, "step": 23442 }, { "epoch": 35.09431137724551, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.127, "step": 23443 }, { "epoch": 35.09580838323353, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1276, "step": 23444 }, { "epoch": 35.09730538922156, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1273, "step": 23445 }, { "epoch": 35.09880239520958, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 23446 }, { "epoch": 35.1002994011976, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1265, "step": 23447 }, { "epoch": 35.10179640718563, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1241, "step": 23448 }, { "epoch": 35.10329341317365, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1245, "step": 23449 }, { "epoch": 35.104790419161674, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.126, "step": 23450 }, { "epoch": 35.1062874251497, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1258, "step": 23451 }, { "epoch": 35.10778443113772, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1292, "step": 23452 }, { "epoch": 35.10928143712575, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.13, "step": 23453 }, { "epoch": 35.11077844311377, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1244, "step": 23454 }, { "epoch": 35.112275449101794, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.129, "step": 23455 }, { "epoch": 35.11377245508982, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1231, "step": 23456 }, { "epoch": 35.115269461077844, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1273, "step": 23457 }, { "epoch": 35.116766467065865, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1192, "step": 23458 }, { "epoch": 35.118263473053894, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.125, "step": 23459 }, { "epoch": 35.119760479041915, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1299, "step": 23460 }, { "epoch": 35.12125748502994, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1235, "step": 23461 }, { "epoch": 35.122754491017965, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1295, "step": 23462 }, { "epoch": 35.124251497005986, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1208, "step": 23463 }, { "epoch": 35.125748502994014, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1292, "step": 23464 }, { "epoch": 35.127245508982035, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 23465 }, { "epoch": 35.12874251497006, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.129, "step": 23466 }, { "epoch": 35.130239520958085, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1272, "step": 23467 }, { "epoch": 35.131736526946106, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1262, "step": 23468 }, { "epoch": 35.133233532934135, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1332, "step": 23469 }, { "epoch": 35.134730538922156, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.128, "step": 23470 }, { "epoch": 35.13622754491018, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1214, "step": 23471 }, { "epoch": 35.137724550898206, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1228, "step": 23472 }, { "epoch": 35.13922155688623, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.132, "step": 23473 }, { "epoch": 35.14071856287425, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1254, "step": 23474 }, { "epoch": 35.14221556886228, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1265, "step": 23475 }, { "epoch": 35.1437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1274, "step": 23476 }, { "epoch": 35.145209580838326, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 23477 }, { "epoch": 35.14670658682635, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1244, "step": 23478 }, { "epoch": 35.14820359281437, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1238, "step": 23479 }, { "epoch": 35.1497005988024, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1236, "step": 23480 }, { "epoch": 35.15119760479042, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1198, "step": 23481 }, { "epoch": 35.15269461077844, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1249, "step": 23482 }, { "epoch": 35.15419161676647, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1299, "step": 23483 }, { "epoch": 35.15568862275449, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 23484 }, { "epoch": 35.15718562874252, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 23485 }, { "epoch": 35.15868263473054, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1241, "step": 23486 }, { "epoch": 35.16017964071856, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.13, "step": 23487 }, { "epoch": 35.16167664670659, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.126, "step": 23488 }, { "epoch": 35.16317365269461, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1245, "step": 23489 }, { "epoch": 35.16467065868263, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1247, "step": 23490 }, { "epoch": 35.16616766467066, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1288, "step": 23491 }, { "epoch": 35.16766467065868, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1287, "step": 23492 }, { "epoch": 35.16916167664671, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1231, "step": 23493 }, { "epoch": 35.17065868263473, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1252, "step": 23494 }, { "epoch": 35.17215568862275, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1187, "step": 23495 }, { "epoch": 35.17365269461078, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1248, "step": 23496 }, { "epoch": 35.1751497005988, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1241, "step": 23497 }, { "epoch": 35.17664670658683, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1265, "step": 23498 }, { "epoch": 35.17814371257485, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1222, "step": 23499 }, { "epoch": 35.17964071856287, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1197, "step": 23500 }, { "epoch": 35.1811377245509, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1236, "step": 23501 }, { "epoch": 35.18263473053892, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1303, "step": 23502 }, { "epoch": 35.18413173652694, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1295, "step": 23503 }, { "epoch": 35.18562874251497, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1308, "step": 23504 }, { "epoch": 35.18712574850299, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1317, "step": 23505 }, { "epoch": 35.18862275449102, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1288, "step": 23506 }, { "epoch": 35.19011976047904, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1284, "step": 23507 }, { "epoch": 35.191616766467064, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1273, "step": 23508 }, { "epoch": 35.19311377245509, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1304, "step": 23509 }, { "epoch": 35.19461077844311, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1282, "step": 23510 }, { "epoch": 35.196107784431135, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1259, "step": 23511 }, { "epoch": 35.19760479041916, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1307, "step": 23512 }, { "epoch": 35.199101796407184, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1298, "step": 23513 }, { "epoch": 35.20059880239521, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1313, "step": 23514 }, { "epoch": 35.202095808383234, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1246, "step": 23515 }, { "epoch": 35.203592814371255, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1275, "step": 23516 }, { "epoch": 35.205089820359284, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1186, "step": 23517 }, { "epoch": 35.206586826347305, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1265, "step": 23518 }, { "epoch": 35.208083832335326, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1249, "step": 23519 }, { "epoch": 35.209580838323355, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1284, "step": 23520 }, { "epoch": 35.211077844311376, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1325, "step": 23521 }, { "epoch": 35.212574850299404, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1275, "step": 23522 }, { "epoch": 35.214071856287426, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 23523 }, { "epoch": 35.21556886227545, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.135, "step": 23524 }, { "epoch": 35.217065868263475, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1271, "step": 23525 }, { "epoch": 35.2185628742515, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1268, "step": 23526 }, { "epoch": 35.22005988023952, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 23527 }, { "epoch": 35.221556886227546, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1306, "step": 23528 }, { "epoch": 35.22305389221557, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1244, "step": 23529 }, { "epoch": 35.224550898203596, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1275, "step": 23530 }, { "epoch": 35.22604790419162, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1286, "step": 23531 }, { "epoch": 35.22754491017964, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1249, "step": 23532 }, { "epoch": 35.22904191616767, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1284, "step": 23533 }, { "epoch": 35.23053892215569, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1256, "step": 23534 }, { "epoch": 35.23203592814371, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.124, "step": 23535 }, { "epoch": 35.23353293413174, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1333, "step": 23536 }, { "epoch": 35.23502994011976, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1256, "step": 23537 }, { "epoch": 35.23652694610779, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1323, "step": 23538 }, { "epoch": 35.23802395209581, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1329, "step": 23539 }, { "epoch": 35.23952095808383, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1347, "step": 23540 }, { "epoch": 35.24101796407186, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 23541 }, { "epoch": 35.24251497005988, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1316, "step": 23542 }, { "epoch": 35.2440119760479, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1312, "step": 23543 }, { "epoch": 35.24550898203593, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1191, "step": 23544 }, { "epoch": 35.24700598802395, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 23545 }, { "epoch": 35.24850299401198, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1245, "step": 23546 }, { "epoch": 35.25, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1252, "step": 23547 }, { "epoch": 35.25149700598802, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1223, "step": 23548 }, { "epoch": 35.25299401197605, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1211, "step": 23549 }, { "epoch": 35.25449101796407, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1274, "step": 23550 }, { "epoch": 35.2559880239521, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1215, "step": 23551 }, { "epoch": 35.25748502994012, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1285, "step": 23552 }, { "epoch": 35.25898203592814, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1352, "step": 23553 }, { "epoch": 35.26047904191617, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1297, "step": 23554 }, { "epoch": 35.26197604790419, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1211, "step": 23555 }, { "epoch": 35.26347305389221, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1288, "step": 23556 }, { "epoch": 35.26497005988024, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1287, "step": 23557 }, { "epoch": 35.26646706586826, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1204, "step": 23558 }, { "epoch": 35.26796407185629, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 23559 }, { "epoch": 35.26946107784431, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1296, "step": 23560 }, { "epoch": 35.27095808383233, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1304, "step": 23561 }, { "epoch": 35.27245508982036, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1274, "step": 23562 }, { "epoch": 35.27395209580838, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1214, "step": 23563 }, { "epoch": 35.275449101796404, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1263, "step": 23564 }, { "epoch": 35.27694610778443, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1194, "step": 23565 }, { "epoch": 35.278443113772454, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1253, "step": 23566 }, { "epoch": 35.27994011976048, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1286, "step": 23567 }, { "epoch": 35.2814371257485, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1349, "step": 23568 }, { "epoch": 35.282934131736525, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1273, "step": 23569 }, { "epoch": 35.28443113772455, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1263, "step": 23570 }, { "epoch": 35.285928143712574, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1245, "step": 23571 }, { "epoch": 35.287425149700596, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1333, "step": 23572 }, { "epoch": 35.288922155688624, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1222, "step": 23573 }, { "epoch": 35.290419161676645, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1295, "step": 23574 }, { "epoch": 35.291916167664674, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.127, "step": 23575 }, { "epoch": 35.293413173652695, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1322, "step": 23576 }, { "epoch": 35.294910179640716, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1253, "step": 23577 }, { "epoch": 35.296407185628745, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1278, "step": 23578 }, { "epoch": 35.297904191616766, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.132, "step": 23579 }, { "epoch": 35.29940119760479, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1277, "step": 23580 }, { "epoch": 35.300898203592816, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1249, "step": 23581 }, { "epoch": 35.30239520958084, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1236, "step": 23582 }, { "epoch": 35.303892215568865, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1275, "step": 23583 }, { "epoch": 35.30538922155689, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1231, "step": 23584 }, { "epoch": 35.30688622754491, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1327, "step": 23585 }, { "epoch": 35.308383233532936, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1325, "step": 23586 }, { "epoch": 35.30988023952096, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1295, "step": 23587 }, { "epoch": 35.31137724550898, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 23588 }, { "epoch": 35.31287425149701, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.131, "step": 23589 }, { "epoch": 35.31437125748503, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1326, "step": 23590 }, { "epoch": 35.31586826347306, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1284, "step": 23591 }, { "epoch": 35.31736526946108, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1299, "step": 23592 }, { "epoch": 35.3188622754491, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1285, "step": 23593 }, { "epoch": 35.32035928143713, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1272, "step": 23594 }, { "epoch": 35.32185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1352, "step": 23595 }, { "epoch": 35.32335329341317, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1273, "step": 23596 }, { "epoch": 35.3248502994012, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 23597 }, { "epoch": 35.32634730538922, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1279, "step": 23598 }, { "epoch": 35.32784431137725, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1242, "step": 23599 }, { "epoch": 35.32934131736527, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.127, "step": 23600 }, { "epoch": 35.33083832335329, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1307, "step": 23601 }, { "epoch": 35.33233532934132, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1314, "step": 23602 }, { "epoch": 35.33383233532934, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1289, "step": 23603 }, { "epoch": 35.33532934131736, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1288, "step": 23604 }, { "epoch": 35.33682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1235, "step": 23605 }, { "epoch": 35.33832335329341, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1278, "step": 23606 }, { "epoch": 35.33982035928144, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1319, "step": 23607 }, { "epoch": 35.34131736526946, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1329, "step": 23608 }, { "epoch": 35.34281437125748, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1303, "step": 23609 }, { "epoch": 35.34431137724551, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1247, "step": 23610 }, { "epoch": 35.34580838323353, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1194, "step": 23611 }, { "epoch": 35.34730538922156, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1254, "step": 23612 }, { "epoch": 35.34880239520958, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1239, "step": 23613 }, { "epoch": 35.3502994011976, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1252, "step": 23614 }, { "epoch": 35.35179640718563, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1263, "step": 23615 }, { "epoch": 35.35329341317365, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1276, "step": 23616 }, { "epoch": 35.354790419161674, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1249, "step": 23617 }, { "epoch": 35.3562874251497, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1262, "step": 23618 }, { "epoch": 35.35778443113772, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1303, "step": 23619 }, { "epoch": 35.35928143712575, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1257, "step": 23620 }, { "epoch": 35.36077844311377, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.132, "step": 23621 }, { "epoch": 35.362275449101794, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1272, "step": 23622 }, { "epoch": 35.36377245508982, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1253, "step": 23623 }, { "epoch": 35.365269461077844, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1261, "step": 23624 }, { "epoch": 35.366766467065865, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1315, "step": 23625 }, { "epoch": 35.368263473053894, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1224, "step": 23626 }, { "epoch": 35.369760479041915, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1273, "step": 23627 }, { "epoch": 35.37125748502994, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1273, "step": 23628 }, { "epoch": 35.372754491017965, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.124, "step": 23629 }, { "epoch": 35.374251497005986, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1195, "step": 23630 }, { "epoch": 35.375748502994014, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.128, "step": 23631 }, { "epoch": 35.377245508982035, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1316, "step": 23632 }, { "epoch": 35.37874251497006, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.124, "step": 23633 }, { "epoch": 35.380239520958085, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 23634 }, { "epoch": 35.381736526946106, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1299, "step": 23635 }, { "epoch": 35.383233532934135, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1223, "step": 23636 }, { "epoch": 35.384730538922156, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.138, "step": 23637 }, { "epoch": 35.38622754491018, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1256, "step": 23638 }, { "epoch": 35.387724550898206, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.124, "step": 23639 }, { "epoch": 35.38922155688623, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1289, "step": 23640 }, { "epoch": 35.39071856287425, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1276, "step": 23641 }, { "epoch": 35.39221556886228, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.132, "step": 23642 }, { "epoch": 35.3937125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1245, "step": 23643 }, { "epoch": 35.395209580838326, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1276, "step": 23644 }, { "epoch": 35.39670658682635, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1235, "step": 23645 }, { "epoch": 35.39820359281437, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1254, "step": 23646 }, { "epoch": 35.3997005988024, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1315, "step": 23647 }, { "epoch": 35.40119760479042, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.129, "step": 23648 }, { "epoch": 35.40269461077844, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1275, "step": 23649 }, { "epoch": 35.40419161676647, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1294, "step": 23650 }, { "epoch": 35.40568862275449, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1294, "step": 23651 }, { "epoch": 35.40718562874252, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1268, "step": 23652 }, { "epoch": 35.40868263473054, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1213, "step": 23653 }, { "epoch": 35.41017964071856, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1256, "step": 23654 }, { "epoch": 35.41167664670659, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1189, "step": 23655 }, { "epoch": 35.41317365269461, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1276, "step": 23656 }, { "epoch": 35.41467065868264, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1135, "step": 23657 }, { "epoch": 35.41616766467066, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1262, "step": 23658 }, { "epoch": 35.41766467065868, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1295, "step": 23659 }, { "epoch": 35.41916167664671, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1325, "step": 23660 }, { "epoch": 35.42065868263473, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1225, "step": 23661 }, { "epoch": 35.42215568862275, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1213, "step": 23662 }, { "epoch": 35.42365269461078, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1292, "step": 23663 }, { "epoch": 35.4251497005988, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1316, "step": 23664 }, { "epoch": 35.42664670658683, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1303, "step": 23665 }, { "epoch": 35.42814371257485, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1262, "step": 23666 }, { "epoch": 35.42964071856287, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1308, "step": 23667 }, { "epoch": 35.4311377245509, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1317, "step": 23668 }, { "epoch": 35.43263473053892, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1203, "step": 23669 }, { "epoch": 35.43413173652694, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1324, "step": 23670 }, { "epoch": 35.43562874251497, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1246, "step": 23671 }, { "epoch": 35.43712574850299, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1253, "step": 23672 }, { "epoch": 35.43862275449102, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1263, "step": 23673 }, { "epoch": 35.44011976047904, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 23674 }, { "epoch": 35.441616766467064, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1262, "step": 23675 }, { "epoch": 35.44311377245509, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1364, "step": 23676 }, { "epoch": 35.44461077844311, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.125, "step": 23677 }, { "epoch": 35.446107784431135, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1328, "step": 23678 }, { "epoch": 35.44760479041916, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.129, "step": 23679 }, { "epoch": 35.449101796407184, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1248, "step": 23680 }, { "epoch": 35.45059880239521, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1245, "step": 23681 }, { "epoch": 35.452095808383234, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1243, "step": 23682 }, { "epoch": 35.453592814371255, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1262, "step": 23683 }, { "epoch": 35.455089820359284, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.132, "step": 23684 }, { "epoch": 35.456586826347305, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1297, "step": 23685 }, { "epoch": 35.458083832335326, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 23686 }, { "epoch": 35.459580838323355, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1339, "step": 23687 }, { "epoch": 35.461077844311376, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1304, "step": 23688 }, { "epoch": 35.462574850299404, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1246, "step": 23689 }, { "epoch": 35.464071856287426, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 23690 }, { "epoch": 35.46556886227545, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1291, "step": 23691 }, { "epoch": 35.467065868263475, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1252, "step": 23692 }, { "epoch": 35.4685628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1245, "step": 23693 }, { "epoch": 35.47005988023952, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1191, "step": 23694 }, { "epoch": 35.471556886227546, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1249, "step": 23695 }, { "epoch": 35.47305389221557, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.129, "step": 23696 }, { "epoch": 35.474550898203596, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1233, "step": 23697 }, { "epoch": 35.47604790419162, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1286, "step": 23698 }, { "epoch": 35.47754491017964, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1252, "step": 23699 }, { "epoch": 35.47904191616767, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.133, "step": 23700 }, { "epoch": 35.48053892215569, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1265, "step": 23701 }, { "epoch": 35.48203592814371, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1256, "step": 23702 }, { "epoch": 35.48353293413174, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1275, "step": 23703 }, { "epoch": 35.48502994011976, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1282, "step": 23704 }, { "epoch": 35.48652694610779, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1239, "step": 23705 }, { "epoch": 35.48802395209581, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1245, "step": 23706 }, { "epoch": 35.48952095808383, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.133, "step": 23707 }, { "epoch": 35.49101796407186, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1319, "step": 23708 }, { "epoch": 35.49251497005988, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1285, "step": 23709 }, { "epoch": 35.4940119760479, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1249, "step": 23710 }, { "epoch": 35.49550898203593, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1291, "step": 23711 }, { "epoch": 35.49700598802395, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1283, "step": 23712 }, { "epoch": 35.49850299401198, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.129, "step": 23713 }, { "epoch": 35.5, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1221, "step": 23714 }, { "epoch": 35.50149700598802, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1259, "step": 23715 }, { "epoch": 35.50299401197605, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1268, "step": 23716 }, { "epoch": 35.50449101796407, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1291, "step": 23717 }, { "epoch": 35.5059880239521, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 23718 }, { "epoch": 35.50748502994012, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1311, "step": 23719 }, { "epoch": 35.50898203592814, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.131, "step": 23720 }, { "epoch": 35.51047904191617, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1283, "step": 23721 }, { "epoch": 35.51197604790419, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1271, "step": 23722 }, { "epoch": 35.51347305389221, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1245, "step": 23723 }, { "epoch": 35.51497005988024, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1211, "step": 23724 }, { "epoch": 35.51646706586826, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1292, "step": 23725 }, { "epoch": 35.51796407185629, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1193, "step": 23726 }, { "epoch": 35.51946107784431, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1315, "step": 23727 }, { "epoch": 35.52095808383233, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1285, "step": 23728 }, { "epoch": 35.52245508982036, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 23729 }, { "epoch": 35.52395209580838, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1222, "step": 23730 }, { "epoch": 35.525449101796404, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 23731 }, { "epoch": 35.52694610778443, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 23732 }, { "epoch": 35.528443113772454, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1269, "step": 23733 }, { "epoch": 35.52994011976048, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1256, "step": 23734 }, { "epoch": 35.5314371257485, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1269, "step": 23735 }, { "epoch": 35.532934131736525, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1257, "step": 23736 }, { "epoch": 35.53443113772455, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1251, "step": 23737 }, { "epoch": 35.535928143712574, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1264, "step": 23738 }, { "epoch": 35.537425149700596, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1289, "step": 23739 }, { "epoch": 35.538922155688624, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 23740 }, { "epoch": 35.540419161676645, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1258, "step": 23741 }, { "epoch": 35.541916167664674, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.126, "step": 23742 }, { "epoch": 35.543413173652695, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1239, "step": 23743 }, { "epoch": 35.544910179640716, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1285, "step": 23744 }, { "epoch": 35.546407185628745, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1335, "step": 23745 }, { "epoch": 35.547904191616766, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1251, "step": 23746 }, { "epoch": 35.54940119760479, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 23747 }, { "epoch": 35.550898203592816, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 23748 }, { "epoch": 35.55239520958084, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1255, "step": 23749 }, { "epoch": 35.553892215568865, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1191, "step": 23750 }, { "epoch": 35.55538922155689, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1308, "step": 23751 }, { "epoch": 35.55688622754491, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1236, "step": 23752 }, { "epoch": 35.558383233532936, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1309, "step": 23753 }, { "epoch": 35.55988023952096, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1309, "step": 23754 }, { "epoch": 35.56137724550898, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1262, "step": 23755 }, { "epoch": 35.56287425149701, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.123, "step": 23756 }, { "epoch": 35.56437125748503, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1338, "step": 23757 }, { "epoch": 35.56586826347306, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1339, "step": 23758 }, { "epoch": 35.56736526946108, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1255, "step": 23759 }, { "epoch": 35.5688622754491, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1244, "step": 23760 }, { "epoch": 35.57035928143713, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1315, "step": 23761 }, { "epoch": 35.57185628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.13, "step": 23762 }, { "epoch": 35.57335329341317, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1228, "step": 23763 }, { "epoch": 35.5748502994012, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1261, "step": 23764 }, { "epoch": 35.57634730538922, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.119, "step": 23765 }, { "epoch": 35.57784431137725, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1231, "step": 23766 }, { "epoch": 35.57934131736527, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.124, "step": 23767 }, { "epoch": 35.58083832335329, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1212, "step": 23768 }, { "epoch": 35.58233532934132, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1239, "step": 23769 }, { "epoch": 35.58383233532934, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1255, "step": 23770 }, { "epoch": 35.58532934131736, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1278, "step": 23771 }, { "epoch": 35.58682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1234, "step": 23772 }, { "epoch": 35.58832335329341, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1223, "step": 23773 }, { "epoch": 35.58982035928144, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 23774 }, { "epoch": 35.59131736526946, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1272, "step": 23775 }, { "epoch": 35.59281437125748, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1177, "step": 23776 }, { "epoch": 35.59431137724551, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1243, "step": 23777 }, { "epoch": 35.59580838323353, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1227, "step": 23778 }, { "epoch": 35.59730538922156, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1216, "step": 23779 }, { "epoch": 35.59880239520958, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.133, "step": 23780 }, { "epoch": 35.6002994011976, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1274, "step": 23781 }, { "epoch": 35.60179640718563, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.13, "step": 23782 }, { "epoch": 35.60329341317365, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1226, "step": 23783 }, { "epoch": 35.604790419161674, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1295, "step": 23784 }, { "epoch": 35.6062874251497, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1238, "step": 23785 }, { "epoch": 35.60778443113772, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1341, "step": 23786 }, { "epoch": 35.60928143712575, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1323, "step": 23787 }, { "epoch": 35.61077844311377, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1287, "step": 23788 }, { "epoch": 35.612275449101794, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1192, "step": 23789 }, { "epoch": 35.61377245508982, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1284, "step": 23790 }, { "epoch": 35.615269461077844, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1319, "step": 23791 }, { "epoch": 35.616766467065865, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1273, "step": 23792 }, { "epoch": 35.618263473053894, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1194, "step": 23793 }, { "epoch": 35.619760479041915, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1242, "step": 23794 }, { "epoch": 35.62125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1265, "step": 23795 }, { "epoch": 35.622754491017965, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1312, "step": 23796 }, { "epoch": 35.624251497005986, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1216, "step": 23797 }, { "epoch": 35.625748502994014, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1231, "step": 23798 }, { "epoch": 35.627245508982035, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 23799 }, { "epoch": 35.62874251497006, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1274, "step": 23800 }, { "epoch": 35.630239520958085, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1262, "step": 23801 }, { "epoch": 35.631736526946106, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1272, "step": 23802 }, { "epoch": 35.633233532934135, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 23803 }, { "epoch": 35.634730538922156, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1262, "step": 23804 }, { "epoch": 35.63622754491018, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1269, "step": 23805 }, { "epoch": 35.637724550898206, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1266, "step": 23806 }, { "epoch": 35.63922155688623, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.13, "step": 23807 }, { "epoch": 35.64071856287425, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1237, "step": 23808 }, { "epoch": 35.64221556886228, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1218, "step": 23809 }, { "epoch": 35.6437125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 23810 }, { "epoch": 35.645209580838326, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1276, "step": 23811 }, { "epoch": 35.64670658682635, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1285, "step": 23812 }, { "epoch": 35.64820359281437, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1341, "step": 23813 }, { "epoch": 35.6497005988024, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1234, "step": 23814 }, { "epoch": 35.65119760479042, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1276, "step": 23815 }, { "epoch": 35.65269461077844, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.126, "step": 23816 }, { "epoch": 35.65419161676647, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1312, "step": 23817 }, { "epoch": 35.65568862275449, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1289, "step": 23818 }, { "epoch": 35.65718562874252, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1313, "step": 23819 }, { "epoch": 35.65868263473054, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1254, "step": 23820 }, { "epoch": 35.66017964071856, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1261, "step": 23821 }, { "epoch": 35.66167664670659, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1264, "step": 23822 }, { "epoch": 35.66317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1289, "step": 23823 }, { "epoch": 35.66467065868264, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1285, "step": 23824 }, { "epoch": 35.66616766467066, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1345, "step": 23825 }, { "epoch": 35.66766467065868, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1269, "step": 23826 }, { "epoch": 35.66916167664671, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1272, "step": 23827 }, { "epoch": 35.67065868263473, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1294, "step": 23828 }, { "epoch": 35.67215568862275, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1298, "step": 23829 }, { "epoch": 35.67365269461078, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1262, "step": 23830 }, { "epoch": 35.6751497005988, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1278, "step": 23831 }, { "epoch": 35.67664670658683, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1239, "step": 23832 }, { "epoch": 35.67814371257485, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1283, "step": 23833 }, { "epoch": 35.67964071856287, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1298, "step": 23834 }, { "epoch": 35.6811377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1264, "step": 23835 }, { "epoch": 35.68263473053892, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1263, "step": 23836 }, { "epoch": 35.68413173652694, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1265, "step": 23837 }, { "epoch": 35.68562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1267, "step": 23838 }, { "epoch": 35.68712574850299, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1306, "step": 23839 }, { "epoch": 35.68862275449102, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1271, "step": 23840 }, { "epoch": 35.69011976047904, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1264, "step": 23841 }, { "epoch": 35.691616766467064, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1231, "step": 23842 }, { "epoch": 35.69311377245509, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1295, "step": 23843 }, { "epoch": 35.69461077844311, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1306, "step": 23844 }, { "epoch": 35.696107784431135, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1294, "step": 23845 }, { "epoch": 35.69760479041916, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 23846 }, { "epoch": 35.699101796407184, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1259, "step": 23847 }, { "epoch": 35.70059880239521, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1277, "step": 23848 }, { "epoch": 35.702095808383234, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1332, "step": 23849 }, { "epoch": 35.703592814371255, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1272, "step": 23850 }, { "epoch": 35.705089820359284, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1248, "step": 23851 }, { "epoch": 35.706586826347305, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1244, "step": 23852 }, { "epoch": 35.708083832335326, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.13, "step": 23853 }, { "epoch": 35.709580838323355, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.132, "step": 23854 }, { "epoch": 35.711077844311376, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1305, "step": 23855 }, { "epoch": 35.712574850299404, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1277, "step": 23856 }, { "epoch": 35.714071856287426, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1266, "step": 23857 }, { "epoch": 35.71556886227545, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1291, "step": 23858 }, { "epoch": 35.717065868263475, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1307, "step": 23859 }, { "epoch": 35.7185628742515, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1269, "step": 23860 }, { "epoch": 35.72005988023952, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1266, "step": 23861 }, { "epoch": 35.721556886227546, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1279, "step": 23862 }, { "epoch": 35.72305389221557, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1263, "step": 23863 }, { "epoch": 35.724550898203596, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1268, "step": 23864 }, { "epoch": 35.72604790419162, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1268, "step": 23865 }, { "epoch": 35.72754491017964, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1247, "step": 23866 }, { "epoch": 35.72904191616767, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1251, "step": 23867 }, { "epoch": 35.73053892215569, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1256, "step": 23868 }, { "epoch": 35.73203592814371, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1302, "step": 23869 }, { "epoch": 35.73353293413174, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1202, "step": 23870 }, { "epoch": 35.73502994011976, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 23871 }, { "epoch": 35.73652694610779, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1242, "step": 23872 }, { "epoch": 35.73802395209581, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.127, "step": 23873 }, { "epoch": 35.73952095808383, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1259, "step": 23874 }, { "epoch": 35.74101796407186, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1234, "step": 23875 }, { "epoch": 35.74251497005988, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1245, "step": 23876 }, { "epoch": 35.7440119760479, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1206, "step": 23877 }, { "epoch": 35.74550898203593, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1222, "step": 23878 }, { "epoch": 35.74700598802395, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 23879 }, { "epoch": 35.74850299401198, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1326, "step": 23880 }, { "epoch": 35.75, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1243, "step": 23881 }, { "epoch": 35.75149700598802, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 23882 }, { "epoch": 35.75299401197605, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1329, "step": 23883 }, { "epoch": 35.75449101796407, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1261, "step": 23884 }, { "epoch": 35.7559880239521, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1275, "step": 23885 }, { "epoch": 35.75748502994012, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1222, "step": 23886 }, { "epoch": 35.75898203592814, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1241, "step": 23887 }, { "epoch": 35.76047904191617, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1265, "step": 23888 }, { "epoch": 35.76197604790419, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1255, "step": 23889 }, { "epoch": 35.76347305389221, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1297, "step": 23890 }, { "epoch": 35.76497005988024, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1322, "step": 23891 }, { "epoch": 35.76646706586826, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 23892 }, { "epoch": 35.76796407185629, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1209, "step": 23893 }, { "epoch": 35.76946107784431, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1294, "step": 23894 }, { "epoch": 35.77095808383233, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1234, "step": 23895 }, { "epoch": 35.77245508982036, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1287, "step": 23896 }, { "epoch": 35.77395209580838, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1189, "step": 23897 }, { "epoch": 35.775449101796404, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1292, "step": 23898 }, { "epoch": 35.77694610778443, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 23899 }, { "epoch": 35.778443113772454, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1258, "step": 23900 }, { "epoch": 35.77994011976048, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1241, "step": 23901 }, { "epoch": 35.7814371257485, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1251, "step": 23902 }, { "epoch": 35.782934131736525, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1303, "step": 23903 }, { "epoch": 35.78443113772455, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1333, "step": 23904 }, { "epoch": 35.785928143712574, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1266, "step": 23905 }, { "epoch": 35.787425149700596, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1284, "step": 23906 }, { "epoch": 35.788922155688624, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1216, "step": 23907 }, { "epoch": 35.790419161676645, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1321, "step": 23908 }, { "epoch": 35.791916167664674, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1321, "step": 23909 }, { "epoch": 35.793413173652695, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1284, "step": 23910 }, { "epoch": 35.794910179640716, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.125, "step": 23911 }, { "epoch": 35.796407185628745, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1264, "step": 23912 }, { "epoch": 35.797904191616766, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1287, "step": 23913 }, { "epoch": 35.79940119760479, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1262, "step": 23914 }, { "epoch": 35.800898203592816, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1291, "step": 23915 }, { "epoch": 35.80239520958084, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1254, "step": 23916 }, { "epoch": 35.803892215568865, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.124, "step": 23917 }, { "epoch": 35.80538922155689, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1273, "step": 23918 }, { "epoch": 35.80688622754491, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 23919 }, { "epoch": 35.808383233532936, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.125, "step": 23920 }, { "epoch": 35.80988023952096, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.125, "step": 23921 }, { "epoch": 35.81137724550898, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1266, "step": 23922 }, { "epoch": 35.81287425149701, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1244, "step": 23923 }, { "epoch": 35.81437125748503, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1276, "step": 23924 }, { "epoch": 35.81586826347306, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1216, "step": 23925 }, { "epoch": 35.81736526946108, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1178, "step": 23926 }, { "epoch": 35.8188622754491, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1248, "step": 23927 }, { "epoch": 35.82035928143713, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1246, "step": 23928 }, { "epoch": 35.82185628742515, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1303, "step": 23929 }, { "epoch": 35.82335329341317, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1279, "step": 23930 }, { "epoch": 35.8248502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1266, "step": 23931 }, { "epoch": 35.82634730538922, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1221, "step": 23932 }, { "epoch": 35.82784431137725, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1227, "step": 23933 }, { "epoch": 35.82934131736527, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1287, "step": 23934 }, { "epoch": 35.83083832335329, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.123, "step": 23935 }, { "epoch": 35.83233532934132, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1293, "step": 23936 }, { "epoch": 35.83383233532934, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1272, "step": 23937 }, { "epoch": 35.83532934131736, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1244, "step": 23938 }, { "epoch": 35.83682634730539, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1203, "step": 23939 }, { "epoch": 35.83832335329341, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1233, "step": 23940 }, { "epoch": 35.83982035928144, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1285, "step": 23941 }, { "epoch": 35.84131736526946, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1299, "step": 23942 }, { "epoch": 35.84281437125748, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.124, "step": 23943 }, { "epoch": 35.84431137724551, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.132, "step": 23944 }, { "epoch": 35.84580838323353, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1224, "step": 23945 }, { "epoch": 35.84730538922156, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1265, "step": 23946 }, { "epoch": 35.84880239520958, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1312, "step": 23947 }, { "epoch": 35.8502994011976, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1271, "step": 23948 }, { "epoch": 35.85179640718563, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1205, "step": 23949 }, { "epoch": 35.85329341317365, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1261, "step": 23950 }, { "epoch": 35.854790419161674, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.131, "step": 23951 }, { "epoch": 35.8562874251497, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1284, "step": 23952 }, { "epoch": 35.85778443113772, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1261, "step": 23953 }, { "epoch": 35.85928143712575, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1211, "step": 23954 }, { "epoch": 35.86077844311377, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1266, "step": 23955 }, { "epoch": 35.862275449101794, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1276, "step": 23956 }, { "epoch": 35.86377245508982, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1333, "step": 23957 }, { "epoch": 35.865269461077844, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.129, "step": 23958 }, { "epoch": 35.866766467065865, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1263, "step": 23959 }, { "epoch": 35.868263473053894, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1314, "step": 23960 }, { "epoch": 35.869760479041915, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1275, "step": 23961 }, { "epoch": 35.87125748502994, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1249, "step": 23962 }, { "epoch": 35.872754491017965, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1265, "step": 23963 }, { "epoch": 35.874251497005986, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1225, "step": 23964 }, { "epoch": 35.875748502994014, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1306, "step": 23965 }, { "epoch": 35.877245508982035, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1272, "step": 23966 }, { "epoch": 35.87874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1277, "step": 23967 }, { "epoch": 35.880239520958085, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1241, "step": 23968 }, { "epoch": 35.881736526946106, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1335, "step": 23969 }, { "epoch": 35.883233532934135, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1273, "step": 23970 }, { "epoch": 35.884730538922156, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1268, "step": 23971 }, { "epoch": 35.88622754491018, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1234, "step": 23972 }, { "epoch": 35.887724550898206, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1282, "step": 23973 }, { "epoch": 35.88922155688623, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1283, "step": 23974 }, { "epoch": 35.89071856287425, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1301, "step": 23975 }, { "epoch": 35.89221556886228, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1333, "step": 23976 }, { "epoch": 35.8937125748503, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1285, "step": 23977 }, { "epoch": 35.895209580838326, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 23978 }, { "epoch": 35.89670658682635, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1184, "step": 23979 }, { "epoch": 35.89820359281437, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1237, "step": 23980 }, { "epoch": 35.8997005988024, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1256, "step": 23981 }, { "epoch": 35.90119760479042, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1227, "step": 23982 }, { "epoch": 35.90269461077844, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1268, "step": 23983 }, { "epoch": 35.90419161676647, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1243, "step": 23984 }, { "epoch": 35.90568862275449, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.13, "step": 23985 }, { "epoch": 35.90718562874252, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1236, "step": 23986 }, { "epoch": 35.90868263473054, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1234, "step": 23987 }, { "epoch": 35.91017964071856, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1293, "step": 23988 }, { "epoch": 35.91167664670659, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1277, "step": 23989 }, { "epoch": 35.91317365269461, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1245, "step": 23990 }, { "epoch": 35.91467065868264, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1235, "step": 23991 }, { "epoch": 35.91616766467066, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1275, "step": 23992 }, { "epoch": 35.91766467065868, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1199, "step": 23993 }, { "epoch": 35.91916167664671, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1314, "step": 23994 }, { "epoch": 35.92065868263473, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 23995 }, { "epoch": 35.92215568862275, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1342, "step": 23996 }, { "epoch": 35.92365269461078, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1263, "step": 23997 }, { "epoch": 35.9251497005988, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1265, "step": 23998 }, { "epoch": 35.92664670658683, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 23999 }, { "epoch": 35.92814371257485, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1311, "step": 24000 }, { "epoch": 35.92964071856287, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.139, "step": 24001 }, { "epoch": 35.9311377245509, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 24002 }, { "epoch": 35.93263473053892, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.133, "step": 24003 }, { "epoch": 35.93413173652694, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1274, "step": 24004 }, { "epoch": 35.93562874251497, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1226, "step": 24005 }, { "epoch": 35.93712574850299, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1227, "step": 24006 }, { "epoch": 35.93862275449102, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.131, "step": 24007 }, { "epoch": 35.94011976047904, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1296, "step": 24008 }, { "epoch": 35.941616766467064, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1221, "step": 24009 }, { "epoch": 35.94311377245509, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1218, "step": 24010 }, { "epoch": 35.94461077844311, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1292, "step": 24011 }, { "epoch": 35.946107784431135, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1256, "step": 24012 }, { "epoch": 35.94760479041916, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1302, "step": 24013 }, { "epoch": 35.949101796407184, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1307, "step": 24014 }, { "epoch": 35.95059880239521, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1215, "step": 24015 }, { "epoch": 35.952095808383234, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1233, "step": 24016 }, { "epoch": 35.953592814371255, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.125, "step": 24017 }, { "epoch": 35.955089820359284, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1295, "step": 24018 }, { "epoch": 35.956586826347305, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1234, "step": 24019 }, { "epoch": 35.958083832335326, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1201, "step": 24020 }, { "epoch": 35.959580838323355, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1211, "step": 24021 }, { "epoch": 35.961077844311376, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.122, "step": 24022 }, { "epoch": 35.962574850299404, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1244, "step": 24023 }, { "epoch": 35.964071856287426, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1262, "step": 24024 }, { "epoch": 35.96556886227545, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1254, "step": 24025 }, { "epoch": 35.967065868263475, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1275, "step": 24026 }, { "epoch": 35.9685628742515, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1377, "step": 24027 }, { "epoch": 35.97005988023952, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1259, "step": 24028 }, { "epoch": 35.971556886227546, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1253, "step": 24029 }, { "epoch": 35.97305389221557, "grad_norm": 0.04931640625, "learning_rate": 0.0008, "loss": 1.1244, "step": 24030 }, { "epoch": 35.974550898203596, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1284, "step": 24031 }, { "epoch": 35.97604790419162, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1241, "step": 24032 }, { "epoch": 35.97754491017964, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1293, "step": 24033 }, { "epoch": 35.97904191616767, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.12, "step": 24034 }, { "epoch": 35.98053892215569, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.127, "step": 24035 }, { "epoch": 35.98203592814371, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1265, "step": 24036 }, { "epoch": 35.98353293413174, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 24037 }, { "epoch": 35.98502994011976, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1238, "step": 24038 }, { "epoch": 35.98652694610779, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 24039 }, { "epoch": 35.98802395209581, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.124, "step": 24040 }, { "epoch": 35.98952095808383, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1198, "step": 24041 }, { "epoch": 35.99101796407186, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1264, "step": 24042 }, { "epoch": 35.99251497005988, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1295, "step": 24043 }, { "epoch": 35.9940119760479, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1216, "step": 24044 }, { "epoch": 35.99550898203593, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1193, "step": 24045 }, { "epoch": 35.99700598802395, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1309, "step": 24046 }, { "epoch": 35.99850299401198, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1279, "step": 24047 }, { "epoch": 36.0, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1256, "step": 24048 }, { "epoch": 36.00149700598802, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1292, "step": 24049 }, { "epoch": 36.00299401197605, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1273, "step": 24050 }, { "epoch": 36.00449101796407, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1232, "step": 24051 }, { "epoch": 36.0059880239521, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1234, "step": 24052 }, { "epoch": 36.00748502994012, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1271, "step": 24053 }, { "epoch": 36.00898203592814, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1308, "step": 24054 }, { "epoch": 36.01047904191617, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1254, "step": 24055 }, { "epoch": 36.01197604790419, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1269, "step": 24056 }, { "epoch": 36.01347305389221, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1217, "step": 24057 }, { "epoch": 36.01497005988024, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1259, "step": 24058 }, { "epoch": 36.01646706586826, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1315, "step": 24059 }, { "epoch": 36.01796407185629, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1288, "step": 24060 }, { "epoch": 36.01946107784431, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1281, "step": 24061 }, { "epoch": 36.02095808383233, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1188, "step": 24062 }, { "epoch": 36.02245508982036, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.127, "step": 24063 }, { "epoch": 36.02395209580838, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1233, "step": 24064 }, { "epoch": 36.025449101796404, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1261, "step": 24065 }, { "epoch": 36.02694610778443, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.126, "step": 24066 }, { "epoch": 36.028443113772454, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.124, "step": 24067 }, { "epoch": 36.02994011976048, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1282, "step": 24068 }, { "epoch": 36.0314371257485, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1266, "step": 24069 }, { "epoch": 36.032934131736525, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1223, "step": 24070 }, { "epoch": 36.03443113772455, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1251, "step": 24071 }, { "epoch": 36.035928143712574, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.125, "step": 24072 }, { "epoch": 36.037425149700596, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.128, "step": 24073 }, { "epoch": 36.038922155688624, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1251, "step": 24074 }, { "epoch": 36.040419161676645, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1312, "step": 24075 }, { "epoch": 36.041916167664674, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1319, "step": 24076 }, { "epoch": 36.043413173652695, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.128, "step": 24077 }, { "epoch": 36.044910179640716, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1248, "step": 24078 }, { "epoch": 36.046407185628745, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 24079 }, { "epoch": 36.047904191616766, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1236, "step": 24080 }, { "epoch": 36.04940119760479, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1275, "step": 24081 }, { "epoch": 36.050898203592816, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1271, "step": 24082 }, { "epoch": 36.05239520958084, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1206, "step": 24083 }, { "epoch": 36.053892215568865, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 24084 }, { "epoch": 36.05538922155689, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1248, "step": 24085 }, { "epoch": 36.05688622754491, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1287, "step": 24086 }, { "epoch": 36.058383233532936, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1211, "step": 24087 }, { "epoch": 36.05988023952096, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1315, "step": 24088 }, { "epoch": 36.06137724550898, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1284, "step": 24089 }, { "epoch": 36.06287425149701, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1326, "step": 24090 }, { "epoch": 36.06437125748503, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1193, "step": 24091 }, { "epoch": 36.06586826347306, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1273, "step": 24092 }, { "epoch": 36.06736526946108, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 24093 }, { "epoch": 36.0688622754491, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1242, "step": 24094 }, { "epoch": 36.07035928143713, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1255, "step": 24095 }, { "epoch": 36.07185628742515, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.124, "step": 24096 }, { "epoch": 36.07335329341317, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1249, "step": 24097 }, { "epoch": 36.0748502994012, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1212, "step": 24098 }, { "epoch": 36.07634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1244, "step": 24099 }, { "epoch": 36.07784431137725, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1259, "step": 24100 }, { "epoch": 36.07934131736527, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1359, "step": 24101 }, { "epoch": 36.08083832335329, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1255, "step": 24102 }, { "epoch": 36.08233532934132, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1306, "step": 24103 }, { "epoch": 36.08383233532934, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1221, "step": 24104 }, { "epoch": 36.08532934131737, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 24105 }, { "epoch": 36.08682634730539, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1284, "step": 24106 }, { "epoch": 36.08832335329341, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.128, "step": 24107 }, { "epoch": 36.08982035928144, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1331, "step": 24108 }, { "epoch": 36.09131736526946, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.127, "step": 24109 }, { "epoch": 36.09281437125748, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1309, "step": 24110 }, { "epoch": 36.09431137724551, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.121, "step": 24111 }, { "epoch": 36.09580838323353, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1276, "step": 24112 }, { "epoch": 36.09730538922156, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1258, "step": 24113 }, { "epoch": 36.09880239520958, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.121, "step": 24114 }, { "epoch": 36.1002994011976, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1291, "step": 24115 }, { "epoch": 36.10179640718563, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1364, "step": 24116 }, { "epoch": 36.10329341317365, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.124, "step": 24117 }, { "epoch": 36.104790419161674, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1256, "step": 24118 }, { "epoch": 36.1062874251497, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1225, "step": 24119 }, { "epoch": 36.10778443113772, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1303, "step": 24120 }, { "epoch": 36.10928143712575, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1251, "step": 24121 }, { "epoch": 36.11077844311377, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1266, "step": 24122 }, { "epoch": 36.112275449101794, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1302, "step": 24123 }, { "epoch": 36.11377245508982, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1251, "step": 24124 }, { "epoch": 36.115269461077844, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1224, "step": 24125 }, { "epoch": 36.116766467065865, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1263, "step": 24126 }, { "epoch": 36.118263473053894, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 24127 }, { "epoch": 36.119760479041915, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1256, "step": 24128 }, { "epoch": 36.12125748502994, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1232, "step": 24129 }, { "epoch": 36.122754491017965, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1272, "step": 24130 }, { "epoch": 36.124251497005986, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1205, "step": 24131 }, { "epoch": 36.125748502994014, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1256, "step": 24132 }, { "epoch": 36.127245508982035, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1258, "step": 24133 }, { "epoch": 36.12874251497006, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1269, "step": 24134 }, { "epoch": 36.130239520958085, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1292, "step": 24135 }, { "epoch": 36.131736526946106, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1247, "step": 24136 }, { "epoch": 36.133233532934135, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1206, "step": 24137 }, { "epoch": 36.134730538922156, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1214, "step": 24138 }, { "epoch": 36.13622754491018, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1266, "step": 24139 }, { "epoch": 36.137724550898206, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1286, "step": 24140 }, { "epoch": 36.13922155688623, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1287, "step": 24141 }, { "epoch": 36.14071856287425, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1282, "step": 24142 }, { "epoch": 36.14221556886228, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1292, "step": 24143 }, { "epoch": 36.1437125748503, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1276, "step": 24144 }, { "epoch": 36.145209580838326, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 24145 }, { "epoch": 36.14670658682635, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1204, "step": 24146 }, { "epoch": 36.14820359281437, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1255, "step": 24147 }, { "epoch": 36.1497005988024, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1192, "step": 24148 }, { "epoch": 36.15119760479042, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1263, "step": 24149 }, { "epoch": 36.15269461077844, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1263, "step": 24150 }, { "epoch": 36.15419161676647, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1212, "step": 24151 }, { "epoch": 36.15568862275449, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1259, "step": 24152 }, { "epoch": 36.15718562874252, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1311, "step": 24153 }, { "epoch": 36.15868263473054, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1225, "step": 24154 }, { "epoch": 36.16017964071856, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1289, "step": 24155 }, { "epoch": 36.16167664670659, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1293, "step": 24156 }, { "epoch": 36.16317365269461, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1264, "step": 24157 }, { "epoch": 36.16467065868263, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 24158 }, { "epoch": 36.16616766467066, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1245, "step": 24159 }, { "epoch": 36.16766467065868, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1249, "step": 24160 }, { "epoch": 36.16916167664671, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1348, "step": 24161 }, { "epoch": 36.17065868263473, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1309, "step": 24162 }, { "epoch": 36.17215568862275, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1277, "step": 24163 }, { "epoch": 36.17365269461078, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.118, "step": 24164 }, { "epoch": 36.1751497005988, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1281, "step": 24165 }, { "epoch": 36.17664670658683, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1326, "step": 24166 }, { "epoch": 36.17814371257485, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1271, "step": 24167 }, { "epoch": 36.17964071856287, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 24168 }, { "epoch": 36.1811377245509, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1314, "step": 24169 }, { "epoch": 36.18263473053892, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1204, "step": 24170 }, { "epoch": 36.18413173652694, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1147, "step": 24171 }, { "epoch": 36.18562874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1259, "step": 24172 }, { "epoch": 36.18712574850299, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1227, "step": 24173 }, { "epoch": 36.18862275449102, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1282, "step": 24174 }, { "epoch": 36.19011976047904, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1249, "step": 24175 }, { "epoch": 36.191616766467064, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1303, "step": 24176 }, { "epoch": 36.19311377245509, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1199, "step": 24177 }, { "epoch": 36.19461077844311, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1233, "step": 24178 }, { "epoch": 36.196107784431135, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 24179 }, { "epoch": 36.19760479041916, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1224, "step": 24180 }, { "epoch": 36.199101796407184, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.127, "step": 24181 }, { "epoch": 36.20059880239521, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1257, "step": 24182 }, { "epoch": 36.202095808383234, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1244, "step": 24183 }, { "epoch": 36.203592814371255, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1242, "step": 24184 }, { "epoch": 36.205089820359284, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1271, "step": 24185 }, { "epoch": 36.206586826347305, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1247, "step": 24186 }, { "epoch": 36.208083832335326, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1236, "step": 24187 }, { "epoch": 36.209580838323355, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.131, "step": 24188 }, { "epoch": 36.211077844311376, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1218, "step": 24189 }, { "epoch": 36.212574850299404, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1225, "step": 24190 }, { "epoch": 36.214071856287426, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1212, "step": 24191 }, { "epoch": 36.21556886227545, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1225, "step": 24192 }, { "epoch": 36.217065868263475, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1311, "step": 24193 }, { "epoch": 36.2185628742515, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1253, "step": 24194 }, { "epoch": 36.22005988023952, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1287, "step": 24195 }, { "epoch": 36.221556886227546, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1242, "step": 24196 }, { "epoch": 36.22305389221557, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1224, "step": 24197 }, { "epoch": 36.224550898203596, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1262, "step": 24198 }, { "epoch": 36.22604790419162, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1229, "step": 24199 }, { "epoch": 36.22754491017964, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1262, "step": 24200 }, { "epoch": 36.22904191616767, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.126, "step": 24201 }, { "epoch": 36.23053892215569, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1216, "step": 24202 }, { "epoch": 36.23203592814371, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1243, "step": 24203 }, { "epoch": 36.23353293413174, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 24204 }, { "epoch": 36.23502994011976, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.124, "step": 24205 }, { "epoch": 36.23652694610779, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1281, "step": 24206 }, { "epoch": 36.23802395209581, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1333, "step": 24207 }, { "epoch": 36.23952095808383, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1327, "step": 24208 }, { "epoch": 36.24101796407186, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1245, "step": 24209 }, { "epoch": 36.24251497005988, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1238, "step": 24210 }, { "epoch": 36.2440119760479, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1235, "step": 24211 }, { "epoch": 36.24550898203593, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1195, "step": 24212 }, { "epoch": 36.24700598802395, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1283, "step": 24213 }, { "epoch": 36.24850299401198, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1219, "step": 24214 }, { "epoch": 36.25, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1295, "step": 24215 }, { "epoch": 36.25149700598802, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1294, "step": 24216 }, { "epoch": 36.25299401197605, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1279, "step": 24217 }, { "epoch": 36.25449101796407, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1302, "step": 24218 }, { "epoch": 36.2559880239521, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1234, "step": 24219 }, { "epoch": 36.25748502994012, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1317, "step": 24220 }, { "epoch": 36.25898203592814, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1287, "step": 24221 }, { "epoch": 36.26047904191617, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1227, "step": 24222 }, { "epoch": 36.26197604790419, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1331, "step": 24223 }, { "epoch": 36.26347305389221, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1294, "step": 24224 }, { "epoch": 36.26497005988024, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1259, "step": 24225 }, { "epoch": 36.26646706586826, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1309, "step": 24226 }, { "epoch": 36.26796407185629, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.124, "step": 24227 }, { "epoch": 36.26946107784431, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1248, "step": 24228 }, { "epoch": 36.27095808383233, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1236, "step": 24229 }, { "epoch": 36.27245508982036, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1246, "step": 24230 }, { "epoch": 36.27395209580838, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1256, "step": 24231 }, { "epoch": 36.275449101796404, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.134, "step": 24232 }, { "epoch": 36.27694610778443, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1199, "step": 24233 }, { "epoch": 36.278443113772454, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.121, "step": 24234 }, { "epoch": 36.27994011976048, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 24235 }, { "epoch": 36.2814371257485, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1241, "step": 24236 }, { "epoch": 36.282934131736525, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1261, "step": 24237 }, { "epoch": 36.28443113772455, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1305, "step": 24238 }, { "epoch": 36.285928143712574, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1226, "step": 24239 }, { "epoch": 36.287425149700596, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 24240 }, { "epoch": 36.288922155688624, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 24241 }, { "epoch": 36.290419161676645, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1218, "step": 24242 }, { "epoch": 36.291916167664674, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1249, "step": 24243 }, { "epoch": 36.293413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1333, "step": 24244 }, { "epoch": 36.294910179640716, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1328, "step": 24245 }, { "epoch": 36.296407185628745, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.129, "step": 24246 }, { "epoch": 36.297904191616766, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.126, "step": 24247 }, { "epoch": 36.29940119760479, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1252, "step": 24248 }, { "epoch": 36.300898203592816, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 24249 }, { "epoch": 36.30239520958084, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 24250 }, { "epoch": 36.303892215568865, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1273, "step": 24251 }, { "epoch": 36.30538922155689, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 24252 }, { "epoch": 36.30688622754491, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1252, "step": 24253 }, { "epoch": 36.308383233532936, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.121, "step": 24254 }, { "epoch": 36.30988023952096, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1214, "step": 24255 }, { "epoch": 36.31137724550898, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1262, "step": 24256 }, { "epoch": 36.31287425149701, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.125, "step": 24257 }, { "epoch": 36.31437125748503, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1292, "step": 24258 }, { "epoch": 36.31586826347306, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1253, "step": 24259 }, { "epoch": 36.31736526946108, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1286, "step": 24260 }, { "epoch": 36.3188622754491, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1211, "step": 24261 }, { "epoch": 36.32035928143713, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1324, "step": 24262 }, { "epoch": 36.32185628742515, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1258, "step": 24263 }, { "epoch": 36.32335329341317, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1286, "step": 24264 }, { "epoch": 36.3248502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.123, "step": 24265 }, { "epoch": 36.32634730538922, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1284, "step": 24266 }, { "epoch": 36.32784431137725, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 24267 }, { "epoch": 36.32934131736527, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1248, "step": 24268 }, { "epoch": 36.33083832335329, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1261, "step": 24269 }, { "epoch": 36.33233532934132, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1274, "step": 24270 }, { "epoch": 36.33383233532934, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1255, "step": 24271 }, { "epoch": 36.33532934131736, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1251, "step": 24272 }, { "epoch": 36.33682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.127, "step": 24273 }, { "epoch": 36.33832335329341, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1236, "step": 24274 }, { "epoch": 36.33982035928144, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1361, "step": 24275 }, { "epoch": 36.34131736526946, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1359, "step": 24276 }, { "epoch": 36.34281437125748, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1268, "step": 24277 }, { "epoch": 36.34431137724551, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1234, "step": 24278 }, { "epoch": 36.34580838323353, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1289, "step": 24279 }, { "epoch": 36.34730538922156, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1217, "step": 24280 }, { "epoch": 36.34880239520958, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1207, "step": 24281 }, { "epoch": 36.3502994011976, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1259, "step": 24282 }, { "epoch": 36.35179640718563, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1322, "step": 24283 }, { "epoch": 36.35329341317365, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1299, "step": 24284 }, { "epoch": 36.354790419161674, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 24285 }, { "epoch": 36.3562874251497, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1199, "step": 24286 }, { "epoch": 36.35778443113772, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1257, "step": 24287 }, { "epoch": 36.35928143712575, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1294, "step": 24288 }, { "epoch": 36.36077844311377, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1261, "step": 24289 }, { "epoch": 36.362275449101794, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1221, "step": 24290 }, { "epoch": 36.36377245508982, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1202, "step": 24291 }, { "epoch": 36.365269461077844, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1265, "step": 24292 }, { "epoch": 36.366766467065865, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.125, "step": 24293 }, { "epoch": 36.368263473053894, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1301, "step": 24294 }, { "epoch": 36.369760479041915, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1298, "step": 24295 }, { "epoch": 36.37125748502994, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1289, "step": 24296 }, { "epoch": 36.372754491017965, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.127, "step": 24297 }, { "epoch": 36.374251497005986, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1293, "step": 24298 }, { "epoch": 36.375748502994014, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1275, "step": 24299 }, { "epoch": 36.377245508982035, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1252, "step": 24300 }, { "epoch": 36.37874251497006, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1262, "step": 24301 }, { "epoch": 36.380239520958085, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1198, "step": 24302 }, { "epoch": 36.381736526946106, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.126, "step": 24303 }, { "epoch": 36.383233532934135, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1312, "step": 24304 }, { "epoch": 36.384730538922156, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1266, "step": 24305 }, { "epoch": 36.38622754491018, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1293, "step": 24306 }, { "epoch": 36.387724550898206, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 24307 }, { "epoch": 36.38922155688623, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1202, "step": 24308 }, { "epoch": 36.39071856287425, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1263, "step": 24309 }, { "epoch": 36.39221556886228, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1334, "step": 24310 }, { "epoch": 36.3937125748503, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1186, "step": 24311 }, { "epoch": 36.395209580838326, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1236, "step": 24312 }, { "epoch": 36.39670658682635, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1343, "step": 24313 }, { "epoch": 36.39820359281437, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1242, "step": 24314 }, { "epoch": 36.3997005988024, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.125, "step": 24315 }, { "epoch": 36.40119760479042, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1244, "step": 24316 }, { "epoch": 36.40269461077844, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 24317 }, { "epoch": 36.40419161676647, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1248, "step": 24318 }, { "epoch": 36.40568862275449, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1215, "step": 24319 }, { "epoch": 36.40718562874252, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1285, "step": 24320 }, { "epoch": 36.40868263473054, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1248, "step": 24321 }, { "epoch": 36.41017964071856, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1268, "step": 24322 }, { "epoch": 36.41167664670659, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1261, "step": 24323 }, { "epoch": 36.41317365269461, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1231, "step": 24324 }, { "epoch": 36.41467065868264, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1304, "step": 24325 }, { "epoch": 36.41616766467066, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1276, "step": 24326 }, { "epoch": 36.41766467065868, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1254, "step": 24327 }, { "epoch": 36.41916167664671, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1279, "step": 24328 }, { "epoch": 36.42065868263473, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 24329 }, { "epoch": 36.42215568862275, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1245, "step": 24330 }, { "epoch": 36.42365269461078, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1264, "step": 24331 }, { "epoch": 36.4251497005988, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1285, "step": 24332 }, { "epoch": 36.42664670658683, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1295, "step": 24333 }, { "epoch": 36.42814371257485, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1264, "step": 24334 }, { "epoch": 36.42964071856287, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1329, "step": 24335 }, { "epoch": 36.4311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1244, "step": 24336 }, { "epoch": 36.43263473053892, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.126, "step": 24337 }, { "epoch": 36.43413173652694, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1371, "step": 24338 }, { "epoch": 36.43562874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1309, "step": 24339 }, { "epoch": 36.43712574850299, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1282, "step": 24340 }, { "epoch": 36.43862275449102, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1268, "step": 24341 }, { "epoch": 36.44011976047904, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1263, "step": 24342 }, { "epoch": 36.441616766467064, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1229, "step": 24343 }, { "epoch": 36.44311377245509, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 24344 }, { "epoch": 36.44461077844311, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 24345 }, { "epoch": 36.446107784431135, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1225, "step": 24346 }, { "epoch": 36.44760479041916, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1239, "step": 24347 }, { "epoch": 36.449101796407184, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1238, "step": 24348 }, { "epoch": 36.45059880239521, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1275, "step": 24349 }, { "epoch": 36.452095808383234, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1274, "step": 24350 }, { "epoch": 36.453592814371255, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1243, "step": 24351 }, { "epoch": 36.455089820359284, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1256, "step": 24352 }, { "epoch": 36.456586826347305, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1288, "step": 24353 }, { "epoch": 36.458083832335326, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1238, "step": 24354 }, { "epoch": 36.459580838323355, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1263, "step": 24355 }, { "epoch": 36.461077844311376, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1296, "step": 24356 }, { "epoch": 36.462574850299404, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.125, "step": 24357 }, { "epoch": 36.464071856287426, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1251, "step": 24358 }, { "epoch": 36.46556886227545, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1277, "step": 24359 }, { "epoch": 36.467065868263475, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1293, "step": 24360 }, { "epoch": 36.4685628742515, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1238, "step": 24361 }, { "epoch": 36.47005988023952, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.125, "step": 24362 }, { "epoch": 36.471556886227546, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1228, "step": 24363 }, { "epoch": 36.47305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1271, "step": 24364 }, { "epoch": 36.474550898203596, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 24365 }, { "epoch": 36.47604790419162, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1226, "step": 24366 }, { "epoch": 36.47754491017964, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1252, "step": 24367 }, { "epoch": 36.47904191616767, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.135, "step": 24368 }, { "epoch": 36.48053892215569, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1235, "step": 24369 }, { "epoch": 36.48203592814371, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1287, "step": 24370 }, { "epoch": 36.48353293413174, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1268, "step": 24371 }, { "epoch": 36.48502994011976, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1295, "step": 24372 }, { "epoch": 36.48652694610779, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.131, "step": 24373 }, { "epoch": 36.48802395209581, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1224, "step": 24374 }, { "epoch": 36.48952095808383, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1279, "step": 24375 }, { "epoch": 36.49101796407186, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.125, "step": 24376 }, { "epoch": 36.49251497005988, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1233, "step": 24377 }, { "epoch": 36.4940119760479, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.129, "step": 24378 }, { "epoch": 36.49550898203593, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1287, "step": 24379 }, { "epoch": 36.49700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1212, "step": 24380 }, { "epoch": 36.49850299401198, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1294, "step": 24381 }, { "epoch": 36.5, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1246, "step": 24382 }, { "epoch": 36.50149700598802, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.131, "step": 24383 }, { "epoch": 36.50299401197605, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1277, "step": 24384 }, { "epoch": 36.50449101796407, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.129, "step": 24385 }, { "epoch": 36.5059880239521, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1337, "step": 24386 }, { "epoch": 36.50748502994012, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.131, "step": 24387 }, { "epoch": 36.50898203592814, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.126, "step": 24388 }, { "epoch": 36.51047904191617, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1346, "step": 24389 }, { "epoch": 36.51197604790419, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1262, "step": 24390 }, { "epoch": 36.51347305389221, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1214, "step": 24391 }, { "epoch": 36.51497005988024, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1295, "step": 24392 }, { "epoch": 36.51646706586826, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1263, "step": 24393 }, { "epoch": 36.51796407185629, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 24394 }, { "epoch": 36.51946107784431, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1295, "step": 24395 }, { "epoch": 36.52095808383233, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1263, "step": 24396 }, { "epoch": 36.52245508982036, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1298, "step": 24397 }, { "epoch": 36.52395209580838, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1225, "step": 24398 }, { "epoch": 36.525449101796404, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1209, "step": 24399 }, { "epoch": 36.52694610778443, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1294, "step": 24400 }, { "epoch": 36.528443113772454, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1221, "step": 24401 }, { "epoch": 36.52994011976048, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.129, "step": 24402 }, { "epoch": 36.5314371257485, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1321, "step": 24403 }, { "epoch": 36.532934131736525, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1198, "step": 24404 }, { "epoch": 36.53443113772455, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1326, "step": 24405 }, { "epoch": 36.535928143712574, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1255, "step": 24406 }, { "epoch": 36.537425149700596, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 24407 }, { "epoch": 36.538922155688624, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1272, "step": 24408 }, { "epoch": 36.540419161676645, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1248, "step": 24409 }, { "epoch": 36.541916167664674, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1224, "step": 24410 }, { "epoch": 36.543413173652695, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1263, "step": 24411 }, { "epoch": 36.544910179640716, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1197, "step": 24412 }, { "epoch": 36.546407185628745, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1228, "step": 24413 }, { "epoch": 36.547904191616766, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1211, "step": 24414 }, { "epoch": 36.54940119760479, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1288, "step": 24415 }, { "epoch": 36.550898203592816, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1237, "step": 24416 }, { "epoch": 36.55239520958084, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1345, "step": 24417 }, { "epoch": 36.553892215568865, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.124, "step": 24418 }, { "epoch": 36.55538922155689, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1245, "step": 24419 }, { "epoch": 36.55688622754491, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1299, "step": 24420 }, { "epoch": 36.558383233532936, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1237, "step": 24421 }, { "epoch": 36.55988023952096, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 24422 }, { "epoch": 36.56137724550898, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1279, "step": 24423 }, { "epoch": 36.56287425149701, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1216, "step": 24424 }, { "epoch": 36.56437125748503, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1277, "step": 24425 }, { "epoch": 36.56586826347306, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1265, "step": 24426 }, { "epoch": 36.56736526946108, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1194, "step": 24427 }, { "epoch": 36.5688622754491, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1228, "step": 24428 }, { "epoch": 36.57035928143713, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1247, "step": 24429 }, { "epoch": 36.57185628742515, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1246, "step": 24430 }, { "epoch": 36.57335329341317, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1234, "step": 24431 }, { "epoch": 36.5748502994012, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1232, "step": 24432 }, { "epoch": 36.57634730538922, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.126, "step": 24433 }, { "epoch": 36.57784431137725, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1301, "step": 24434 }, { "epoch": 36.57934131736527, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1249, "step": 24435 }, { "epoch": 36.58083832335329, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1309, "step": 24436 }, { "epoch": 36.58233532934132, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1283, "step": 24437 }, { "epoch": 36.58383233532934, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 24438 }, { "epoch": 36.58532934131736, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1236, "step": 24439 }, { "epoch": 36.58682634730539, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1201, "step": 24440 }, { "epoch": 36.58832335329341, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1325, "step": 24441 }, { "epoch": 36.58982035928144, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1232, "step": 24442 }, { "epoch": 36.59131736526946, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1234, "step": 24443 }, { "epoch": 36.59281437125748, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1302, "step": 24444 }, { "epoch": 36.59431137724551, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1324, "step": 24445 }, { "epoch": 36.59580838323353, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.133, "step": 24446 }, { "epoch": 36.59730538922156, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1178, "step": 24447 }, { "epoch": 36.59880239520958, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1284, "step": 24448 }, { "epoch": 36.6002994011976, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1286, "step": 24449 }, { "epoch": 36.60179640718563, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1232, "step": 24450 }, { "epoch": 36.60329341317365, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1292, "step": 24451 }, { "epoch": 36.604790419161674, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.126, "step": 24452 }, { "epoch": 36.6062874251497, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1297, "step": 24453 }, { "epoch": 36.60778443113772, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1204, "step": 24454 }, { "epoch": 36.60928143712575, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1259, "step": 24455 }, { "epoch": 36.61077844311377, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1299, "step": 24456 }, { "epoch": 36.612275449101794, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1248, "step": 24457 }, { "epoch": 36.61377245508982, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1253, "step": 24458 }, { "epoch": 36.615269461077844, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1282, "step": 24459 }, { "epoch": 36.616766467065865, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1281, "step": 24460 }, { "epoch": 36.618263473053894, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1241, "step": 24461 }, { "epoch": 36.619760479041915, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1181, "step": 24462 }, { "epoch": 36.62125748502994, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1242, "step": 24463 }, { "epoch": 36.622754491017965, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1283, "step": 24464 }, { "epoch": 36.624251497005986, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.128, "step": 24465 }, { "epoch": 36.625748502994014, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1323, "step": 24466 }, { "epoch": 36.627245508982035, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1243, "step": 24467 }, { "epoch": 36.62874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.129, "step": 24468 }, { "epoch": 36.630239520958085, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1338, "step": 24469 }, { "epoch": 36.631736526946106, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1256, "step": 24470 }, { "epoch": 36.633233532934135, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1298, "step": 24471 }, { "epoch": 36.634730538922156, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1284, "step": 24472 }, { "epoch": 36.63622754491018, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1248, "step": 24473 }, { "epoch": 36.637724550898206, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1197, "step": 24474 }, { "epoch": 36.63922155688623, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1152, "step": 24475 }, { "epoch": 36.64071856287425, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1269, "step": 24476 }, { "epoch": 36.64221556886228, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1293, "step": 24477 }, { "epoch": 36.6437125748503, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1265, "step": 24478 }, { "epoch": 36.645209580838326, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.126, "step": 24479 }, { "epoch": 36.64670658682635, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 24480 }, { "epoch": 36.64820359281437, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1319, "step": 24481 }, { "epoch": 36.6497005988024, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 24482 }, { "epoch": 36.65119760479042, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1198, "step": 24483 }, { "epoch": 36.65269461077844, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1129, "step": 24484 }, { "epoch": 36.65419161676647, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1223, "step": 24485 }, { "epoch": 36.65568862275449, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1238, "step": 24486 }, { "epoch": 36.65718562874252, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 24487 }, { "epoch": 36.65868263473054, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1282, "step": 24488 }, { "epoch": 36.66017964071856, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1198, "step": 24489 }, { "epoch": 36.66167664670659, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1246, "step": 24490 }, { "epoch": 36.66317365269461, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1275, "step": 24491 }, { "epoch": 36.66467065868264, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1207, "step": 24492 }, { "epoch": 36.66616766467066, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 24493 }, { "epoch": 36.66766467065868, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1259, "step": 24494 }, { "epoch": 36.66916167664671, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1239, "step": 24495 }, { "epoch": 36.67065868263473, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.123, "step": 24496 }, { "epoch": 36.67215568862275, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1298, "step": 24497 }, { "epoch": 36.67365269461078, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1273, "step": 24498 }, { "epoch": 36.6751497005988, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1264, "step": 24499 }, { "epoch": 36.67664670658683, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1263, "step": 24500 }, { "epoch": 36.67814371257485, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1221, "step": 24501 }, { "epoch": 36.67964071856287, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1224, "step": 24502 }, { "epoch": 36.6811377245509, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1283, "step": 24503 }, { "epoch": 36.68263473053892, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1315, "step": 24504 }, { "epoch": 36.68413173652694, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.119, "step": 24505 }, { "epoch": 36.68562874251497, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1225, "step": 24506 }, { "epoch": 36.68712574850299, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1217, "step": 24507 }, { "epoch": 36.68862275449102, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1263, "step": 24508 }, { "epoch": 36.69011976047904, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1343, "step": 24509 }, { "epoch": 36.691616766467064, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1305, "step": 24510 }, { "epoch": 36.69311377245509, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1216, "step": 24511 }, { "epoch": 36.69461077844311, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1257, "step": 24512 }, { "epoch": 36.696107784431135, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1264, "step": 24513 }, { "epoch": 36.69760479041916, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1211, "step": 24514 }, { "epoch": 36.699101796407184, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1247, "step": 24515 }, { "epoch": 36.70059880239521, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 24516 }, { "epoch": 36.702095808383234, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1242, "step": 24517 }, { "epoch": 36.703592814371255, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1209, "step": 24518 }, { "epoch": 36.705089820359284, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.125, "step": 24519 }, { "epoch": 36.706586826347305, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1238, "step": 24520 }, { "epoch": 36.708083832335326, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 24521 }, { "epoch": 36.709580838323355, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1196, "step": 24522 }, { "epoch": 36.711077844311376, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1232, "step": 24523 }, { "epoch": 36.712574850299404, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1254, "step": 24524 }, { "epoch": 36.714071856287426, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1231, "step": 24525 }, { "epoch": 36.71556886227545, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.127, "step": 24526 }, { "epoch": 36.717065868263475, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1175, "step": 24527 }, { "epoch": 36.7185628742515, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1314, "step": 24528 }, { "epoch": 36.72005988023952, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1287, "step": 24529 }, { "epoch": 36.721556886227546, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1271, "step": 24530 }, { "epoch": 36.72305389221557, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1328, "step": 24531 }, { "epoch": 36.724550898203596, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1253, "step": 24532 }, { "epoch": 36.72604790419162, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 24533 }, { "epoch": 36.72754491017964, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1239, "step": 24534 }, { "epoch": 36.72904191616767, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1211, "step": 24535 }, { "epoch": 36.73053892215569, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1243, "step": 24536 }, { "epoch": 36.73203592814371, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1358, "step": 24537 }, { "epoch": 36.73353293413174, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 24538 }, { "epoch": 36.73502994011976, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1331, "step": 24539 }, { "epoch": 36.73652694610779, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1268, "step": 24540 }, { "epoch": 36.73802395209581, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1266, "step": 24541 }, { "epoch": 36.73952095808383, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1228, "step": 24542 }, { "epoch": 36.74101796407186, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1286, "step": 24543 }, { "epoch": 36.74251497005988, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1348, "step": 24544 }, { "epoch": 36.7440119760479, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1255, "step": 24545 }, { "epoch": 36.74550898203593, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1227, "step": 24546 }, { "epoch": 36.74700598802395, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1264, "step": 24547 }, { "epoch": 36.74850299401198, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.127, "step": 24548 }, { "epoch": 36.75, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1245, "step": 24549 }, { "epoch": 36.75149700598802, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1268, "step": 24550 }, { "epoch": 36.75299401197605, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1246, "step": 24551 }, { "epoch": 36.75449101796407, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1322, "step": 24552 }, { "epoch": 36.7559880239521, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1248, "step": 24553 }, { "epoch": 36.75748502994012, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 24554 }, { "epoch": 36.75898203592814, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1211, "step": 24555 }, { "epoch": 36.76047904191617, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1246, "step": 24556 }, { "epoch": 36.76197604790419, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1317, "step": 24557 }, { "epoch": 36.76347305389221, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1282, "step": 24558 }, { "epoch": 36.76497005988024, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1264, "step": 24559 }, { "epoch": 36.76646706586826, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.128, "step": 24560 }, { "epoch": 36.76796407185629, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1233, "step": 24561 }, { "epoch": 36.76946107784431, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1227, "step": 24562 }, { "epoch": 36.77095808383233, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1268, "step": 24563 }, { "epoch": 36.77245508982036, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1307, "step": 24564 }, { "epoch": 36.77395209580838, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1275, "step": 24565 }, { "epoch": 36.775449101796404, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1265, "step": 24566 }, { "epoch": 36.77694610778443, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1264, "step": 24567 }, { "epoch": 36.778443113772454, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1303, "step": 24568 }, { "epoch": 36.77994011976048, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1228, "step": 24569 }, { "epoch": 36.7814371257485, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1241, "step": 24570 }, { "epoch": 36.782934131736525, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1242, "step": 24571 }, { "epoch": 36.78443113772455, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.122, "step": 24572 }, { "epoch": 36.785928143712574, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1252, "step": 24573 }, { "epoch": 36.787425149700596, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1248, "step": 24574 }, { "epoch": 36.788922155688624, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1244, "step": 24575 }, { "epoch": 36.790419161676645, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1277, "step": 24576 }, { "epoch": 36.791916167664674, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1352, "step": 24577 }, { "epoch": 36.793413173652695, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 24578 }, { "epoch": 36.794910179640716, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1293, "step": 24579 }, { "epoch": 36.796407185628745, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1273, "step": 24580 }, { "epoch": 36.797904191616766, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.127, "step": 24581 }, { "epoch": 36.79940119760479, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1198, "step": 24582 }, { "epoch": 36.800898203592816, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1251, "step": 24583 }, { "epoch": 36.80239520958084, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1226, "step": 24584 }, { "epoch": 36.803892215568865, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1291, "step": 24585 }, { "epoch": 36.80538922155689, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1253, "step": 24586 }, { "epoch": 36.80688622754491, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1257, "step": 24587 }, { "epoch": 36.808383233532936, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1168, "step": 24588 }, { "epoch": 36.80988023952096, "grad_norm": 0.049560546875, "learning_rate": 0.0008, "loss": 1.1174, "step": 24589 }, { "epoch": 36.81137724550898, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1259, "step": 24590 }, { "epoch": 36.81287425149701, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1285, "step": 24591 }, { "epoch": 36.81437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1259, "step": 24592 }, { "epoch": 36.81586826347306, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1277, "step": 24593 }, { "epoch": 36.81736526946108, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 24594 }, { "epoch": 36.8188622754491, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1322, "step": 24595 }, { "epoch": 36.82035928143713, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1362, "step": 24596 }, { "epoch": 36.82185628742515, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1247, "step": 24597 }, { "epoch": 36.82335329341317, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1263, "step": 24598 }, { "epoch": 36.8248502994012, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.137, "step": 24599 }, { "epoch": 36.82634730538922, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1236, "step": 24600 }, { "epoch": 36.82784431137725, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1256, "step": 24601 }, { "epoch": 36.82934131736527, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1211, "step": 24602 }, { "epoch": 36.83083832335329, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1247, "step": 24603 }, { "epoch": 36.83233532934132, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1231, "step": 24604 }, { "epoch": 36.83383233532934, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1258, "step": 24605 }, { "epoch": 36.83532934131736, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 24606 }, { "epoch": 36.83682634730539, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1285, "step": 24607 }, { "epoch": 36.83832335329341, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1262, "step": 24608 }, { "epoch": 36.83982035928144, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.121, "step": 24609 }, { "epoch": 36.84131736526946, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1235, "step": 24610 }, { "epoch": 36.84281437125748, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1271, "step": 24611 }, { "epoch": 36.84431137724551, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1254, "step": 24612 }, { "epoch": 36.84580838323353, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1257, "step": 24613 }, { "epoch": 36.84730538922156, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1222, "step": 24614 }, { "epoch": 36.84880239520958, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1253, "step": 24615 }, { "epoch": 36.8502994011976, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1169, "step": 24616 }, { "epoch": 36.85179640718563, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1215, "step": 24617 }, { "epoch": 36.85329341317365, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1279, "step": 24618 }, { "epoch": 36.854790419161674, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1247, "step": 24619 }, { "epoch": 36.8562874251497, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1268, "step": 24620 }, { "epoch": 36.85778443113772, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1225, "step": 24621 }, { "epoch": 36.85928143712575, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1239, "step": 24622 }, { "epoch": 36.86077844311377, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1261, "step": 24623 }, { "epoch": 36.862275449101794, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1234, "step": 24624 }, { "epoch": 36.86377245508982, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1238, "step": 24625 }, { "epoch": 36.865269461077844, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1282, "step": 24626 }, { "epoch": 36.866766467065865, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1205, "step": 24627 }, { "epoch": 36.868263473053894, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 24628 }, { "epoch": 36.869760479041915, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1265, "step": 24629 }, { "epoch": 36.87125748502994, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1293, "step": 24630 }, { "epoch": 36.872754491017965, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.126, "step": 24631 }, { "epoch": 36.874251497005986, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.123, "step": 24632 }, { "epoch": 36.875748502994014, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.122, "step": 24633 }, { "epoch": 36.877245508982035, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1198, "step": 24634 }, { "epoch": 36.87874251497006, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1261, "step": 24635 }, { "epoch": 36.880239520958085, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1288, "step": 24636 }, { "epoch": 36.881736526946106, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1287, "step": 24637 }, { "epoch": 36.883233532934135, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1271, "step": 24638 }, { "epoch": 36.884730538922156, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1193, "step": 24639 }, { "epoch": 36.88622754491018, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1253, "step": 24640 }, { "epoch": 36.887724550898206, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1259, "step": 24641 }, { "epoch": 36.88922155688623, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1265, "step": 24642 }, { "epoch": 36.89071856287425, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1283, "step": 24643 }, { "epoch": 36.89221556886228, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1299, "step": 24644 }, { "epoch": 36.8937125748503, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 24645 }, { "epoch": 36.895209580838326, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1286, "step": 24646 }, { "epoch": 36.89670658682635, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 24647 }, { "epoch": 36.89820359281437, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1228, "step": 24648 }, { "epoch": 36.8997005988024, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.121, "step": 24649 }, { "epoch": 36.90119760479042, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1331, "step": 24650 }, { "epoch": 36.90269461077844, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1308, "step": 24651 }, { "epoch": 36.90419161676647, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 24652 }, { "epoch": 36.90568862275449, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1256, "step": 24653 }, { "epoch": 36.90718562874252, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.124, "step": 24654 }, { "epoch": 36.90868263473054, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1228, "step": 24655 }, { "epoch": 36.91017964071856, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.123, "step": 24656 }, { "epoch": 36.91167664670659, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1297, "step": 24657 }, { "epoch": 36.91317365269461, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1217, "step": 24658 }, { "epoch": 36.91467065868264, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.122, "step": 24659 }, { "epoch": 36.91616766467066, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.124, "step": 24660 }, { "epoch": 36.91766467065868, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 24661 }, { "epoch": 36.91916167664671, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1325, "step": 24662 }, { "epoch": 36.92065868263473, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1246, "step": 24663 }, { "epoch": 36.92215568862275, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1232, "step": 24664 }, { "epoch": 36.92365269461078, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1254, "step": 24665 }, { "epoch": 36.9251497005988, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1221, "step": 24666 }, { "epoch": 36.92664670658683, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1222, "step": 24667 }, { "epoch": 36.92814371257485, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1182, "step": 24668 }, { "epoch": 36.92964071856287, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.13, "step": 24669 }, { "epoch": 36.9311377245509, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.125, "step": 24670 }, { "epoch": 36.93263473053892, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 24671 }, { "epoch": 36.93413173652694, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1244, "step": 24672 }, { "epoch": 36.93562874251497, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1262, "step": 24673 }, { "epoch": 36.93712574850299, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1261, "step": 24674 }, { "epoch": 36.93862275449102, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1268, "step": 24675 }, { "epoch": 36.94011976047904, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.121, "step": 24676 }, { "epoch": 36.941616766467064, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.128, "step": 24677 }, { "epoch": 36.94311377245509, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1174, "step": 24678 }, { "epoch": 36.94461077844311, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1253, "step": 24679 }, { "epoch": 36.946107784431135, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1273, "step": 24680 }, { "epoch": 36.94760479041916, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1254, "step": 24681 }, { "epoch": 36.949101796407184, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1222, "step": 24682 }, { "epoch": 36.95059880239521, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1292, "step": 24683 }, { "epoch": 36.952095808383234, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1313, "step": 24684 }, { "epoch": 36.953592814371255, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 24685 }, { "epoch": 36.955089820359284, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1237, "step": 24686 }, { "epoch": 36.956586826347305, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1298, "step": 24687 }, { "epoch": 36.958083832335326, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.129, "step": 24688 }, { "epoch": 36.959580838323355, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1244, "step": 24689 }, { "epoch": 36.961077844311376, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1266, "step": 24690 }, { "epoch": 36.962574850299404, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1276, "step": 24691 }, { "epoch": 36.964071856287426, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1211, "step": 24692 }, { "epoch": 36.96556886227545, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1199, "step": 24693 }, { "epoch": 36.967065868263475, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1191, "step": 24694 }, { "epoch": 36.9685628742515, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1247, "step": 24695 }, { "epoch": 36.97005988023952, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1322, "step": 24696 }, { "epoch": 36.971556886227546, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1319, "step": 24697 }, { "epoch": 36.97305389221557, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.128, "step": 24698 }, { "epoch": 36.974550898203596, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1257, "step": 24699 }, { "epoch": 36.97604790419162, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1218, "step": 24700 }, { "epoch": 36.97754491017964, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1265, "step": 24701 }, { "epoch": 36.97904191616767, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1252, "step": 24702 }, { "epoch": 36.98053892215569, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1271, "step": 24703 }, { "epoch": 36.98203592814371, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1256, "step": 24704 }, { "epoch": 36.98353293413174, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1272, "step": 24705 }, { "epoch": 36.98502994011976, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.128, "step": 24706 }, { "epoch": 36.98652694610779, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1241, "step": 24707 }, { "epoch": 36.98802395209581, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1255, "step": 24708 }, { "epoch": 36.98952095808383, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1259, "step": 24709 }, { "epoch": 36.99101796407186, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 24710 }, { "epoch": 36.99251497005988, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1336, "step": 24711 }, { "epoch": 36.9940119760479, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1325, "step": 24712 }, { "epoch": 36.99550898203593, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1237, "step": 24713 }, { "epoch": 36.99700598802395, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1285, "step": 24714 }, { "epoch": 36.99850299401198, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1244, "step": 24715 }, { "epoch": 37.0, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1192, "step": 24716 }, { "epoch": 37.00149700598802, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1276, "step": 24717 }, { "epoch": 37.00299401197605, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1228, "step": 24718 }, { "epoch": 37.00449101796407, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1238, "step": 24719 }, { "epoch": 37.0059880239521, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 24720 }, { "epoch": 37.00748502994012, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1187, "step": 24721 }, { "epoch": 37.00898203592814, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1213, "step": 24722 }, { "epoch": 37.01047904191617, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 24723 }, { "epoch": 37.01197604790419, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.124, "step": 24724 }, { "epoch": 37.01347305389221, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1259, "step": 24725 }, { "epoch": 37.01497005988024, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1205, "step": 24726 }, { "epoch": 37.01646706586826, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1215, "step": 24727 }, { "epoch": 37.01796407185629, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1284, "step": 24728 }, { "epoch": 37.01946107784431, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1208, "step": 24729 }, { "epoch": 37.02095808383233, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.12, "step": 24730 }, { "epoch": 37.02245508982036, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1267, "step": 24731 }, { "epoch": 37.02395209580838, "grad_norm": 0.050048828125, "learning_rate": 0.0008, "loss": 1.1282, "step": 24732 }, { "epoch": 37.025449101796404, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1247, "step": 24733 }, { "epoch": 37.02694610778443, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.124, "step": 24734 }, { "epoch": 37.028443113772454, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1278, "step": 24735 }, { "epoch": 37.02994011976048, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1243, "step": 24736 }, { "epoch": 37.0314371257485, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.126, "step": 24737 }, { "epoch": 37.032934131736525, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1226, "step": 24738 }, { "epoch": 37.03443113772455, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1266, "step": 24739 }, { "epoch": 37.035928143712574, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1269, "step": 24740 }, { "epoch": 37.037425149700596, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1202, "step": 24741 }, { "epoch": 37.038922155688624, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1264, "step": 24742 }, { "epoch": 37.040419161676645, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1292, "step": 24743 }, { "epoch": 37.041916167664674, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1267, "step": 24744 }, { "epoch": 37.043413173652695, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1253, "step": 24745 }, { "epoch": 37.044910179640716, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1219, "step": 24746 }, { "epoch": 37.046407185628745, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.123, "step": 24747 }, { "epoch": 37.047904191616766, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1178, "step": 24748 }, { "epoch": 37.04940119760479, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1293, "step": 24749 }, { "epoch": 37.050898203592816, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.127, "step": 24750 }, { "epoch": 37.05239520958084, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1262, "step": 24751 }, { "epoch": 37.053892215568865, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.122, "step": 24752 }, { "epoch": 37.05538922155689, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1232, "step": 24753 }, { "epoch": 37.05688622754491, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.123, "step": 24754 }, { "epoch": 37.058383233532936, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1261, "step": 24755 }, { "epoch": 37.05988023952096, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1248, "step": 24756 }, { "epoch": 37.06137724550898, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1339, "step": 24757 }, { "epoch": 37.06287425149701, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.125, "step": 24758 }, { "epoch": 37.06437125748503, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1258, "step": 24759 }, { "epoch": 37.06586826347306, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1276, "step": 24760 }, { "epoch": 37.06736526946108, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1276, "step": 24761 }, { "epoch": 37.0688622754491, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1255, "step": 24762 }, { "epoch": 37.07035928143713, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 24763 }, { "epoch": 37.07185628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1297, "step": 24764 }, { "epoch": 37.07335329341317, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 24765 }, { "epoch": 37.0748502994012, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1288, "step": 24766 }, { "epoch": 37.07634730538922, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.123, "step": 24767 }, { "epoch": 37.07784431137725, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1267, "step": 24768 }, { "epoch": 37.07934131736527, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1291, "step": 24769 }, { "epoch": 37.08083832335329, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.122, "step": 24770 }, { "epoch": 37.08233532934132, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.126, "step": 24771 }, { "epoch": 37.08383233532934, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1227, "step": 24772 }, { "epoch": 37.08532934131737, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1232, "step": 24773 }, { "epoch": 37.08682634730539, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1226, "step": 24774 }, { "epoch": 37.08832335329341, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1268, "step": 24775 }, { "epoch": 37.08982035928144, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1296, "step": 24776 }, { "epoch": 37.09131736526946, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1236, "step": 24777 }, { "epoch": 37.09281437125748, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1301, "step": 24778 }, { "epoch": 37.09431137724551, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1267, "step": 24779 }, { "epoch": 37.09580838323353, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1237, "step": 24780 }, { "epoch": 37.09730538922156, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1234, "step": 24781 }, { "epoch": 37.09880239520958, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1322, "step": 24782 }, { "epoch": 37.1002994011976, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1222, "step": 24783 }, { "epoch": 37.10179640718563, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.121, "step": 24784 }, { "epoch": 37.10329341317365, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1317, "step": 24785 }, { "epoch": 37.104790419161674, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.126, "step": 24786 }, { "epoch": 37.1062874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1252, "step": 24787 }, { "epoch": 37.10778443113772, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1259, "step": 24788 }, { "epoch": 37.10928143712575, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1268, "step": 24789 }, { "epoch": 37.11077844311377, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1286, "step": 24790 }, { "epoch": 37.112275449101794, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1276, "step": 24791 }, { "epoch": 37.11377245508982, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1273, "step": 24792 }, { "epoch": 37.115269461077844, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1235, "step": 24793 }, { "epoch": 37.116766467065865, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1247, "step": 24794 }, { "epoch": 37.118263473053894, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1258, "step": 24795 }, { "epoch": 37.119760479041915, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1257, "step": 24796 }, { "epoch": 37.12125748502994, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1267, "step": 24797 }, { "epoch": 37.122754491017965, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1304, "step": 24798 }, { "epoch": 37.124251497005986, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.129, "step": 24799 }, { "epoch": 37.125748502994014, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1211, "step": 24800 }, { "epoch": 37.127245508982035, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1309, "step": 24801 }, { "epoch": 37.12874251497006, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1237, "step": 24802 }, { "epoch": 37.130239520958085, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1301, "step": 24803 }, { "epoch": 37.131736526946106, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1246, "step": 24804 }, { "epoch": 37.133233532934135, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1212, "step": 24805 }, { "epoch": 37.134730538922156, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1223, "step": 24806 }, { "epoch": 37.13622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1251, "step": 24807 }, { "epoch": 37.137724550898206, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1313, "step": 24808 }, { "epoch": 37.13922155688623, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1236, "step": 24809 }, { "epoch": 37.14071856287425, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1316, "step": 24810 }, { "epoch": 37.14221556886228, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1273, "step": 24811 }, { "epoch": 37.1437125748503, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1278, "step": 24812 }, { "epoch": 37.145209580838326, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1228, "step": 24813 }, { "epoch": 37.14670658682635, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1245, "step": 24814 }, { "epoch": 37.14820359281437, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1157, "step": 24815 }, { "epoch": 37.1497005988024, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1264, "step": 24816 }, { "epoch": 37.15119760479042, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1197, "step": 24817 }, { "epoch": 37.15269461077844, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1299, "step": 24818 }, { "epoch": 37.15419161676647, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1317, "step": 24819 }, { "epoch": 37.15568862275449, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1284, "step": 24820 }, { "epoch": 37.15718562874252, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1287, "step": 24821 }, { "epoch": 37.15868263473054, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1277, "step": 24822 }, { "epoch": 37.16017964071856, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.128, "step": 24823 }, { "epoch": 37.16167664670659, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 24824 }, { "epoch": 37.16317365269461, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1244, "step": 24825 }, { "epoch": 37.16467065868263, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.125, "step": 24826 }, { "epoch": 37.16616766467066, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.132, "step": 24827 }, { "epoch": 37.16766467065868, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 24828 }, { "epoch": 37.16916167664671, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1255, "step": 24829 }, { "epoch": 37.17065868263473, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1308, "step": 24830 }, { "epoch": 37.17215568862275, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1309, "step": 24831 }, { "epoch": 37.17365269461078, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1252, "step": 24832 }, { "epoch": 37.1751497005988, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1237, "step": 24833 }, { "epoch": 37.17664670658683, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.124, "step": 24834 }, { "epoch": 37.17814371257485, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1294, "step": 24835 }, { "epoch": 37.17964071856287, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1258, "step": 24836 }, { "epoch": 37.1811377245509, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1234, "step": 24837 }, { "epoch": 37.18263473053892, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1261, "step": 24838 }, { "epoch": 37.18413173652694, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.124, "step": 24839 }, { "epoch": 37.18562874251497, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1182, "step": 24840 }, { "epoch": 37.18712574850299, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1241, "step": 24841 }, { "epoch": 37.18862275449102, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 24842 }, { "epoch": 37.19011976047904, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.125, "step": 24843 }, { "epoch": 37.191616766467064, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1239, "step": 24844 }, { "epoch": 37.19311377245509, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1318, "step": 24845 }, { "epoch": 37.19461077844311, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1222, "step": 24846 }, { "epoch": 37.196107784431135, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1207, "step": 24847 }, { "epoch": 37.19760479041916, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1262, "step": 24848 }, { "epoch": 37.199101796407184, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1257, "step": 24849 }, { "epoch": 37.20059880239521, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1286, "step": 24850 }, { "epoch": 37.202095808383234, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1269, "step": 24851 }, { "epoch": 37.203592814371255, "grad_norm": 0.05029296875, "learning_rate": 0.0008, "loss": 1.1232, "step": 24852 }, { "epoch": 37.205089820359284, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.123, "step": 24853 }, { "epoch": 37.206586826347305, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1192, "step": 24854 }, { "epoch": 37.208083832335326, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1222, "step": 24855 }, { "epoch": 37.209580838323355, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.124, "step": 24856 }, { "epoch": 37.211077844311376, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1313, "step": 24857 }, { "epoch": 37.212574850299404, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 24858 }, { "epoch": 37.214071856287426, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1227, "step": 24859 }, { "epoch": 37.21556886227545, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1232, "step": 24860 }, { "epoch": 37.217065868263475, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.126, "step": 24861 }, { "epoch": 37.2185628742515, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1235, "step": 24862 }, { "epoch": 37.22005988023952, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1229, "step": 24863 }, { "epoch": 37.221556886227546, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1236, "step": 24864 }, { "epoch": 37.22305389221557, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1308, "step": 24865 }, { "epoch": 37.224550898203596, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 24866 }, { "epoch": 37.22604790419162, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1238, "step": 24867 }, { "epoch": 37.22754491017964, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 24868 }, { "epoch": 37.22904191616767, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1254, "step": 24869 }, { "epoch": 37.23053892215569, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1213, "step": 24870 }, { "epoch": 37.23203592814371, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1281, "step": 24871 }, { "epoch": 37.23353293413174, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1147, "step": 24872 }, { "epoch": 37.23502994011976, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1225, "step": 24873 }, { "epoch": 37.23652694610779, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.126, "step": 24874 }, { "epoch": 37.23802395209581, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1261, "step": 24875 }, { "epoch": 37.23952095808383, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1168, "step": 24876 }, { "epoch": 37.24101796407186, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1259, "step": 24877 }, { "epoch": 37.24251497005988, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1247, "step": 24878 }, { "epoch": 37.2440119760479, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 24879 }, { "epoch": 37.24550898203593, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 24880 }, { "epoch": 37.24700598802395, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1247, "step": 24881 }, { "epoch": 37.24850299401198, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1243, "step": 24882 }, { "epoch": 37.25, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1242, "step": 24883 }, { "epoch": 37.25149700598802, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1237, "step": 24884 }, { "epoch": 37.25299401197605, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1299, "step": 24885 }, { "epoch": 37.25449101796407, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1338, "step": 24886 }, { "epoch": 37.2559880239521, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1231, "step": 24887 }, { "epoch": 37.25748502994012, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1273, "step": 24888 }, { "epoch": 37.25898203592814, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1249, "step": 24889 }, { "epoch": 37.26047904191617, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1268, "step": 24890 }, { "epoch": 37.26197604790419, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1248, "step": 24891 }, { "epoch": 37.26347305389221, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1206, "step": 24892 }, { "epoch": 37.26497005988024, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1189, "step": 24893 }, { "epoch": 37.26646706586826, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1235, "step": 24894 }, { "epoch": 37.26796407185629, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1263, "step": 24895 }, { "epoch": 37.26946107784431, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1208, "step": 24896 }, { "epoch": 37.27095808383233, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1247, "step": 24897 }, { "epoch": 37.27245508982036, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1258, "step": 24898 }, { "epoch": 37.27395209580838, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1261, "step": 24899 }, { "epoch": 37.275449101796404, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1276, "step": 24900 }, { "epoch": 37.27694610778443, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1278, "step": 24901 }, { "epoch": 37.278443113772454, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.127, "step": 24902 }, { "epoch": 37.27994011976048, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.126, "step": 24903 }, { "epoch": 37.2814371257485, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.127, "step": 24904 }, { "epoch": 37.282934131736525, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1185, "step": 24905 }, { "epoch": 37.28443113772455, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1186, "step": 24906 }, { "epoch": 37.285928143712574, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1265, "step": 24907 }, { "epoch": 37.287425149700596, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1285, "step": 24908 }, { "epoch": 37.288922155688624, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1231, "step": 24909 }, { "epoch": 37.290419161676645, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1299, "step": 24910 }, { "epoch": 37.291916167664674, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1294, "step": 24911 }, { "epoch": 37.293413173652695, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1238, "step": 24912 }, { "epoch": 37.294910179640716, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1278, "step": 24913 }, { "epoch": 37.296407185628745, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 24914 }, { "epoch": 37.297904191616766, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1252, "step": 24915 }, { "epoch": 37.29940119760479, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.128, "step": 24916 }, { "epoch": 37.300898203592816, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1329, "step": 24917 }, { "epoch": 37.30239520958084, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1306, "step": 24918 }, { "epoch": 37.303892215568865, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1275, "step": 24919 }, { "epoch": 37.30538922155689, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1225, "step": 24920 }, { "epoch": 37.30688622754491, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1256, "step": 24921 }, { "epoch": 37.308383233532936, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1316, "step": 24922 }, { "epoch": 37.30988023952096, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1273, "step": 24923 }, { "epoch": 37.31137724550898, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1271, "step": 24924 }, { "epoch": 37.31287425149701, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1249, "step": 24925 }, { "epoch": 37.31437125748503, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1235, "step": 24926 }, { "epoch": 37.31586826347306, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1259, "step": 24927 }, { "epoch": 37.31736526946108, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1278, "step": 24928 }, { "epoch": 37.3188622754491, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 24929 }, { "epoch": 37.32035928143713, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1209, "step": 24930 }, { "epoch": 37.32185628742515, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1258, "step": 24931 }, { "epoch": 37.32335329341317, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1221, "step": 24932 }, { "epoch": 37.3248502994012, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1266, "step": 24933 }, { "epoch": 37.32634730538922, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1257, "step": 24934 }, { "epoch": 37.32784431137725, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1245, "step": 24935 }, { "epoch": 37.32934131736527, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1288, "step": 24936 }, { "epoch": 37.33083832335329, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1243, "step": 24937 }, { "epoch": 37.33233532934132, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1322, "step": 24938 }, { "epoch": 37.33383233532934, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1279, "step": 24939 }, { "epoch": 37.33532934131736, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1221, "step": 24940 }, { "epoch": 37.33682634730539, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1208, "step": 24941 }, { "epoch": 37.33832335329341, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1249, "step": 24942 }, { "epoch": 37.33982035928144, "grad_norm": 0.048095703125, "learning_rate": 0.0008, "loss": 1.1297, "step": 24943 }, { "epoch": 37.34131736526946, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1253, "step": 24944 }, { "epoch": 37.34281437125748, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1283, "step": 24945 }, { "epoch": 37.34431137724551, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1282, "step": 24946 }, { "epoch": 37.34580838323353, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1274, "step": 24947 }, { "epoch": 37.34730538922156, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1264, "step": 24948 }, { "epoch": 37.34880239520958, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1263, "step": 24949 }, { "epoch": 37.3502994011976, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1282, "step": 24950 }, { "epoch": 37.35179640718563, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1201, "step": 24951 }, { "epoch": 37.35329341317365, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1251, "step": 24952 }, { "epoch": 37.354790419161674, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.124, "step": 24953 }, { "epoch": 37.3562874251497, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1273, "step": 24954 }, { "epoch": 37.35778443113772, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 24955 }, { "epoch": 37.35928143712575, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.121, "step": 24956 }, { "epoch": 37.36077844311377, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1259, "step": 24957 }, { "epoch": 37.362275449101794, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1254, "step": 24958 }, { "epoch": 37.36377245508982, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.125, "step": 24959 }, { "epoch": 37.365269461077844, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1218, "step": 24960 }, { "epoch": 37.366766467065865, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1315, "step": 24961 }, { "epoch": 37.368263473053894, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1294, "step": 24962 }, { "epoch": 37.369760479041915, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.127, "step": 24963 }, { "epoch": 37.37125748502994, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1224, "step": 24964 }, { "epoch": 37.372754491017965, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1266, "step": 24965 }, { "epoch": 37.374251497005986, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.128, "step": 24966 }, { "epoch": 37.375748502994014, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.127, "step": 24967 }, { "epoch": 37.377245508982035, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1245, "step": 24968 }, { "epoch": 37.37874251497006, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1194, "step": 24969 }, { "epoch": 37.380239520958085, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1241, "step": 24970 }, { "epoch": 37.381736526946106, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1335, "step": 24971 }, { "epoch": 37.383233532934135, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1255, "step": 24972 }, { "epoch": 37.384730538922156, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1287, "step": 24973 }, { "epoch": 37.38622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1147, "step": 24974 }, { "epoch": 37.387724550898206, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1257, "step": 24975 }, { "epoch": 37.38922155688623, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1245, "step": 24976 }, { "epoch": 37.39071856287425, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 24977 }, { "epoch": 37.39221556886228, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1239, "step": 24978 }, { "epoch": 37.3937125748503, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1215, "step": 24979 }, { "epoch": 37.395209580838326, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1201, "step": 24980 }, { "epoch": 37.39670658682635, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1241, "step": 24981 }, { "epoch": 37.39820359281437, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1206, "step": 24982 }, { "epoch": 37.3997005988024, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1232, "step": 24983 }, { "epoch": 37.40119760479042, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1258, "step": 24984 }, { "epoch": 37.40269461077844, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1271, "step": 24985 }, { "epoch": 37.40419161676647, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1284, "step": 24986 }, { "epoch": 37.40568862275449, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1194, "step": 24987 }, { "epoch": 37.40718562874252, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1269, "step": 24988 }, { "epoch": 37.40868263473054, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1233, "step": 24989 }, { "epoch": 37.41017964071856, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1309, "step": 24990 }, { "epoch": 37.41167664670659, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1297, "step": 24991 }, { "epoch": 37.41317365269461, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1256, "step": 24992 }, { "epoch": 37.41467065868264, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.122, "step": 24993 }, { "epoch": 37.41616766467066, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1277, "step": 24994 }, { "epoch": 37.41766467065868, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 24995 }, { "epoch": 37.41916167664671, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1272, "step": 24996 }, { "epoch": 37.42065868263473, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1247, "step": 24997 }, { "epoch": 37.42215568862275, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1229, "step": 24998 }, { "epoch": 37.42365269461078, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1209, "step": 24999 }, { "epoch": 37.4251497005988, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1289, "step": 25000 }, { "epoch": 37.42664670658683, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1221, "step": 25001 }, { "epoch": 37.42814371257485, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1217, "step": 25002 }, { "epoch": 37.42964071856287, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1279, "step": 25003 }, { "epoch": 37.4311377245509, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1213, "step": 25004 }, { "epoch": 37.43263473053892, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1255, "step": 25005 }, { "epoch": 37.43413173652694, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1225, "step": 25006 }, { "epoch": 37.43562874251497, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1282, "step": 25007 }, { "epoch": 37.43712574850299, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1266, "step": 25008 }, { "epoch": 37.43862275449102, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 25009 }, { "epoch": 37.44011976047904, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1255, "step": 25010 }, { "epoch": 37.441616766467064, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.125, "step": 25011 }, { "epoch": 37.44311377245509, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1325, "step": 25012 }, { "epoch": 37.44461077844311, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1218, "step": 25013 }, { "epoch": 37.446107784431135, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1239, "step": 25014 }, { "epoch": 37.44760479041916, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1208, "step": 25015 }, { "epoch": 37.449101796407184, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1211, "step": 25016 }, { "epoch": 37.45059880239521, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1267, "step": 25017 }, { "epoch": 37.452095808383234, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1244, "step": 25018 }, { "epoch": 37.453592814371255, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.127, "step": 25019 }, { "epoch": 37.455089820359284, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1188, "step": 25020 }, { "epoch": 37.456586826347305, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1268, "step": 25021 }, { "epoch": 37.458083832335326, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1282, "step": 25022 }, { "epoch": 37.459580838323355, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1264, "step": 25023 }, { "epoch": 37.461077844311376, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1307, "step": 25024 }, { "epoch": 37.462574850299404, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1199, "step": 25025 }, { "epoch": 37.464071856287426, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.123, "step": 25026 }, { "epoch": 37.46556886227545, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1219, "step": 25027 }, { "epoch": 37.467065868263475, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1252, "step": 25028 }, { "epoch": 37.4685628742515, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 25029 }, { "epoch": 37.47005988023952, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1296, "step": 25030 }, { "epoch": 37.471556886227546, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.127, "step": 25031 }, { "epoch": 37.47305389221557, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1282, "step": 25032 }, { "epoch": 37.474550898203596, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1267, "step": 25033 }, { "epoch": 37.47604790419162, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1245, "step": 25034 }, { "epoch": 37.47754491017964, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1162, "step": 25035 }, { "epoch": 37.47904191616767, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1243, "step": 25036 }, { "epoch": 37.48053892215569, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1329, "step": 25037 }, { "epoch": 37.48203592814371, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1241, "step": 25038 }, { "epoch": 37.48353293413174, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.129, "step": 25039 }, { "epoch": 37.48502994011976, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1272, "step": 25040 }, { "epoch": 37.48652694610779, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 25041 }, { "epoch": 37.48802395209581, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1236, "step": 25042 }, { "epoch": 37.48952095808383, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1262, "step": 25043 }, { "epoch": 37.49101796407186, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1272, "step": 25044 }, { "epoch": 37.49251497005988, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1325, "step": 25045 }, { "epoch": 37.4940119760479, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1246, "step": 25046 }, { "epoch": 37.49550898203593, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1303, "step": 25047 }, { "epoch": 37.49700598802395, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1227, "step": 25048 }, { "epoch": 37.49850299401198, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1204, "step": 25049 }, { "epoch": 37.5, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1202, "step": 25050 }, { "epoch": 37.50149700598802, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1239, "step": 25051 }, { "epoch": 37.50299401197605, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1241, "step": 25052 }, { "epoch": 37.50449101796407, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1255, "step": 25053 }, { "epoch": 37.5059880239521, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1249, "step": 25054 }, { "epoch": 37.50748502994012, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1251, "step": 25055 }, { "epoch": 37.50898203592814, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1213, "step": 25056 }, { "epoch": 37.51047904191617, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1254, "step": 25057 }, { "epoch": 37.51197604790419, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1239, "step": 25058 }, { "epoch": 37.51347305389221, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1261, "step": 25059 }, { "epoch": 37.51497005988024, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1266, "step": 25060 }, { "epoch": 37.51646706586826, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1212, "step": 25061 }, { "epoch": 37.51796407185629, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1179, "step": 25062 }, { "epoch": 37.51946107784431, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1281, "step": 25063 }, { "epoch": 37.52095808383233, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1282, "step": 25064 }, { "epoch": 37.52245508982036, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1275, "step": 25065 }, { "epoch": 37.52395209580838, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1262, "step": 25066 }, { "epoch": 37.525449101796404, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1215, "step": 25067 }, { "epoch": 37.52694610778443, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1275, "step": 25068 }, { "epoch": 37.528443113772454, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1288, "step": 25069 }, { "epoch": 37.52994011976048, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.13, "step": 25070 }, { "epoch": 37.5314371257485, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1248, "step": 25071 }, { "epoch": 37.532934131736525, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1254, "step": 25072 }, { "epoch": 37.53443113772455, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1288, "step": 25073 }, { "epoch": 37.535928143712574, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 25074 }, { "epoch": 37.537425149700596, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1219, "step": 25075 }, { "epoch": 37.538922155688624, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1237, "step": 25076 }, { "epoch": 37.540419161676645, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1217, "step": 25077 }, { "epoch": 37.541916167664674, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1286, "step": 25078 }, { "epoch": 37.543413173652695, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.122, "step": 25079 }, { "epoch": 37.544910179640716, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1215, "step": 25080 }, { "epoch": 37.546407185628745, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1275, "step": 25081 }, { "epoch": 37.547904191616766, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1346, "step": 25082 }, { "epoch": 37.54940119760479, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1191, "step": 25083 }, { "epoch": 37.550898203592816, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1223, "step": 25084 }, { "epoch": 37.55239520958084, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1283, "step": 25085 }, { "epoch": 37.553892215568865, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 25086 }, { "epoch": 37.55538922155689, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1235, "step": 25087 }, { "epoch": 37.55688622754491, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1293, "step": 25088 }, { "epoch": 37.558383233532936, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1269, "step": 25089 }, { "epoch": 37.55988023952096, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1245, "step": 25090 }, { "epoch": 37.56137724550898, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1241, "step": 25091 }, { "epoch": 37.56287425149701, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1225, "step": 25092 }, { "epoch": 37.56437125748503, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1269, "step": 25093 }, { "epoch": 37.56586826347306, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1209, "step": 25094 }, { "epoch": 37.56736526946108, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1227, "step": 25095 }, { "epoch": 37.5688622754491, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1265, "step": 25096 }, { "epoch": 37.57035928143713, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 25097 }, { "epoch": 37.57185628742515, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1258, "step": 25098 }, { "epoch": 37.57335329341317, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1284, "step": 25099 }, { "epoch": 37.5748502994012, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1257, "step": 25100 }, { "epoch": 37.57634730538922, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.129, "step": 25101 }, { "epoch": 37.57784431137725, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1244, "step": 25102 }, { "epoch": 37.57934131736527, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1276, "step": 25103 }, { "epoch": 37.58083832335329, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 25104 }, { "epoch": 37.58233532934132, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1284, "step": 25105 }, { "epoch": 37.58383233532934, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1219, "step": 25106 }, { "epoch": 37.58532934131736, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1319, "step": 25107 }, { "epoch": 37.58682634730539, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1211, "step": 25108 }, { "epoch": 37.58832335329341, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1278, "step": 25109 }, { "epoch": 37.58982035928144, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1257, "step": 25110 }, { "epoch": 37.59131736526946, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1317, "step": 25111 }, { "epoch": 37.59281437125748, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1281, "step": 25112 }, { "epoch": 37.59431137724551, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1324, "step": 25113 }, { "epoch": 37.59580838323353, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1262, "step": 25114 }, { "epoch": 37.59730538922156, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1281, "step": 25115 }, { "epoch": 37.59880239520958, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1245, "step": 25116 }, { "epoch": 37.6002994011976, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1277, "step": 25117 }, { "epoch": 37.60179640718563, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1305, "step": 25118 }, { "epoch": 37.60329341317365, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1204, "step": 25119 }, { "epoch": 37.604790419161674, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1201, "step": 25120 }, { "epoch": 37.6062874251497, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1257, "step": 25121 }, { "epoch": 37.60778443113772, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1262, "step": 25122 }, { "epoch": 37.60928143712575, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1236, "step": 25123 }, { "epoch": 37.61077844311377, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1253, "step": 25124 }, { "epoch": 37.612275449101794, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1212, "step": 25125 }, { "epoch": 37.61377245508982, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1272, "step": 25126 }, { "epoch": 37.615269461077844, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1196, "step": 25127 }, { "epoch": 37.616766467065865, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1278, "step": 25128 }, { "epoch": 37.618263473053894, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.129, "step": 25129 }, { "epoch": 37.619760479041915, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1308, "step": 25130 }, { "epoch": 37.62125748502994, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1256, "step": 25131 }, { "epoch": 37.622754491017965, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1248, "step": 25132 }, { "epoch": 37.624251497005986, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1277, "step": 25133 }, { "epoch": 37.625748502994014, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1311, "step": 25134 }, { "epoch": 37.627245508982035, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1273, "step": 25135 }, { "epoch": 37.62874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1246, "step": 25136 }, { "epoch": 37.630239520958085, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1232, "step": 25137 }, { "epoch": 37.631736526946106, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1159, "step": 25138 }, { "epoch": 37.633233532934135, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1254, "step": 25139 }, { "epoch": 37.634730538922156, "grad_norm": 0.051025390625, "learning_rate": 0.0008, "loss": 1.1233, "step": 25140 }, { "epoch": 37.63622754491018, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1238, "step": 25141 }, { "epoch": 37.637724550898206, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1274, "step": 25142 }, { "epoch": 37.63922155688623, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1221, "step": 25143 }, { "epoch": 37.64071856287425, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1267, "step": 25144 }, { "epoch": 37.64221556886228, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1272, "step": 25145 }, { "epoch": 37.6437125748503, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1295, "step": 25146 }, { "epoch": 37.645209580838326, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.13, "step": 25147 }, { "epoch": 37.64670658682635, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1317, "step": 25148 }, { "epoch": 37.64820359281437, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1232, "step": 25149 }, { "epoch": 37.6497005988024, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1186, "step": 25150 }, { "epoch": 37.65119760479042, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1272, "step": 25151 }, { "epoch": 37.65269461077844, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1221, "step": 25152 }, { "epoch": 37.65419161676647, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1223, "step": 25153 }, { "epoch": 37.65568862275449, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1278, "step": 25154 }, { "epoch": 37.65718562874252, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1275, "step": 25155 }, { "epoch": 37.65868263473054, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1247, "step": 25156 }, { "epoch": 37.66017964071856, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1268, "step": 25157 }, { "epoch": 37.66167664670659, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1233, "step": 25158 }, { "epoch": 37.66317365269461, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1274, "step": 25159 }, { "epoch": 37.66467065868264, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1219, "step": 25160 }, { "epoch": 37.66616766467066, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1308, "step": 25161 }, { "epoch": 37.66766467065868, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1272, "step": 25162 }, { "epoch": 37.66916167664671, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1261, "step": 25163 }, { "epoch": 37.67065868263473, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 25164 }, { "epoch": 37.67215568862275, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1204, "step": 25165 }, { "epoch": 37.67365269461078, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1251, "step": 25166 }, { "epoch": 37.6751497005988, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1297, "step": 25167 }, { "epoch": 37.67664670658683, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1264, "step": 25168 }, { "epoch": 37.67814371257485, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1313, "step": 25169 }, { "epoch": 37.67964071856287, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1231, "step": 25170 }, { "epoch": 37.6811377245509, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1243, "step": 25171 }, { "epoch": 37.68263473053892, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1294, "step": 25172 }, { "epoch": 37.68413173652694, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1267, "step": 25173 }, { "epoch": 37.68562874251497, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.121, "step": 25174 }, { "epoch": 37.68712574850299, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1247, "step": 25175 }, { "epoch": 37.68862275449102, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1263, "step": 25176 }, { "epoch": 37.69011976047904, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.121, "step": 25177 }, { "epoch": 37.691616766467064, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1227, "step": 25178 }, { "epoch": 37.69311377245509, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1225, "step": 25179 }, { "epoch": 37.69461077844311, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1313, "step": 25180 }, { "epoch": 37.696107784431135, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.121, "step": 25181 }, { "epoch": 37.69760479041916, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1179, "step": 25182 }, { "epoch": 37.699101796407184, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1209, "step": 25183 }, { "epoch": 37.70059880239521, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1166, "step": 25184 }, { "epoch": 37.702095808383234, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.126, "step": 25185 }, { "epoch": 37.703592814371255, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1239, "step": 25186 }, { "epoch": 37.705089820359284, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.131, "step": 25187 }, { "epoch": 37.706586826347305, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1341, "step": 25188 }, { "epoch": 37.708083832335326, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1305, "step": 25189 }, { "epoch": 37.709580838323355, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1247, "step": 25190 }, { "epoch": 37.711077844311376, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.128, "step": 25191 }, { "epoch": 37.712574850299404, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.125, "step": 25192 }, { "epoch": 37.714071856287426, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1234, "step": 25193 }, { "epoch": 37.71556886227545, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1241, "step": 25194 }, { "epoch": 37.717065868263475, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1219, "step": 25195 }, { "epoch": 37.7185628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1272, "step": 25196 }, { "epoch": 37.72005988023952, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1239, "step": 25197 }, { "epoch": 37.721556886227546, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1308, "step": 25198 }, { "epoch": 37.72305389221557, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1271, "step": 25199 }, { "epoch": 37.724550898203596, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.129, "step": 25200 }, { "epoch": 37.72604790419162, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1257, "step": 25201 }, { "epoch": 37.72754491017964, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1207, "step": 25202 }, { "epoch": 37.72904191616767, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1187, "step": 25203 }, { "epoch": 37.73053892215569, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1213, "step": 25204 }, { "epoch": 37.73203592814371, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1328, "step": 25205 }, { "epoch": 37.73353293413174, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1192, "step": 25206 }, { "epoch": 37.73502994011976, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1217, "step": 25207 }, { "epoch": 37.73652694610779, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1202, "step": 25208 }, { "epoch": 37.73802395209581, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1267, "step": 25209 }, { "epoch": 37.73952095808383, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1219, "step": 25210 }, { "epoch": 37.74101796407186, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1242, "step": 25211 }, { "epoch": 37.74251497005988, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1254, "step": 25212 }, { "epoch": 37.7440119760479, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1272, "step": 25213 }, { "epoch": 37.74550898203593, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1255, "step": 25214 }, { "epoch": 37.74700598802395, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 25215 }, { "epoch": 37.74850299401198, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1255, "step": 25216 }, { "epoch": 37.75, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1222, "step": 25217 }, { "epoch": 37.75149700598802, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1238, "step": 25218 }, { "epoch": 37.75299401197605, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 25219 }, { "epoch": 37.75449101796407, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1304, "step": 25220 }, { "epoch": 37.7559880239521, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1197, "step": 25221 }, { "epoch": 37.75748502994012, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1232, "step": 25222 }, { "epoch": 37.75898203592814, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1272, "step": 25223 }, { "epoch": 37.76047904191617, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1272, "step": 25224 }, { "epoch": 37.76197604790419, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1285, "step": 25225 }, { "epoch": 37.76347305389221, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1321, "step": 25226 }, { "epoch": 37.76497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1312, "step": 25227 }, { "epoch": 37.76646706586826, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1262, "step": 25228 }, { "epoch": 37.76796407185629, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1342, "step": 25229 }, { "epoch": 37.76946107784431, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1255, "step": 25230 }, { "epoch": 37.77095808383233, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1198, "step": 25231 }, { "epoch": 37.77245508982036, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1213, "step": 25232 }, { "epoch": 37.77395209580838, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1194, "step": 25233 }, { "epoch": 37.775449101796404, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1255, "step": 25234 }, { "epoch": 37.77694610778443, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1271, "step": 25235 }, { "epoch": 37.778443113772454, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1243, "step": 25236 }, { "epoch": 37.77994011976048, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 25237 }, { "epoch": 37.7814371257485, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1252, "step": 25238 }, { "epoch": 37.782934131736525, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1322, "step": 25239 }, { "epoch": 37.78443113772455, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1273, "step": 25240 }, { "epoch": 37.785928143712574, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1239, "step": 25241 }, { "epoch": 37.787425149700596, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1311, "step": 25242 }, { "epoch": 37.788922155688624, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1252, "step": 25243 }, { "epoch": 37.790419161676645, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.117, "step": 25244 }, { "epoch": 37.791916167664674, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1283, "step": 25245 }, { "epoch": 37.793413173652695, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1293, "step": 25246 }, { "epoch": 37.794910179640716, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1231, "step": 25247 }, { "epoch": 37.796407185628745, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 25248 }, { "epoch": 37.797904191616766, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1286, "step": 25249 }, { "epoch": 37.79940119760479, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1223, "step": 25250 }, { "epoch": 37.800898203592816, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.129, "step": 25251 }, { "epoch": 37.80239520958084, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1238, "step": 25252 }, { "epoch": 37.803892215568865, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1269, "step": 25253 }, { "epoch": 37.80538922155689, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1269, "step": 25254 }, { "epoch": 37.80688622754491, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1231, "step": 25255 }, { "epoch": 37.808383233532936, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1208, "step": 25256 }, { "epoch": 37.80988023952096, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 25257 }, { "epoch": 37.81137724550898, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1316, "step": 25258 }, { "epoch": 37.81287425149701, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1252, "step": 25259 }, { "epoch": 37.81437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1326, "step": 25260 }, { "epoch": 37.81586826347306, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1271, "step": 25261 }, { "epoch": 37.81736526946108, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1289, "step": 25262 }, { "epoch": 37.8188622754491, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 25263 }, { "epoch": 37.82035928143713, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.125, "step": 25264 }, { "epoch": 37.82185628742515, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.126, "step": 25265 }, { "epoch": 37.82335329341317, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1303, "step": 25266 }, { "epoch": 37.8248502994012, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1314, "step": 25267 }, { "epoch": 37.82634730538922, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.123, "step": 25268 }, { "epoch": 37.82784431137725, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.126, "step": 25269 }, { "epoch": 37.82934131736527, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1185, "step": 25270 }, { "epoch": 37.83083832335329, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.121, "step": 25271 }, { "epoch": 37.83233532934132, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1249, "step": 25272 }, { "epoch": 37.83383233532934, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1264, "step": 25273 }, { "epoch": 37.83532934131736, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1218, "step": 25274 }, { "epoch": 37.83682634730539, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1235, "step": 25275 }, { "epoch": 37.83832335329341, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1186, "step": 25276 }, { "epoch": 37.83982035928144, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1251, "step": 25277 }, { "epoch": 37.84131736526946, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1199, "step": 25278 }, { "epoch": 37.84281437125748, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1215, "step": 25279 }, { "epoch": 37.84431137724551, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1245, "step": 25280 }, { "epoch": 37.84580838323353, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1267, "step": 25281 }, { "epoch": 37.84730538922156, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1197, "step": 25282 }, { "epoch": 37.84880239520958, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.12, "step": 25283 }, { "epoch": 37.8502994011976, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1247, "step": 25284 }, { "epoch": 37.85179640718563, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1289, "step": 25285 }, { "epoch": 37.85329341317365, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1205, "step": 25286 }, { "epoch": 37.854790419161674, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 25287 }, { "epoch": 37.8562874251497, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.127, "step": 25288 }, { "epoch": 37.85778443113772, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1233, "step": 25289 }, { "epoch": 37.85928143712575, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.128, "step": 25290 }, { "epoch": 37.86077844311377, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1276, "step": 25291 }, { "epoch": 37.862275449101794, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1233, "step": 25292 }, { "epoch": 37.86377245508982, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1243, "step": 25293 }, { "epoch": 37.865269461077844, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.127, "step": 25294 }, { "epoch": 37.866766467065865, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1168, "step": 25295 }, { "epoch": 37.868263473053894, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1163, "step": 25296 }, { "epoch": 37.869760479041915, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1269, "step": 25297 }, { "epoch": 37.87125748502994, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1268, "step": 25298 }, { "epoch": 37.872754491017965, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1268, "step": 25299 }, { "epoch": 37.874251497005986, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1268, "step": 25300 }, { "epoch": 37.875748502994014, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1314, "step": 25301 }, { "epoch": 37.877245508982035, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1245, "step": 25302 }, { "epoch": 37.87874251497006, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.128, "step": 25303 }, { "epoch": 37.880239520958085, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1315, "step": 25304 }, { "epoch": 37.881736526946106, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.127, "step": 25305 }, { "epoch": 37.883233532934135, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1318, "step": 25306 }, { "epoch": 37.884730538922156, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1212, "step": 25307 }, { "epoch": 37.88622754491018, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1231, "step": 25308 }, { "epoch": 37.887724550898206, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1312, "step": 25309 }, { "epoch": 37.88922155688623, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.123, "step": 25310 }, { "epoch": 37.89071856287425, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1263, "step": 25311 }, { "epoch": 37.89221556886228, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1178, "step": 25312 }, { "epoch": 37.8937125748503, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1251, "step": 25313 }, { "epoch": 37.895209580838326, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.122, "step": 25314 }, { "epoch": 37.89670658682635, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1189, "step": 25315 }, { "epoch": 37.89820359281437, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1246, "step": 25316 }, { "epoch": 37.8997005988024, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1235, "step": 25317 }, { "epoch": 37.90119760479042, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1205, "step": 25318 }, { "epoch": 37.90269461077844, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1293, "step": 25319 }, { "epoch": 37.90419161676647, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1219, "step": 25320 }, { "epoch": 37.90568862275449, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1191, "step": 25321 }, { "epoch": 37.90718562874252, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1304, "step": 25322 }, { "epoch": 37.90868263473054, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1284, "step": 25323 }, { "epoch": 37.91017964071856, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1199, "step": 25324 }, { "epoch": 37.91167664670659, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.131, "step": 25325 }, { "epoch": 37.91317365269461, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1262, "step": 25326 }, { "epoch": 37.91467065868264, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1262, "step": 25327 }, { "epoch": 37.91616766467066, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1241, "step": 25328 }, { "epoch": 37.91766467065868, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.121, "step": 25329 }, { "epoch": 37.91916167664671, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1264, "step": 25330 }, { "epoch": 37.92065868263473, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1208, "step": 25331 }, { "epoch": 37.92215568862275, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1253, "step": 25332 }, { "epoch": 37.92365269461078, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1265, "step": 25333 }, { "epoch": 37.9251497005988, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1245, "step": 25334 }, { "epoch": 37.92664670658683, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1304, "step": 25335 }, { "epoch": 37.92814371257485, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1219, "step": 25336 }, { "epoch": 37.92964071856287, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1272, "step": 25337 }, { "epoch": 37.9311377245509, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1299, "step": 25338 }, { "epoch": 37.93263473053892, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.127, "step": 25339 }, { "epoch": 37.93413173652694, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 25340 }, { "epoch": 37.93562874251497, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.134, "step": 25341 }, { "epoch": 37.93712574850299, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1261, "step": 25342 }, { "epoch": 37.93862275449102, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1289, "step": 25343 }, { "epoch": 37.94011976047904, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1249, "step": 25344 }, { "epoch": 37.941616766467064, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1259, "step": 25345 }, { "epoch": 37.94311377245509, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1302, "step": 25346 }, { "epoch": 37.94461077844311, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1226, "step": 25347 }, { "epoch": 37.946107784431135, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1254, "step": 25348 }, { "epoch": 37.94760479041916, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1215, "step": 25349 }, { "epoch": 37.949101796407184, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1293, "step": 25350 }, { "epoch": 37.95059880239521, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1267, "step": 25351 }, { "epoch": 37.952095808383234, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1266, "step": 25352 }, { "epoch": 37.953592814371255, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1305, "step": 25353 }, { "epoch": 37.955089820359284, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.122, "step": 25354 }, { "epoch": 37.956586826347305, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.13, "step": 25355 }, { "epoch": 37.958083832335326, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1251, "step": 25356 }, { "epoch": 37.959580838323355, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1269, "step": 25357 }, { "epoch": 37.961077844311376, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1264, "step": 25358 }, { "epoch": 37.962574850299404, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1275, "step": 25359 }, { "epoch": 37.964071856287426, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1284, "step": 25360 }, { "epoch": 37.96556886227545, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1272, "step": 25361 }, { "epoch": 37.967065868263475, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1255, "step": 25362 }, { "epoch": 37.9685628742515, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1217, "step": 25363 }, { "epoch": 37.97005988023952, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1244, "step": 25364 }, { "epoch": 37.971556886227546, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 25365 }, { "epoch": 37.97305389221557, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1278, "step": 25366 }, { "epoch": 37.974550898203596, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1263, "step": 25367 }, { "epoch": 37.97604790419162, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.127, "step": 25368 }, { "epoch": 37.97754491017964, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1153, "step": 25369 }, { "epoch": 37.97904191616767, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1216, "step": 25370 }, { "epoch": 37.98053892215569, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1268, "step": 25371 }, { "epoch": 37.98203592814371, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.125, "step": 25372 }, { "epoch": 37.98353293413174, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1252, "step": 25373 }, { "epoch": 37.98502994011976, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1305, "step": 25374 }, { "epoch": 37.98652694610779, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1255, "step": 25375 }, { "epoch": 37.98802395209581, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 25376 }, { "epoch": 37.98952095808383, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1251, "step": 25377 }, { "epoch": 37.99101796407186, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1276, "step": 25378 }, { "epoch": 37.99251497005988, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1251, "step": 25379 }, { "epoch": 37.9940119760479, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1176, "step": 25380 }, { "epoch": 37.99550898203593, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.127, "step": 25381 }, { "epoch": 37.99700598802395, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1257, "step": 25382 }, { "epoch": 37.99850299401198, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1192, "step": 25383 }, { "epoch": 38.0, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1261, "step": 25384 }, { "epoch": 38.00149700598802, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1174, "step": 25385 }, { "epoch": 38.00299401197605, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1226, "step": 25386 }, { "epoch": 38.00449101796407, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1189, "step": 25387 }, { "epoch": 38.0059880239521, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1274, "step": 25388 }, { "epoch": 38.00748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1209, "step": 25389 }, { "epoch": 38.00898203592814, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 25390 }, { "epoch": 38.01047904191617, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1206, "step": 25391 }, { "epoch": 38.01197604790419, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1218, "step": 25392 }, { "epoch": 38.01347305389221, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1224, "step": 25393 }, { "epoch": 38.01497005988024, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1234, "step": 25394 }, { "epoch": 38.01646706586826, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1244, "step": 25395 }, { "epoch": 38.01796407185629, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1226, "step": 25396 }, { "epoch": 38.01946107784431, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1235, "step": 25397 }, { "epoch": 38.02095808383233, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1243, "step": 25398 }, { "epoch": 38.02245508982036, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1224, "step": 25399 }, { "epoch": 38.02395209580838, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1248, "step": 25400 }, { "epoch": 38.025449101796404, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1191, "step": 25401 }, { "epoch": 38.02694610778443, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1253, "step": 25402 }, { "epoch": 38.028443113772454, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1236, "step": 25403 }, { "epoch": 38.02994011976048, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1311, "step": 25404 }, { "epoch": 38.0314371257485, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1236, "step": 25405 }, { "epoch": 38.032934131736525, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1262, "step": 25406 }, { "epoch": 38.03443113772455, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1246, "step": 25407 }, { "epoch": 38.035928143712574, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1257, "step": 25408 }, { "epoch": 38.037425149700596, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1245, "step": 25409 }, { "epoch": 38.038922155688624, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1235, "step": 25410 }, { "epoch": 38.040419161676645, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1236, "step": 25411 }, { "epoch": 38.041916167664674, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.134, "step": 25412 }, { "epoch": 38.043413173652695, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.126, "step": 25413 }, { "epoch": 38.044910179640716, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 25414 }, { "epoch": 38.046407185628745, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1282, "step": 25415 }, { "epoch": 38.047904191616766, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.124, "step": 25416 }, { "epoch": 38.04940119760479, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1212, "step": 25417 }, { "epoch": 38.050898203592816, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1238, "step": 25418 }, { "epoch": 38.05239520958084, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.123, "step": 25419 }, { "epoch": 38.053892215568865, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1239, "step": 25420 }, { "epoch": 38.05538922155689, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1248, "step": 25421 }, { "epoch": 38.05688622754491, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1253, "step": 25422 }, { "epoch": 38.058383233532936, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1232, "step": 25423 }, { "epoch": 38.05988023952096, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1283, "step": 25424 }, { "epoch": 38.06137724550898, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1318, "step": 25425 }, { "epoch": 38.06287425149701, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1183, "step": 25426 }, { "epoch": 38.06437125748503, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.125, "step": 25427 }, { "epoch": 38.06586826347306, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1201, "step": 25428 }, { "epoch": 38.06736526946108, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1208, "step": 25429 }, { "epoch": 38.0688622754491, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1245, "step": 25430 }, { "epoch": 38.07035928143713, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1242, "step": 25431 }, { "epoch": 38.07185628742515, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.128, "step": 25432 }, { "epoch": 38.07335329341317, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1257, "step": 25433 }, { "epoch": 38.0748502994012, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1306, "step": 25434 }, { "epoch": 38.07634730538922, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1241, "step": 25435 }, { "epoch": 38.07784431137725, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1273, "step": 25436 }, { "epoch": 38.07934131736527, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1274, "step": 25437 }, { "epoch": 38.08083832335329, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1229, "step": 25438 }, { "epoch": 38.08233532934132, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1238, "step": 25439 }, { "epoch": 38.08383233532934, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1265, "step": 25440 }, { "epoch": 38.08532934131737, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1231, "step": 25441 }, { "epoch": 38.08682634730539, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1207, "step": 25442 }, { "epoch": 38.08832335329341, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1275, "step": 25443 }, { "epoch": 38.08982035928144, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1292, "step": 25444 }, { "epoch": 38.09131736526946, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1223, "step": 25445 }, { "epoch": 38.09281437125748, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1278, "step": 25446 }, { "epoch": 38.09431137724551, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1241, "step": 25447 }, { "epoch": 38.09580838323353, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1251, "step": 25448 }, { "epoch": 38.09730538922156, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1213, "step": 25449 }, { "epoch": 38.09880239520958, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1229, "step": 25450 }, { "epoch": 38.1002994011976, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1257, "step": 25451 }, { "epoch": 38.10179640718563, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.123, "step": 25452 }, { "epoch": 38.10329341317365, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1267, "step": 25453 }, { "epoch": 38.104790419161674, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1263, "step": 25454 }, { "epoch": 38.1062874251497, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1204, "step": 25455 }, { "epoch": 38.10778443113772, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1184, "step": 25456 }, { "epoch": 38.10928143712575, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1223, "step": 25457 }, { "epoch": 38.11077844311377, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1234, "step": 25458 }, { "epoch": 38.112275449101794, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1278, "step": 25459 }, { "epoch": 38.11377245508982, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1209, "step": 25460 }, { "epoch": 38.115269461077844, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.127, "step": 25461 }, { "epoch": 38.116766467065865, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1211, "step": 25462 }, { "epoch": 38.118263473053894, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.13, "step": 25463 }, { "epoch": 38.119760479041915, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1291, "step": 25464 }, { "epoch": 38.12125748502994, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1281, "step": 25465 }, { "epoch": 38.122754491017965, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1316, "step": 25466 }, { "epoch": 38.124251497005986, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1297, "step": 25467 }, { "epoch": 38.125748502994014, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.121, "step": 25468 }, { "epoch": 38.127245508982035, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1231, "step": 25469 }, { "epoch": 38.12874251497006, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1246, "step": 25470 }, { "epoch": 38.130239520958085, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1227, "step": 25471 }, { "epoch": 38.131736526946106, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1284, "step": 25472 }, { "epoch": 38.133233532934135, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1226, "step": 25473 }, { "epoch": 38.134730538922156, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1246, "step": 25474 }, { "epoch": 38.13622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1214, "step": 25475 }, { "epoch": 38.137724550898206, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1259, "step": 25476 }, { "epoch": 38.13922155688623, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1235, "step": 25477 }, { "epoch": 38.14071856287425, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1248, "step": 25478 }, { "epoch": 38.14221556886228, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.122, "step": 25479 }, { "epoch": 38.1437125748503, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1214, "step": 25480 }, { "epoch": 38.145209580838326, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1178, "step": 25481 }, { "epoch": 38.14670658682635, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1233, "step": 25482 }, { "epoch": 38.14820359281437, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1234, "step": 25483 }, { "epoch": 38.1497005988024, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1259, "step": 25484 }, { "epoch": 38.15119760479042, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1202, "step": 25485 }, { "epoch": 38.15269461077844, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.122, "step": 25486 }, { "epoch": 38.15419161676647, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1248, "step": 25487 }, { "epoch": 38.15568862275449, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1218, "step": 25488 }, { "epoch": 38.15718562874252, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.127, "step": 25489 }, { "epoch": 38.15868263473054, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1264, "step": 25490 }, { "epoch": 38.16017964071856, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.123, "step": 25491 }, { "epoch": 38.16167664670659, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1292, "step": 25492 }, { "epoch": 38.16317365269461, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1254, "step": 25493 }, { "epoch": 38.16467065868263, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.122, "step": 25494 }, { "epoch": 38.16616766467066, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1256, "step": 25495 }, { "epoch": 38.16766467065868, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1237, "step": 25496 }, { "epoch": 38.16916167664671, "grad_norm": 0.051025390625, "learning_rate": 0.0008, "loss": 1.1298, "step": 25497 }, { "epoch": 38.17065868263473, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 25498 }, { "epoch": 38.17215568862275, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1193, "step": 25499 }, { "epoch": 38.17365269461078, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.124, "step": 25500 }, { "epoch": 38.1751497005988, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 25501 }, { "epoch": 38.17664670658683, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1199, "step": 25502 }, { "epoch": 38.17814371257485, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1266, "step": 25503 }, { "epoch": 38.17964071856287, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1228, "step": 25504 }, { "epoch": 38.1811377245509, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1297, "step": 25505 }, { "epoch": 38.18263473053892, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1254, "step": 25506 }, { "epoch": 38.18413173652694, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1279, "step": 25507 }, { "epoch": 38.18562874251497, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1216, "step": 25508 }, { "epoch": 38.18712574850299, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1272, "step": 25509 }, { "epoch": 38.18862275449102, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1216, "step": 25510 }, { "epoch": 38.19011976047904, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1253, "step": 25511 }, { "epoch": 38.191616766467064, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.125, "step": 25512 }, { "epoch": 38.19311377245509, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1202, "step": 25513 }, { "epoch": 38.19461077844311, "grad_norm": 0.2265625, "learning_rate": 0.0008, "loss": 1.1321, "step": 25514 }, { "epoch": 38.196107784431135, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1274, "step": 25515 }, { "epoch": 38.19760479041916, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1215, "step": 25516 }, { "epoch": 38.199101796407184, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1264, "step": 25517 }, { "epoch": 38.20059880239521, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1241, "step": 25518 }, { "epoch": 38.202095808383234, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.124, "step": 25519 }, { "epoch": 38.203592814371255, "grad_norm": 0.1650390625, "learning_rate": 0.0008, "loss": 1.1251, "step": 25520 }, { "epoch": 38.205089820359284, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1265, "step": 25521 }, { "epoch": 38.206586826347305, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1291, "step": 25522 }, { "epoch": 38.208083832335326, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1223, "step": 25523 }, { "epoch": 38.209580838323355, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1275, "step": 25524 }, { "epoch": 38.211077844311376, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1314, "step": 25525 }, { "epoch": 38.212574850299404, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1217, "step": 25526 }, { "epoch": 38.214071856287426, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1266, "step": 25527 }, { "epoch": 38.21556886227545, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1237, "step": 25528 }, { "epoch": 38.217065868263475, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1263, "step": 25529 }, { "epoch": 38.2185628742515, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.125, "step": 25530 }, { "epoch": 38.22005988023952, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1289, "step": 25531 }, { "epoch": 38.221556886227546, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1242, "step": 25532 }, { "epoch": 38.22305389221557, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1269, "step": 25533 }, { "epoch": 38.224550898203596, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.125, "step": 25534 }, { "epoch": 38.22604790419162, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1288, "step": 25535 }, { "epoch": 38.22754491017964, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1192, "step": 25536 }, { "epoch": 38.22904191616767, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1298, "step": 25537 }, { "epoch": 38.23053892215569, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1298, "step": 25538 }, { "epoch": 38.23203592814371, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1215, "step": 25539 }, { "epoch": 38.23353293413174, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1227, "step": 25540 }, { "epoch": 38.23502994011976, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1191, "step": 25541 }, { "epoch": 38.23652694610779, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1232, "step": 25542 }, { "epoch": 38.23802395209581, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1242, "step": 25543 }, { "epoch": 38.23952095808383, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1234, "step": 25544 }, { "epoch": 38.24101796407186, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1196, "step": 25545 }, { "epoch": 38.24251497005988, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1247, "step": 25546 }, { "epoch": 38.2440119760479, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1253, "step": 25547 }, { "epoch": 38.24550898203593, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1291, "step": 25548 }, { "epoch": 38.24700598802395, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1249, "step": 25549 }, { "epoch": 38.24850299401198, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.127, "step": 25550 }, { "epoch": 38.25, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1246, "step": 25551 }, { "epoch": 38.25149700598802, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 25552 }, { "epoch": 38.25299401197605, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1224, "step": 25553 }, { "epoch": 38.25449101796407, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1245, "step": 25554 }, { "epoch": 38.2559880239521, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1194, "step": 25555 }, { "epoch": 38.25748502994012, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1284, "step": 25556 }, { "epoch": 38.25898203592814, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 25557 }, { "epoch": 38.26047904191617, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1252, "step": 25558 }, { "epoch": 38.26197604790419, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1298, "step": 25559 }, { "epoch": 38.26347305389221, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1234, "step": 25560 }, { "epoch": 38.26497005988024, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1176, "step": 25561 }, { "epoch": 38.26646706586826, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1249, "step": 25562 }, { "epoch": 38.26796407185629, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1239, "step": 25563 }, { "epoch": 38.26946107784431, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1296, "step": 25564 }, { "epoch": 38.27095808383233, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1216, "step": 25565 }, { "epoch": 38.27245508982036, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1228, "step": 25566 }, { "epoch": 38.27395209580838, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1269, "step": 25567 }, { "epoch": 38.275449101796404, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1273, "step": 25568 }, { "epoch": 38.27694610778443, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1245, "step": 25569 }, { "epoch": 38.278443113772454, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1315, "step": 25570 }, { "epoch": 38.27994011976048, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1321, "step": 25571 }, { "epoch": 38.2814371257485, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1211, "step": 25572 }, { "epoch": 38.282934131736525, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.128, "step": 25573 }, { "epoch": 38.28443113772455, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1187, "step": 25574 }, { "epoch": 38.285928143712574, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1222, "step": 25575 }, { "epoch": 38.287425149700596, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1218, "step": 25576 }, { "epoch": 38.288922155688624, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1216, "step": 25577 }, { "epoch": 38.290419161676645, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1269, "step": 25578 }, { "epoch": 38.291916167664674, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1245, "step": 25579 }, { "epoch": 38.293413173652695, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1224, "step": 25580 }, { "epoch": 38.294910179640716, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1286, "step": 25581 }, { "epoch": 38.296407185628745, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1258, "step": 25582 }, { "epoch": 38.297904191616766, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1213, "step": 25583 }, { "epoch": 38.29940119760479, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1248, "step": 25584 }, { "epoch": 38.300898203592816, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1259, "step": 25585 }, { "epoch": 38.30239520958084, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1243, "step": 25586 }, { "epoch": 38.303892215568865, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1241, "step": 25587 }, { "epoch": 38.30538922155689, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1284, "step": 25588 }, { "epoch": 38.30688622754491, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1229, "step": 25589 }, { "epoch": 38.308383233532936, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 25590 }, { "epoch": 38.30988023952096, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1212, "step": 25591 }, { "epoch": 38.31137724550898, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1259, "step": 25592 }, { "epoch": 38.31287425149701, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.126, "step": 25593 }, { "epoch": 38.31437125748503, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1221, "step": 25594 }, { "epoch": 38.31586826347306, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.125, "step": 25595 }, { "epoch": 38.31736526946108, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1253, "step": 25596 }, { "epoch": 38.3188622754491, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1174, "step": 25597 }, { "epoch": 38.32035928143713, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1244, "step": 25598 }, { "epoch": 38.32185628742515, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1236, "step": 25599 }, { "epoch": 38.32335329341317, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.119, "step": 25600 }, { "epoch": 38.3248502994012, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1243, "step": 25601 }, { "epoch": 38.32634730538922, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1214, "step": 25602 }, { "epoch": 38.32784431137725, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1246, "step": 25603 }, { "epoch": 38.32934131736527, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1289, "step": 25604 }, { "epoch": 38.33083832335329, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1307, "step": 25605 }, { "epoch": 38.33233532934132, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1242, "step": 25606 }, { "epoch": 38.33383233532934, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1205, "step": 25607 }, { "epoch": 38.33532934131736, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1307, "step": 25608 }, { "epoch": 38.33682634730539, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1282, "step": 25609 }, { "epoch": 38.33832335329341, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1258, "step": 25610 }, { "epoch": 38.33982035928144, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1319, "step": 25611 }, { "epoch": 38.34131736526946, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1329, "step": 25612 }, { "epoch": 38.34281437125748, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.125, "step": 25613 }, { "epoch": 38.34431137724551, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1212, "step": 25614 }, { "epoch": 38.34580838323353, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1273, "step": 25615 }, { "epoch": 38.34730538922156, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1227, "step": 25616 }, { "epoch": 38.34880239520958, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1253, "step": 25617 }, { "epoch": 38.3502994011976, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1284, "step": 25618 }, { "epoch": 38.35179640718563, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1265, "step": 25619 }, { "epoch": 38.35329341317365, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1237, "step": 25620 }, { "epoch": 38.354790419161674, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.123, "step": 25621 }, { "epoch": 38.3562874251497, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1303, "step": 25622 }, { "epoch": 38.35778443113772, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1319, "step": 25623 }, { "epoch": 38.35928143712575, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1256, "step": 25624 }, { "epoch": 38.36077844311377, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1182, "step": 25625 }, { "epoch": 38.362275449101794, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1244, "step": 25626 }, { "epoch": 38.36377245508982, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1238, "step": 25627 }, { "epoch": 38.365269461077844, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1247, "step": 25628 }, { "epoch": 38.366766467065865, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.129, "step": 25629 }, { "epoch": 38.368263473053894, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1261, "step": 25630 }, { "epoch": 38.369760479041915, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1257, "step": 25631 }, { "epoch": 38.37125748502994, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1242, "step": 25632 }, { "epoch": 38.372754491017965, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1289, "step": 25633 }, { "epoch": 38.374251497005986, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1246, "step": 25634 }, { "epoch": 38.375748502994014, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1256, "step": 25635 }, { "epoch": 38.377245508982035, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1291, "step": 25636 }, { "epoch": 38.37874251497006, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1206, "step": 25637 }, { "epoch": 38.380239520958085, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1224, "step": 25638 }, { "epoch": 38.381736526946106, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.127, "step": 25639 }, { "epoch": 38.383233532934135, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1265, "step": 25640 }, { "epoch": 38.384730538922156, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1246, "step": 25641 }, { "epoch": 38.38622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1324, "step": 25642 }, { "epoch": 38.387724550898206, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1259, "step": 25643 }, { "epoch": 38.38922155688623, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1298, "step": 25644 }, { "epoch": 38.39071856287425, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.123, "step": 25645 }, { "epoch": 38.39221556886228, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1266, "step": 25646 }, { "epoch": 38.3937125748503, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.12, "step": 25647 }, { "epoch": 38.395209580838326, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1252, "step": 25648 }, { "epoch": 38.39670658682635, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.122, "step": 25649 }, { "epoch": 38.39820359281437, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.122, "step": 25650 }, { "epoch": 38.3997005988024, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1251, "step": 25651 }, { "epoch": 38.40119760479042, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1223, "step": 25652 }, { "epoch": 38.40269461077844, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1209, "step": 25653 }, { "epoch": 38.40419161676647, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1256, "step": 25654 }, { "epoch": 38.40568862275449, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1243, "step": 25655 }, { "epoch": 38.40718562874252, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1252, "step": 25656 }, { "epoch": 38.40868263473054, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1272, "step": 25657 }, { "epoch": 38.41017964071856, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1282, "step": 25658 }, { "epoch": 38.41167664670659, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1259, "step": 25659 }, { "epoch": 38.41317365269461, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1245, "step": 25660 }, { "epoch": 38.41467065868264, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1182, "step": 25661 }, { "epoch": 38.41616766467066, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1265, "step": 25662 }, { "epoch": 38.41766467065868, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1232, "step": 25663 }, { "epoch": 38.41916167664671, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1288, "step": 25664 }, { "epoch": 38.42065868263473, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1228, "step": 25665 }, { "epoch": 38.42215568862275, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1161, "step": 25666 }, { "epoch": 38.42365269461078, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1221, "step": 25667 }, { "epoch": 38.4251497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1224, "step": 25668 }, { "epoch": 38.42664670658683, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1263, "step": 25669 }, { "epoch": 38.42814371257485, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1215, "step": 25670 }, { "epoch": 38.42964071856287, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1234, "step": 25671 }, { "epoch": 38.4311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1253, "step": 25672 }, { "epoch": 38.43263473053892, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1232, "step": 25673 }, { "epoch": 38.43413173652694, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1265, "step": 25674 }, { "epoch": 38.43562874251497, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1315, "step": 25675 }, { "epoch": 38.43712574850299, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1266, "step": 25676 }, { "epoch": 38.43862275449102, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1248, "step": 25677 }, { "epoch": 38.44011976047904, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1214, "step": 25678 }, { "epoch": 38.441616766467064, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1283, "step": 25679 }, { "epoch": 38.44311377245509, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1259, "step": 25680 }, { "epoch": 38.44461077844311, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1229, "step": 25681 }, { "epoch": 38.446107784431135, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1279, "step": 25682 }, { "epoch": 38.44760479041916, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1227, "step": 25683 }, { "epoch": 38.449101796407184, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1207, "step": 25684 }, { "epoch": 38.45059880239521, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1259, "step": 25685 }, { "epoch": 38.452095808383234, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.125, "step": 25686 }, { "epoch": 38.453592814371255, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1198, "step": 25687 }, { "epoch": 38.455089820359284, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1184, "step": 25688 }, { "epoch": 38.456586826347305, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1268, "step": 25689 }, { "epoch": 38.458083832335326, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1253, "step": 25690 }, { "epoch": 38.459580838323355, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1247, "step": 25691 }, { "epoch": 38.461077844311376, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1176, "step": 25692 }, { "epoch": 38.462574850299404, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.127, "step": 25693 }, { "epoch": 38.464071856287426, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.125, "step": 25694 }, { "epoch": 38.46556886227545, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1206, "step": 25695 }, { "epoch": 38.467065868263475, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 25696 }, { "epoch": 38.4685628742515, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1306, "step": 25697 }, { "epoch": 38.47005988023952, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1231, "step": 25698 }, { "epoch": 38.471556886227546, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1232, "step": 25699 }, { "epoch": 38.47305389221557, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 25700 }, { "epoch": 38.474550898203596, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1255, "step": 25701 }, { "epoch": 38.47604790419162, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1226, "step": 25702 }, { "epoch": 38.47754491017964, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1187, "step": 25703 }, { "epoch": 38.47904191616767, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1268, "step": 25704 }, { "epoch": 38.48053892215569, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1301, "step": 25705 }, { "epoch": 38.48203592814371, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1226, "step": 25706 }, { "epoch": 38.48353293413174, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1324, "step": 25707 }, { "epoch": 38.48502994011976, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1248, "step": 25708 }, { "epoch": 38.48652694610779, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1232, "step": 25709 }, { "epoch": 38.48802395209581, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1253, "step": 25710 }, { "epoch": 38.48952095808383, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1291, "step": 25711 }, { "epoch": 38.49101796407186, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1243, "step": 25712 }, { "epoch": 38.49251497005988, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1266, "step": 25713 }, { "epoch": 38.4940119760479, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 25714 }, { "epoch": 38.49550898203593, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1229, "step": 25715 }, { "epoch": 38.49700598802395, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1291, "step": 25716 }, { "epoch": 38.49850299401198, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1249, "step": 25717 }, { "epoch": 38.5, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.13, "step": 25718 }, { "epoch": 38.50149700598802, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1258, "step": 25719 }, { "epoch": 38.50299401197605, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1216, "step": 25720 }, { "epoch": 38.50449101796407, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1211, "step": 25721 }, { "epoch": 38.5059880239521, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1203, "step": 25722 }, { "epoch": 38.50748502994012, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1295, "step": 25723 }, { "epoch": 38.50898203592814, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1252, "step": 25724 }, { "epoch": 38.51047904191617, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.12, "step": 25725 }, { "epoch": 38.51197604790419, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 25726 }, { "epoch": 38.51347305389221, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1225, "step": 25727 }, { "epoch": 38.51497005988024, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1208, "step": 25728 }, { "epoch": 38.51646706586826, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1287, "step": 25729 }, { "epoch": 38.51796407185629, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1289, "step": 25730 }, { "epoch": 38.51946107784431, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1256, "step": 25731 }, { "epoch": 38.52095808383233, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1287, "step": 25732 }, { "epoch": 38.52245508982036, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1267, "step": 25733 }, { "epoch": 38.52395209580838, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1278, "step": 25734 }, { "epoch": 38.525449101796404, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1261, "step": 25735 }, { "epoch": 38.52694610778443, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1291, "step": 25736 }, { "epoch": 38.528443113772454, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1265, "step": 25737 }, { "epoch": 38.52994011976048, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1324, "step": 25738 }, { "epoch": 38.5314371257485, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1269, "step": 25739 }, { "epoch": 38.532934131736525, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 25740 }, { "epoch": 38.53443113772455, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1215, "step": 25741 }, { "epoch": 38.535928143712574, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1332, "step": 25742 }, { "epoch": 38.537425149700596, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1285, "step": 25743 }, { "epoch": 38.538922155688624, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1223, "step": 25744 }, { "epoch": 38.540419161676645, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1232, "step": 25745 }, { "epoch": 38.541916167664674, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1257, "step": 25746 }, { "epoch": 38.543413173652695, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1173, "step": 25747 }, { "epoch": 38.544910179640716, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1236, "step": 25748 }, { "epoch": 38.546407185628745, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1216, "step": 25749 }, { "epoch": 38.547904191616766, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1268, "step": 25750 }, { "epoch": 38.54940119760479, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1215, "step": 25751 }, { "epoch": 38.550898203592816, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1193, "step": 25752 }, { "epoch": 38.55239520958084, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 25753 }, { "epoch": 38.553892215568865, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1301, "step": 25754 }, { "epoch": 38.55538922155689, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1215, "step": 25755 }, { "epoch": 38.55688622754491, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1318, "step": 25756 }, { "epoch": 38.558383233532936, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1356, "step": 25757 }, { "epoch": 38.55988023952096, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1251, "step": 25758 }, { "epoch": 38.56137724550898, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 25759 }, { "epoch": 38.56287425149701, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1189, "step": 25760 }, { "epoch": 38.56437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1231, "step": 25761 }, { "epoch": 38.56586826347306, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1241, "step": 25762 }, { "epoch": 38.56736526946108, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1216, "step": 25763 }, { "epoch": 38.5688622754491, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1222, "step": 25764 }, { "epoch": 38.57035928143713, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1235, "step": 25765 }, { "epoch": 38.57185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1208, "step": 25766 }, { "epoch": 38.57335329341317, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1266, "step": 25767 }, { "epoch": 38.5748502994012, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1255, "step": 25768 }, { "epoch": 38.57634730538922, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1288, "step": 25769 }, { "epoch": 38.57784431137725, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1275, "step": 25770 }, { "epoch": 38.57934131736527, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1207, "step": 25771 }, { "epoch": 38.58083832335329, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.126, "step": 25772 }, { "epoch": 38.58233532934132, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1228, "step": 25773 }, { "epoch": 38.58383233532934, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1293, "step": 25774 }, { "epoch": 38.58532934131736, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1301, "step": 25775 }, { "epoch": 38.58682634730539, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1273, "step": 25776 }, { "epoch": 38.58832335329341, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1286, "step": 25777 }, { "epoch": 38.58982035928144, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1223, "step": 25778 }, { "epoch": 38.59131736526946, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1244, "step": 25779 }, { "epoch": 38.59281437125748, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1265, "step": 25780 }, { "epoch": 38.59431137724551, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1326, "step": 25781 }, { "epoch": 38.59580838323353, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1238, "step": 25782 }, { "epoch": 38.59730538922156, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1301, "step": 25783 }, { "epoch": 38.59880239520958, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1223, "step": 25784 }, { "epoch": 38.6002994011976, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1256, "step": 25785 }, { "epoch": 38.60179640718563, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1265, "step": 25786 }, { "epoch": 38.60329341317365, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1224, "step": 25787 }, { "epoch": 38.604790419161674, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1226, "step": 25788 }, { "epoch": 38.6062874251497, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1259, "step": 25789 }, { "epoch": 38.60778443113772, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1218, "step": 25790 }, { "epoch": 38.60928143712575, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1279, "step": 25791 }, { "epoch": 38.61077844311377, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1285, "step": 25792 }, { "epoch": 38.612275449101794, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.135, "step": 25793 }, { "epoch": 38.61377245508982, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1294, "step": 25794 }, { "epoch": 38.615269461077844, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1269, "step": 25795 }, { "epoch": 38.616766467065865, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.113, "step": 25796 }, { "epoch": 38.618263473053894, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.123, "step": 25797 }, { "epoch": 38.619760479041915, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1217, "step": 25798 }, { "epoch": 38.62125748502994, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1167, "step": 25799 }, { "epoch": 38.622754491017965, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1236, "step": 25800 }, { "epoch": 38.624251497005986, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1271, "step": 25801 }, { "epoch": 38.625748502994014, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1202, "step": 25802 }, { "epoch": 38.627245508982035, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1239, "step": 25803 }, { "epoch": 38.62874251497006, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1279, "step": 25804 }, { "epoch": 38.630239520958085, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1209, "step": 25805 }, { "epoch": 38.631736526946106, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1212, "step": 25806 }, { "epoch": 38.633233532934135, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1217, "step": 25807 }, { "epoch": 38.634730538922156, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1251, "step": 25808 }, { "epoch": 38.63622754491018, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1218, "step": 25809 }, { "epoch": 38.637724550898206, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1245, "step": 25810 }, { "epoch": 38.63922155688623, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1286, "step": 25811 }, { "epoch": 38.64071856287425, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1273, "step": 25812 }, { "epoch": 38.64221556886228, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 25813 }, { "epoch": 38.6437125748503, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1239, "step": 25814 }, { "epoch": 38.645209580838326, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1193, "step": 25815 }, { "epoch": 38.64670658682635, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1233, "step": 25816 }, { "epoch": 38.64820359281437, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1279, "step": 25817 }, { "epoch": 38.6497005988024, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1184, "step": 25818 }, { "epoch": 38.65119760479042, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1263, "step": 25819 }, { "epoch": 38.65269461077844, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1257, "step": 25820 }, { "epoch": 38.65419161676647, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1222, "step": 25821 }, { "epoch": 38.65568862275449, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.12, "step": 25822 }, { "epoch": 38.65718562874252, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1265, "step": 25823 }, { "epoch": 38.65868263473054, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1256, "step": 25824 }, { "epoch": 38.66017964071856, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1297, "step": 25825 }, { "epoch": 38.66167664670659, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1265, "step": 25826 }, { "epoch": 38.66317365269461, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.127, "step": 25827 }, { "epoch": 38.66467065868264, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1303, "step": 25828 }, { "epoch": 38.66616766467066, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.124, "step": 25829 }, { "epoch": 38.66766467065868, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1225, "step": 25830 }, { "epoch": 38.66916167664671, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1226, "step": 25831 }, { "epoch": 38.67065868263473, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1249, "step": 25832 }, { "epoch": 38.67215568862275, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.126, "step": 25833 }, { "epoch": 38.67365269461078, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1229, "step": 25834 }, { "epoch": 38.6751497005988, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1272, "step": 25835 }, { "epoch": 38.67664670658683, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1255, "step": 25836 }, { "epoch": 38.67814371257485, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1255, "step": 25837 }, { "epoch": 38.67964071856287, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1283, "step": 25838 }, { "epoch": 38.6811377245509, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1225, "step": 25839 }, { "epoch": 38.68263473053892, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1228, "step": 25840 }, { "epoch": 38.68413173652694, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.123, "step": 25841 }, { "epoch": 38.68562874251497, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1295, "step": 25842 }, { "epoch": 38.68712574850299, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1231, "step": 25843 }, { "epoch": 38.68862275449102, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1305, "step": 25844 }, { "epoch": 38.69011976047904, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1314, "step": 25845 }, { "epoch": 38.691616766467064, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.125, "step": 25846 }, { "epoch": 38.69311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1224, "step": 25847 }, { "epoch": 38.69461077844311, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 25848 }, { "epoch": 38.696107784431135, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1237, "step": 25849 }, { "epoch": 38.69760479041916, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1264, "step": 25850 }, { "epoch": 38.699101796407184, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 25851 }, { "epoch": 38.70059880239521, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1288, "step": 25852 }, { "epoch": 38.702095808383234, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1279, "step": 25853 }, { "epoch": 38.703592814371255, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 25854 }, { "epoch": 38.705089820359284, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1236, "step": 25855 }, { "epoch": 38.706586826347305, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1315, "step": 25856 }, { "epoch": 38.708083832335326, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1254, "step": 25857 }, { "epoch": 38.709580838323355, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1221, "step": 25858 }, { "epoch": 38.711077844311376, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1181, "step": 25859 }, { "epoch": 38.712574850299404, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.12, "step": 25860 }, { "epoch": 38.714071856287426, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1238, "step": 25861 }, { "epoch": 38.71556886227545, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1251, "step": 25862 }, { "epoch": 38.717065868263475, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1237, "step": 25863 }, { "epoch": 38.7185628742515, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1194, "step": 25864 }, { "epoch": 38.72005988023952, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1238, "step": 25865 }, { "epoch": 38.721556886227546, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1233, "step": 25866 }, { "epoch": 38.72305389221557, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1303, "step": 25867 }, { "epoch": 38.724550898203596, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1258, "step": 25868 }, { "epoch": 38.72604790419162, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1277, "step": 25869 }, { "epoch": 38.72754491017964, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1227, "step": 25870 }, { "epoch": 38.72904191616767, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1245, "step": 25871 }, { "epoch": 38.73053892215569, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1206, "step": 25872 }, { "epoch": 38.73203592814371, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1221, "step": 25873 }, { "epoch": 38.73353293413174, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1263, "step": 25874 }, { "epoch": 38.73502994011976, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1222, "step": 25875 }, { "epoch": 38.73652694610779, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1287, "step": 25876 }, { "epoch": 38.73802395209581, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1215, "step": 25877 }, { "epoch": 38.73952095808383, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1217, "step": 25878 }, { "epoch": 38.74101796407186, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1249, "step": 25879 }, { "epoch": 38.74251497005988, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1241, "step": 25880 }, { "epoch": 38.7440119760479, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.119, "step": 25881 }, { "epoch": 38.74550898203593, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1281, "step": 25882 }, { "epoch": 38.74700598802395, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1233, "step": 25883 }, { "epoch": 38.74850299401198, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1205, "step": 25884 }, { "epoch": 38.75, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1289, "step": 25885 }, { "epoch": 38.75149700598802, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1242, "step": 25886 }, { "epoch": 38.75299401197605, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.128, "step": 25887 }, { "epoch": 38.75449101796407, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1211, "step": 25888 }, { "epoch": 38.7559880239521, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1238, "step": 25889 }, { "epoch": 38.75748502994012, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1284, "step": 25890 }, { "epoch": 38.75898203592814, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1239, "step": 25891 }, { "epoch": 38.76047904191617, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1306, "step": 25892 }, { "epoch": 38.76197604790419, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1322, "step": 25893 }, { "epoch": 38.76347305389221, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1242, "step": 25894 }, { "epoch": 38.76497005988024, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1275, "step": 25895 }, { "epoch": 38.76646706586826, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1202, "step": 25896 }, { "epoch": 38.76796407185629, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1309, "step": 25897 }, { "epoch": 38.76946107784431, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1272, "step": 25898 }, { "epoch": 38.77095808383233, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1246, "step": 25899 }, { "epoch": 38.77245508982036, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1227, "step": 25900 }, { "epoch": 38.77395209580838, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1283, "step": 25901 }, { "epoch": 38.775449101796404, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1252, "step": 25902 }, { "epoch": 38.77694610778443, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.121, "step": 25903 }, { "epoch": 38.778443113772454, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1197, "step": 25904 }, { "epoch": 38.77994011976048, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 25905 }, { "epoch": 38.7814371257485, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1205, "step": 25906 }, { "epoch": 38.782934131736525, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1213, "step": 25907 }, { "epoch": 38.78443113772455, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1266, "step": 25908 }, { "epoch": 38.785928143712574, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1198, "step": 25909 }, { "epoch": 38.787425149700596, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.121, "step": 25910 }, { "epoch": 38.788922155688624, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1295, "step": 25911 }, { "epoch": 38.790419161676645, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 25912 }, { "epoch": 38.791916167664674, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1214, "step": 25913 }, { "epoch": 38.793413173652695, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1248, "step": 25914 }, { "epoch": 38.794910179640716, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1302, "step": 25915 }, { "epoch": 38.796407185628745, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1281, "step": 25916 }, { "epoch": 38.797904191616766, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1246, "step": 25917 }, { "epoch": 38.79940119760479, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1279, "step": 25918 }, { "epoch": 38.800898203592816, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.123, "step": 25919 }, { "epoch": 38.80239520958084, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1261, "step": 25920 }, { "epoch": 38.803892215568865, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1254, "step": 25921 }, { "epoch": 38.80538922155689, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1222, "step": 25922 }, { "epoch": 38.80688622754491, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 25923 }, { "epoch": 38.808383233532936, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1311, "step": 25924 }, { "epoch": 38.80988023952096, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1197, "step": 25925 }, { "epoch": 38.81137724550898, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1241, "step": 25926 }, { "epoch": 38.81287425149701, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1183, "step": 25927 }, { "epoch": 38.81437125748503, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1303, "step": 25928 }, { "epoch": 38.81586826347306, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1245, "step": 25929 }, { "epoch": 38.81736526946108, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1289, "step": 25930 }, { "epoch": 38.8188622754491, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1273, "step": 25931 }, { "epoch": 38.82035928143713, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1178, "step": 25932 }, { "epoch": 38.82185628742515, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1256, "step": 25933 }, { "epoch": 38.82335329341317, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1262, "step": 25934 }, { "epoch": 38.8248502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1258, "step": 25935 }, { "epoch": 38.82634730538922, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1253, "step": 25936 }, { "epoch": 38.82784431137725, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1201, "step": 25937 }, { "epoch": 38.82934131736527, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1219, "step": 25938 }, { "epoch": 38.83083832335329, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1235, "step": 25939 }, { "epoch": 38.83233532934132, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.125, "step": 25940 }, { "epoch": 38.83383233532934, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1318, "step": 25941 }, { "epoch": 38.83532934131736, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1244, "step": 25942 }, { "epoch": 38.83682634730539, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1258, "step": 25943 }, { "epoch": 38.83832335329341, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1231, "step": 25944 }, { "epoch": 38.83982035928144, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.127, "step": 25945 }, { "epoch": 38.84131736526946, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1235, "step": 25946 }, { "epoch": 38.84281437125748, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1182, "step": 25947 }, { "epoch": 38.84431137724551, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1227, "step": 25948 }, { "epoch": 38.84580838323353, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.125, "step": 25949 }, { "epoch": 38.84730538922156, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.122, "step": 25950 }, { "epoch": 38.84880239520958, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1223, "step": 25951 }, { "epoch": 38.8502994011976, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1303, "step": 25952 }, { "epoch": 38.85179640718563, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1295, "step": 25953 }, { "epoch": 38.85329341317365, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1264, "step": 25954 }, { "epoch": 38.854790419161674, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1273, "step": 25955 }, { "epoch": 38.8562874251497, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1226, "step": 25956 }, { "epoch": 38.85778443113772, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.123, "step": 25957 }, { "epoch": 38.85928143712575, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1297, "step": 25958 }, { "epoch": 38.86077844311377, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1235, "step": 25959 }, { "epoch": 38.862275449101794, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1284, "step": 25960 }, { "epoch": 38.86377245508982, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1237, "step": 25961 }, { "epoch": 38.865269461077844, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1305, "step": 25962 }, { "epoch": 38.866766467065865, "grad_norm": 0.05029296875, "learning_rate": 0.0008, "loss": 1.1183, "step": 25963 }, { "epoch": 38.868263473053894, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1157, "step": 25964 }, { "epoch": 38.869760479041915, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1277, "step": 25965 }, { "epoch": 38.87125748502994, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1196, "step": 25966 }, { "epoch": 38.872754491017965, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1217, "step": 25967 }, { "epoch": 38.874251497005986, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1237, "step": 25968 }, { "epoch": 38.875748502994014, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1245, "step": 25969 }, { "epoch": 38.877245508982035, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1202, "step": 25970 }, { "epoch": 38.87874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1242, "step": 25971 }, { "epoch": 38.880239520958085, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1303, "step": 25972 }, { "epoch": 38.881736526946106, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.12, "step": 25973 }, { "epoch": 38.883233532934135, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1247, "step": 25974 }, { "epoch": 38.884730538922156, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1241, "step": 25975 }, { "epoch": 38.88622754491018, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1268, "step": 25976 }, { "epoch": 38.887724550898206, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1291, "step": 25977 }, { "epoch": 38.88922155688623, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1293, "step": 25978 }, { "epoch": 38.89071856287425, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1205, "step": 25979 }, { "epoch": 38.89221556886228, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1263, "step": 25980 }, { "epoch": 38.8937125748503, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.135, "step": 25981 }, { "epoch": 38.895209580838326, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1291, "step": 25982 }, { "epoch": 38.89670658682635, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1261, "step": 25983 }, { "epoch": 38.89820359281437, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.126, "step": 25984 }, { "epoch": 38.8997005988024, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1288, "step": 25985 }, { "epoch": 38.90119760479042, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1268, "step": 25986 }, { "epoch": 38.90269461077844, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.124, "step": 25987 }, { "epoch": 38.90419161676647, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1251, "step": 25988 }, { "epoch": 38.90568862275449, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1216, "step": 25989 }, { "epoch": 38.90718562874252, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1185, "step": 25990 }, { "epoch": 38.90868263473054, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.125, "step": 25991 }, { "epoch": 38.91017964071856, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1257, "step": 25992 }, { "epoch": 38.91167664670659, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.124, "step": 25993 }, { "epoch": 38.91317365269461, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1234, "step": 25994 }, { "epoch": 38.91467065868264, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1229, "step": 25995 }, { "epoch": 38.91616766467066, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1257, "step": 25996 }, { "epoch": 38.91766467065868, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1272, "step": 25997 }, { "epoch": 38.91916167664671, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1276, "step": 25998 }, { "epoch": 38.92065868263473, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1268, "step": 25999 }, { "epoch": 38.92215568862275, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1267, "step": 26000 }, { "epoch": 38.92365269461078, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1205, "step": 26001 }, { "epoch": 38.9251497005988, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1247, "step": 26002 }, { "epoch": 38.92664670658683, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1191, "step": 26003 }, { "epoch": 38.92814371257485, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1196, "step": 26004 }, { "epoch": 38.92964071856287, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1299, "step": 26005 }, { "epoch": 38.9311377245509, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1205, "step": 26006 }, { "epoch": 38.93263473053892, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1253, "step": 26007 }, { "epoch": 38.93413173652694, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1268, "step": 26008 }, { "epoch": 38.93562874251497, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1199, "step": 26009 }, { "epoch": 38.93712574850299, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1297, "step": 26010 }, { "epoch": 38.93862275449102, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 26011 }, { "epoch": 38.94011976047904, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1286, "step": 26012 }, { "epoch": 38.941616766467064, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.128, "step": 26013 }, { "epoch": 38.94311377245509, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1218, "step": 26014 }, { "epoch": 38.94461077844311, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1225, "step": 26015 }, { "epoch": 38.946107784431135, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1261, "step": 26016 }, { "epoch": 38.94760479041916, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1226, "step": 26017 }, { "epoch": 38.949101796407184, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1267, "step": 26018 }, { "epoch": 38.95059880239521, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1241, "step": 26019 }, { "epoch": 38.952095808383234, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1205, "step": 26020 }, { "epoch": 38.953592814371255, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1289, "step": 26021 }, { "epoch": 38.955089820359284, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1297, "step": 26022 }, { "epoch": 38.956586826347305, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1248, "step": 26023 }, { "epoch": 38.958083832335326, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1324, "step": 26024 }, { "epoch": 38.959580838323355, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1275, "step": 26025 }, { "epoch": 38.961077844311376, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1253, "step": 26026 }, { "epoch": 38.962574850299404, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1217, "step": 26027 }, { "epoch": 38.964071856287426, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1244, "step": 26028 }, { "epoch": 38.96556886227545, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1206, "step": 26029 }, { "epoch": 38.967065868263475, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1279, "step": 26030 }, { "epoch": 38.9685628742515, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1237, "step": 26031 }, { "epoch": 38.97005988023952, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1326, "step": 26032 }, { "epoch": 38.971556886227546, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1261, "step": 26033 }, { "epoch": 38.97305389221557, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 26034 }, { "epoch": 38.974550898203596, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1295, "step": 26035 }, { "epoch": 38.97604790419162, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1248, "step": 26036 }, { "epoch": 38.97754491017964, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1243, "step": 26037 }, { "epoch": 38.97904191616767, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1224, "step": 26038 }, { "epoch": 38.98053892215569, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1219, "step": 26039 }, { "epoch": 38.98203592814371, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1253, "step": 26040 }, { "epoch": 38.98353293413174, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1249, "step": 26041 }, { "epoch": 38.98502994011976, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1255, "step": 26042 }, { "epoch": 38.98652694610779, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 26043 }, { "epoch": 38.98802395209581, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1258, "step": 26044 }, { "epoch": 38.98952095808383, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1247, "step": 26045 }, { "epoch": 38.99101796407186, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1242, "step": 26046 }, { "epoch": 38.99251497005988, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1231, "step": 26047 }, { "epoch": 38.9940119760479, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1244, "step": 26048 }, { "epoch": 38.99550898203593, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1253, "step": 26049 }, { "epoch": 38.99700598802395, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1243, "step": 26050 }, { "epoch": 38.99850299401198, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1294, "step": 26051 }, { "epoch": 39.0, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1237, "step": 26052 }, { "epoch": 39.00149700598802, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1141, "step": 26053 }, { "epoch": 39.00299401197605, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1227, "step": 26054 }, { "epoch": 39.00449101796407, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1302, "step": 26055 }, { "epoch": 39.0059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1302, "step": 26056 }, { "epoch": 39.00748502994012, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1192, "step": 26057 }, { "epoch": 39.00898203592814, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1257, "step": 26058 }, { "epoch": 39.01047904191617, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1208, "step": 26059 }, { "epoch": 39.01197604790419, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1237, "step": 26060 }, { "epoch": 39.01347305389221, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1255, "step": 26061 }, { "epoch": 39.01497005988024, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1265, "step": 26062 }, { "epoch": 39.01646706586826, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1228, "step": 26063 }, { "epoch": 39.01796407185629, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1259, "step": 26064 }, { "epoch": 39.01946107784431, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1235, "step": 26065 }, { "epoch": 39.02095808383233, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1263, "step": 26066 }, { "epoch": 39.02245508982036, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1275, "step": 26067 }, { "epoch": 39.02395209580838, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1326, "step": 26068 }, { "epoch": 39.025449101796404, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1229, "step": 26069 }, { "epoch": 39.02694610778443, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1339, "step": 26070 }, { "epoch": 39.028443113772454, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1218, "step": 26071 }, { "epoch": 39.02994011976048, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1286, "step": 26072 }, { "epoch": 39.0314371257485, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1253, "step": 26073 }, { "epoch": 39.032934131736525, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1226, "step": 26074 }, { "epoch": 39.03443113772455, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1276, "step": 26075 }, { "epoch": 39.035928143712574, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1259, "step": 26076 }, { "epoch": 39.037425149700596, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1291, "step": 26077 }, { "epoch": 39.038922155688624, "grad_norm": 0.05029296875, "learning_rate": 0.0008, "loss": 1.1263, "step": 26078 }, { "epoch": 39.040419161676645, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1272, "step": 26079 }, { "epoch": 39.041916167664674, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1263, "step": 26080 }, { "epoch": 39.043413173652695, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1217, "step": 26081 }, { "epoch": 39.044910179640716, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1209, "step": 26082 }, { "epoch": 39.046407185628745, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1284, "step": 26083 }, { "epoch": 39.047904191616766, "grad_norm": 0.04833984375, "learning_rate": 0.0008, "loss": 1.1273, "step": 26084 }, { "epoch": 39.04940119760479, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1252, "step": 26085 }, { "epoch": 39.050898203592816, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1228, "step": 26086 }, { "epoch": 39.05239520958084, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1304, "step": 26087 }, { "epoch": 39.053892215568865, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1221, "step": 26088 }, { "epoch": 39.05538922155689, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1268, "step": 26089 }, { "epoch": 39.05688622754491, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1321, "step": 26090 }, { "epoch": 39.058383233532936, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1234, "step": 26091 }, { "epoch": 39.05988023952096, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1289, "step": 26092 }, { "epoch": 39.06137724550898, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1183, "step": 26093 }, { "epoch": 39.06287425149701, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1241, "step": 26094 }, { "epoch": 39.06437125748503, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1223, "step": 26095 }, { "epoch": 39.06586826347306, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1272, "step": 26096 }, { "epoch": 39.06736526946108, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1217, "step": 26097 }, { "epoch": 39.0688622754491, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1292, "step": 26098 }, { "epoch": 39.07035928143713, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.125, "step": 26099 }, { "epoch": 39.07185628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1252, "step": 26100 }, { "epoch": 39.07335329341317, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1193, "step": 26101 }, { "epoch": 39.0748502994012, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1248, "step": 26102 }, { "epoch": 39.07634730538922, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1203, "step": 26103 }, { "epoch": 39.07784431137725, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1241, "step": 26104 }, { "epoch": 39.07934131736527, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1223, "step": 26105 }, { "epoch": 39.08083832335329, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1222, "step": 26106 }, { "epoch": 39.08233532934132, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1266, "step": 26107 }, { "epoch": 39.08383233532934, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.134, "step": 26108 }, { "epoch": 39.08532934131737, "grad_norm": 0.051025390625, "learning_rate": 0.0008, "loss": 1.1296, "step": 26109 }, { "epoch": 39.08682634730539, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1295, "step": 26110 }, { "epoch": 39.08832335329341, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1198, "step": 26111 }, { "epoch": 39.08982035928144, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1215, "step": 26112 }, { "epoch": 39.09131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1295, "step": 26113 }, { "epoch": 39.09281437125748, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1193, "step": 26114 }, { "epoch": 39.09431137724551, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1222, "step": 26115 }, { "epoch": 39.09580838323353, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1213, "step": 26116 }, { "epoch": 39.09730538922156, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.124, "step": 26117 }, { "epoch": 39.09880239520958, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1231, "step": 26118 }, { "epoch": 39.1002994011976, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1259, "step": 26119 }, { "epoch": 39.10179640718563, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1244, "step": 26120 }, { "epoch": 39.10329341317365, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1216, "step": 26121 }, { "epoch": 39.104790419161674, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1306, "step": 26122 }, { "epoch": 39.1062874251497, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1299, "step": 26123 }, { "epoch": 39.10778443113772, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.124, "step": 26124 }, { "epoch": 39.10928143712575, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1289, "step": 26125 }, { "epoch": 39.11077844311377, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1276, "step": 26126 }, { "epoch": 39.112275449101794, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1253, "step": 26127 }, { "epoch": 39.11377245508982, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1203, "step": 26128 }, { "epoch": 39.115269461077844, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1247, "step": 26129 }, { "epoch": 39.116766467065865, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1299, "step": 26130 }, { "epoch": 39.118263473053894, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1229, "step": 26131 }, { "epoch": 39.119760479041915, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.124, "step": 26132 }, { "epoch": 39.12125748502994, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1191, "step": 26133 }, { "epoch": 39.122754491017965, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1194, "step": 26134 }, { "epoch": 39.124251497005986, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1247, "step": 26135 }, { "epoch": 39.125748502994014, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1196, "step": 26136 }, { "epoch": 39.127245508982035, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1267, "step": 26137 }, { "epoch": 39.12874251497006, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1293, "step": 26138 }, { "epoch": 39.130239520958085, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.1246, "step": 26139 }, { "epoch": 39.131736526946106, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.121, "step": 26140 }, { "epoch": 39.133233532934135, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1331, "step": 26141 }, { "epoch": 39.134730538922156, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1238, "step": 26142 }, { "epoch": 39.13622754491018, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1229, "step": 26143 }, { "epoch": 39.137724550898206, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1206, "step": 26144 }, { "epoch": 39.13922155688623, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 26145 }, { "epoch": 39.14071856287425, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1235, "step": 26146 }, { "epoch": 39.14221556886228, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1253, "step": 26147 }, { "epoch": 39.1437125748503, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1245, "step": 26148 }, { "epoch": 39.145209580838326, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1191, "step": 26149 }, { "epoch": 39.14670658682635, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1344, "step": 26150 }, { "epoch": 39.14820359281437, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1206, "step": 26151 }, { "epoch": 39.1497005988024, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1255, "step": 26152 }, { "epoch": 39.15119760479042, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1276, "step": 26153 }, { "epoch": 39.15269461077844, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1301, "step": 26154 }, { "epoch": 39.15419161676647, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1167, "step": 26155 }, { "epoch": 39.15568862275449, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1224, "step": 26156 }, { "epoch": 39.15718562874252, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1356, "step": 26157 }, { "epoch": 39.15868263473054, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1269, "step": 26158 }, { "epoch": 39.16017964071856, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.122, "step": 26159 }, { "epoch": 39.16167664670659, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1207, "step": 26160 }, { "epoch": 39.16317365269461, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1263, "step": 26161 }, { "epoch": 39.16467065868263, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1212, "step": 26162 }, { "epoch": 39.16616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1218, "step": 26163 }, { "epoch": 39.16766467065868, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1265, "step": 26164 }, { "epoch": 39.16916167664671, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1245, "step": 26165 }, { "epoch": 39.17065868263473, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1246, "step": 26166 }, { "epoch": 39.17215568862275, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1238, "step": 26167 }, { "epoch": 39.17365269461078, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1269, "step": 26168 }, { "epoch": 39.1751497005988, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1271, "step": 26169 }, { "epoch": 39.17664670658683, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1297, "step": 26170 }, { "epoch": 39.17814371257485, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1202, "step": 26171 }, { "epoch": 39.17964071856287, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1165, "step": 26172 }, { "epoch": 39.1811377245509, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1271, "step": 26173 }, { "epoch": 39.18263473053892, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1246, "step": 26174 }, { "epoch": 39.18413173652694, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1245, "step": 26175 }, { "epoch": 39.18562874251497, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 26176 }, { "epoch": 39.18712574850299, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1209, "step": 26177 }, { "epoch": 39.18862275449102, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1244, "step": 26178 }, { "epoch": 39.19011976047904, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1212, "step": 26179 }, { "epoch": 39.191616766467064, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1271, "step": 26180 }, { "epoch": 39.19311377245509, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1206, "step": 26181 }, { "epoch": 39.19461077844311, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.128, "step": 26182 }, { "epoch": 39.196107784431135, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.119, "step": 26183 }, { "epoch": 39.19760479041916, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1203, "step": 26184 }, { "epoch": 39.199101796407184, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1199, "step": 26185 }, { "epoch": 39.20059880239521, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.128, "step": 26186 }, { "epoch": 39.202095808383234, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1234, "step": 26187 }, { "epoch": 39.203592814371255, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1275, "step": 26188 }, { "epoch": 39.205089820359284, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1197, "step": 26189 }, { "epoch": 39.206586826347305, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1243, "step": 26190 }, { "epoch": 39.208083832335326, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1256, "step": 26191 }, { "epoch": 39.209580838323355, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1266, "step": 26192 }, { "epoch": 39.211077844311376, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1319, "step": 26193 }, { "epoch": 39.212574850299404, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1227, "step": 26194 }, { "epoch": 39.214071856287426, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1249, "step": 26195 }, { "epoch": 39.21556886227545, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.123, "step": 26196 }, { "epoch": 39.217065868263475, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 26197 }, { "epoch": 39.2185628742515, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1245, "step": 26198 }, { "epoch": 39.22005988023952, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1249, "step": 26199 }, { "epoch": 39.221556886227546, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1202, "step": 26200 }, { "epoch": 39.22305389221557, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1275, "step": 26201 }, { "epoch": 39.224550898203596, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 26202 }, { "epoch": 39.22604790419162, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.129, "step": 26203 }, { "epoch": 39.22754491017964, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1202, "step": 26204 }, { "epoch": 39.22904191616767, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1241, "step": 26205 }, { "epoch": 39.23053892215569, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1209, "step": 26206 }, { "epoch": 39.23203592814371, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1258, "step": 26207 }, { "epoch": 39.23353293413174, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1208, "step": 26208 }, { "epoch": 39.23502994011976, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1209, "step": 26209 }, { "epoch": 39.23652694610779, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1249, "step": 26210 }, { "epoch": 39.23802395209581, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 26211 }, { "epoch": 39.23952095808383, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1232, "step": 26212 }, { "epoch": 39.24101796407186, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1236, "step": 26213 }, { "epoch": 39.24251497005988, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1216, "step": 26214 }, { "epoch": 39.2440119760479, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1226, "step": 26215 }, { "epoch": 39.24550898203593, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1175, "step": 26216 }, { "epoch": 39.24700598802395, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1233, "step": 26217 }, { "epoch": 39.24850299401198, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1186, "step": 26218 }, { "epoch": 39.25, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1239, "step": 26219 }, { "epoch": 39.25149700598802, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1306, "step": 26220 }, { "epoch": 39.25299401197605, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1186, "step": 26221 }, { "epoch": 39.25449101796407, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1248, "step": 26222 }, { "epoch": 39.2559880239521, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1223, "step": 26223 }, { "epoch": 39.25748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1273, "step": 26224 }, { "epoch": 39.25898203592814, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1235, "step": 26225 }, { "epoch": 39.26047904191617, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1231, "step": 26226 }, { "epoch": 39.26197604790419, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 26227 }, { "epoch": 39.26347305389221, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.123, "step": 26228 }, { "epoch": 39.26497005988024, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1283, "step": 26229 }, { "epoch": 39.26646706586826, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1209, "step": 26230 }, { "epoch": 39.26796407185629, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1189, "step": 26231 }, { "epoch": 39.26946107784431, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 26232 }, { "epoch": 39.27095808383233, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1292, "step": 26233 }, { "epoch": 39.27245508982036, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1282, "step": 26234 }, { "epoch": 39.27395209580838, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1211, "step": 26235 }, { "epoch": 39.275449101796404, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1264, "step": 26236 }, { "epoch": 39.27694610778443, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1286, "step": 26237 }, { "epoch": 39.278443113772454, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1234, "step": 26238 }, { "epoch": 39.27994011976048, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1277, "step": 26239 }, { "epoch": 39.2814371257485, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.121, "step": 26240 }, { "epoch": 39.282934131736525, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.128, "step": 26241 }, { "epoch": 39.28443113772455, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1205, "step": 26242 }, { "epoch": 39.285928143712574, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1186, "step": 26243 }, { "epoch": 39.287425149700596, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1196, "step": 26244 }, { "epoch": 39.288922155688624, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1326, "step": 26245 }, { "epoch": 39.290419161676645, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1308, "step": 26246 }, { "epoch": 39.291916167664674, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1261, "step": 26247 }, { "epoch": 39.293413173652695, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1208, "step": 26248 }, { "epoch": 39.294910179640716, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1271, "step": 26249 }, { "epoch": 39.296407185628745, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1207, "step": 26250 }, { "epoch": 39.297904191616766, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1149, "step": 26251 }, { "epoch": 39.29940119760479, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1232, "step": 26252 }, { "epoch": 39.300898203592816, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1209, "step": 26253 }, { "epoch": 39.30239520958084, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1348, "step": 26254 }, { "epoch": 39.303892215568865, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1199, "step": 26255 }, { "epoch": 39.30538922155689, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1215, "step": 26256 }, { "epoch": 39.30688622754491, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1229, "step": 26257 }, { "epoch": 39.308383233532936, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1204, "step": 26258 }, { "epoch": 39.30988023952096, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1299, "step": 26259 }, { "epoch": 39.31137724550898, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1234, "step": 26260 }, { "epoch": 39.31287425149701, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1255, "step": 26261 }, { "epoch": 39.31437125748503, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.125, "step": 26262 }, { "epoch": 39.31586826347306, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1246, "step": 26263 }, { "epoch": 39.31736526946108, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1249, "step": 26264 }, { "epoch": 39.3188622754491, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1259, "step": 26265 }, { "epoch": 39.32035928143713, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1237, "step": 26266 }, { "epoch": 39.32185628742515, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1227, "step": 26267 }, { "epoch": 39.32335329341317, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1218, "step": 26268 }, { "epoch": 39.3248502994012, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1226, "step": 26269 }, { "epoch": 39.32634730538922, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1235, "step": 26270 }, { "epoch": 39.32784431137725, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1216, "step": 26271 }, { "epoch": 39.32934131736527, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1271, "step": 26272 }, { "epoch": 39.33083832335329, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1251, "step": 26273 }, { "epoch": 39.33233532934132, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1228, "step": 26274 }, { "epoch": 39.33383233532934, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1237, "step": 26275 }, { "epoch": 39.33532934131736, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1207, "step": 26276 }, { "epoch": 39.33682634730539, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1296, "step": 26277 }, { "epoch": 39.33832335329341, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1303, "step": 26278 }, { "epoch": 39.33982035928144, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1258, "step": 26279 }, { "epoch": 39.34131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1202, "step": 26280 }, { "epoch": 39.34281437125748, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.126, "step": 26281 }, { "epoch": 39.34431137724551, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1219, "step": 26282 }, { "epoch": 39.34580838323353, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1257, "step": 26283 }, { "epoch": 39.34730538922156, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1214, "step": 26284 }, { "epoch": 39.34880239520958, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1212, "step": 26285 }, { "epoch": 39.3502994011976, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1236, "step": 26286 }, { "epoch": 39.35179640718563, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1202, "step": 26287 }, { "epoch": 39.35329341317365, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1279, "step": 26288 }, { "epoch": 39.354790419161674, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1205, "step": 26289 }, { "epoch": 39.3562874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1211, "step": 26290 }, { "epoch": 39.35778443113772, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1241, "step": 26291 }, { "epoch": 39.35928143712575, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1251, "step": 26292 }, { "epoch": 39.36077844311377, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1227, "step": 26293 }, { "epoch": 39.362275449101794, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1258, "step": 26294 }, { "epoch": 39.36377245508982, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1246, "step": 26295 }, { "epoch": 39.365269461077844, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1293, "step": 26296 }, { "epoch": 39.366766467065865, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1265, "step": 26297 }, { "epoch": 39.368263473053894, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1302, "step": 26298 }, { "epoch": 39.369760479041915, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1236, "step": 26299 }, { "epoch": 39.37125748502994, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1274, "step": 26300 }, { "epoch": 39.372754491017965, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1268, "step": 26301 }, { "epoch": 39.374251497005986, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1246, "step": 26302 }, { "epoch": 39.375748502994014, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1267, "step": 26303 }, { "epoch": 39.377245508982035, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1272, "step": 26304 }, { "epoch": 39.37874251497006, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1314, "step": 26305 }, { "epoch": 39.380239520958085, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1265, "step": 26306 }, { "epoch": 39.381736526946106, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 26307 }, { "epoch": 39.383233532934135, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1196, "step": 26308 }, { "epoch": 39.384730538922156, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1294, "step": 26309 }, { "epoch": 39.38622754491018, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1286, "step": 26310 }, { "epoch": 39.387724550898206, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1178, "step": 26311 }, { "epoch": 39.38922155688623, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1279, "step": 26312 }, { "epoch": 39.39071856287425, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1245, "step": 26313 }, { "epoch": 39.39221556886228, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1309, "step": 26314 }, { "epoch": 39.3937125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1179, "step": 26315 }, { "epoch": 39.395209580838326, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1323, "step": 26316 }, { "epoch": 39.39670658682635, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1228, "step": 26317 }, { "epoch": 39.39820359281437, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1296, "step": 26318 }, { "epoch": 39.3997005988024, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.133, "step": 26319 }, { "epoch": 39.40119760479042, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.128, "step": 26320 }, { "epoch": 39.40269461077844, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1301, "step": 26321 }, { "epoch": 39.40419161676647, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1227, "step": 26322 }, { "epoch": 39.40568862275449, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.125, "step": 26323 }, { "epoch": 39.40718562874252, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1243, "step": 26324 }, { "epoch": 39.40868263473054, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1241, "step": 26325 }, { "epoch": 39.41017964071856, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1245, "step": 26326 }, { "epoch": 39.41167664670659, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1245, "step": 26327 }, { "epoch": 39.41317365269461, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 26328 }, { "epoch": 39.41467065868264, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1222, "step": 26329 }, { "epoch": 39.41616766467066, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1193, "step": 26330 }, { "epoch": 39.41766467065868, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1237, "step": 26331 }, { "epoch": 39.41916167664671, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1253, "step": 26332 }, { "epoch": 39.42065868263473, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1292, "step": 26333 }, { "epoch": 39.42215568862275, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1239, "step": 26334 }, { "epoch": 39.42365269461078, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1258, "step": 26335 }, { "epoch": 39.4251497005988, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1228, "step": 26336 }, { "epoch": 39.42664670658683, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1289, "step": 26337 }, { "epoch": 39.42814371257485, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1227, "step": 26338 }, { "epoch": 39.42964071856287, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1254, "step": 26339 }, { "epoch": 39.4311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.124, "step": 26340 }, { "epoch": 39.43263473053892, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1185, "step": 26341 }, { "epoch": 39.43413173652694, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1257, "step": 26342 }, { "epoch": 39.43562874251497, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1272, "step": 26343 }, { "epoch": 39.43712574850299, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1231, "step": 26344 }, { "epoch": 39.43862275449102, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1192, "step": 26345 }, { "epoch": 39.44011976047904, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1131, "step": 26346 }, { "epoch": 39.441616766467064, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1315, "step": 26347 }, { "epoch": 39.44311377245509, "grad_norm": 0.05029296875, "learning_rate": 0.0008, "loss": 1.1195, "step": 26348 }, { "epoch": 39.44461077844311, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1295, "step": 26349 }, { "epoch": 39.446107784431135, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1263, "step": 26350 }, { "epoch": 39.44760479041916, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1215, "step": 26351 }, { "epoch": 39.449101796407184, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1241, "step": 26352 }, { "epoch": 39.45059880239521, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1286, "step": 26353 }, { "epoch": 39.452095808383234, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1259, "step": 26354 }, { "epoch": 39.453592814371255, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1227, "step": 26355 }, { "epoch": 39.455089820359284, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1215, "step": 26356 }, { "epoch": 39.456586826347305, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1321, "step": 26357 }, { "epoch": 39.458083832335326, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1229, "step": 26358 }, { "epoch": 39.459580838323355, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1266, "step": 26359 }, { "epoch": 39.461077844311376, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1261, "step": 26360 }, { "epoch": 39.462574850299404, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1204, "step": 26361 }, { "epoch": 39.464071856287426, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1245, "step": 26362 }, { "epoch": 39.46556886227545, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1241, "step": 26363 }, { "epoch": 39.467065868263475, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1292, "step": 26364 }, { "epoch": 39.4685628742515, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1253, "step": 26365 }, { "epoch": 39.47005988023952, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 26366 }, { "epoch": 39.471556886227546, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1198, "step": 26367 }, { "epoch": 39.47305389221557, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1227, "step": 26368 }, { "epoch": 39.474550898203596, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1243, "step": 26369 }, { "epoch": 39.47604790419162, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1158, "step": 26370 }, { "epoch": 39.47754491017964, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1138, "step": 26371 }, { "epoch": 39.47904191616767, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.127, "step": 26372 }, { "epoch": 39.48053892215569, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.127, "step": 26373 }, { "epoch": 39.48203592814371, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1251, "step": 26374 }, { "epoch": 39.48353293413174, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1258, "step": 26375 }, { "epoch": 39.48502994011976, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1275, "step": 26376 }, { "epoch": 39.48652694610779, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1233, "step": 26377 }, { "epoch": 39.48802395209581, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1284, "step": 26378 }, { "epoch": 39.48952095808383, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1235, "step": 26379 }, { "epoch": 39.49101796407186, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1236, "step": 26380 }, { "epoch": 39.49251497005988, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1214, "step": 26381 }, { "epoch": 39.4940119760479, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1266, "step": 26382 }, { "epoch": 39.49550898203593, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1279, "step": 26383 }, { "epoch": 39.49700598802395, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1253, "step": 26384 }, { "epoch": 39.49850299401198, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1263, "step": 26385 }, { "epoch": 39.5, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1245, "step": 26386 }, { "epoch": 39.50149700598802, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1288, "step": 26387 }, { "epoch": 39.50299401197605, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1288, "step": 26388 }, { "epoch": 39.50449101796407, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1243, "step": 26389 }, { "epoch": 39.5059880239521, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1254, "step": 26390 }, { "epoch": 39.50748502994012, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.121, "step": 26391 }, { "epoch": 39.50898203592814, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1247, "step": 26392 }, { "epoch": 39.51047904191617, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1245, "step": 26393 }, { "epoch": 39.51197604790419, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1215, "step": 26394 }, { "epoch": 39.51347305389221, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1286, "step": 26395 }, { "epoch": 39.51497005988024, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1217, "step": 26396 }, { "epoch": 39.51646706586826, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1241, "step": 26397 }, { "epoch": 39.51796407185629, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1218, "step": 26398 }, { "epoch": 39.51946107784431, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.121, "step": 26399 }, { "epoch": 39.52095808383233, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1185, "step": 26400 }, { "epoch": 39.52245508982036, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1196, "step": 26401 }, { "epoch": 39.52395209580838, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1238, "step": 26402 }, { "epoch": 39.525449101796404, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1263, "step": 26403 }, { "epoch": 39.52694610778443, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.122, "step": 26404 }, { "epoch": 39.528443113772454, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1183, "step": 26405 }, { "epoch": 39.52994011976048, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1286, "step": 26406 }, { "epoch": 39.5314371257485, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1185, "step": 26407 }, { "epoch": 39.532934131736525, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 26408 }, { "epoch": 39.53443113772455, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1271, "step": 26409 }, { "epoch": 39.535928143712574, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.124, "step": 26410 }, { "epoch": 39.537425149700596, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1316, "step": 26411 }, { "epoch": 39.538922155688624, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1232, "step": 26412 }, { "epoch": 39.540419161676645, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1241, "step": 26413 }, { "epoch": 39.541916167664674, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1231, "step": 26414 }, { "epoch": 39.543413173652695, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1216, "step": 26415 }, { "epoch": 39.544910179640716, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1206, "step": 26416 }, { "epoch": 39.546407185628745, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1223, "step": 26417 }, { "epoch": 39.547904191616766, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 26418 }, { "epoch": 39.54940119760479, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1249, "step": 26419 }, { "epoch": 39.550898203592816, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1193, "step": 26420 }, { "epoch": 39.55239520958084, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.129, "step": 26421 }, { "epoch": 39.553892215568865, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1271, "step": 26422 }, { "epoch": 39.55538922155689, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.125, "step": 26423 }, { "epoch": 39.55688622754491, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1274, "step": 26424 }, { "epoch": 39.558383233532936, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1235, "step": 26425 }, { "epoch": 39.55988023952096, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1249, "step": 26426 }, { "epoch": 39.56137724550898, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.126, "step": 26427 }, { "epoch": 39.56287425149701, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.124, "step": 26428 }, { "epoch": 39.56437125748503, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 26429 }, { "epoch": 39.56586826347306, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1246, "step": 26430 }, { "epoch": 39.56736526946108, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.121, "step": 26431 }, { "epoch": 39.5688622754491, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1277, "step": 26432 }, { "epoch": 39.57035928143713, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1319, "step": 26433 }, { "epoch": 39.57185628742515, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1226, "step": 26434 }, { "epoch": 39.57335329341317, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1252, "step": 26435 }, { "epoch": 39.5748502994012, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1229, "step": 26436 }, { "epoch": 39.57634730538922, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1256, "step": 26437 }, { "epoch": 39.57784431137725, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1265, "step": 26438 }, { "epoch": 39.57934131736527, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1254, "step": 26439 }, { "epoch": 39.58083832335329, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1216, "step": 26440 }, { "epoch": 39.58233532934132, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1318, "step": 26441 }, { "epoch": 39.58383233532934, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.121, "step": 26442 }, { "epoch": 39.58532934131736, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1232, "step": 26443 }, { "epoch": 39.58682634730539, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1174, "step": 26444 }, { "epoch": 39.58832335329341, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1302, "step": 26445 }, { "epoch": 39.58982035928144, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.131, "step": 26446 }, { "epoch": 39.59131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1236, "step": 26447 }, { "epoch": 39.59281437125748, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1269, "step": 26448 }, { "epoch": 39.59431137724551, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1165, "step": 26449 }, { "epoch": 39.59580838323353, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1258, "step": 26450 }, { "epoch": 39.59730538922156, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1273, "step": 26451 }, { "epoch": 39.59880239520958, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1238, "step": 26452 }, { "epoch": 39.6002994011976, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1211, "step": 26453 }, { "epoch": 39.60179640718563, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1268, "step": 26454 }, { "epoch": 39.60329341317365, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1177, "step": 26455 }, { "epoch": 39.604790419161674, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1287, "step": 26456 }, { "epoch": 39.6062874251497, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.123, "step": 26457 }, { "epoch": 39.60778443113772, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1306, "step": 26458 }, { "epoch": 39.60928143712575, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1223, "step": 26459 }, { "epoch": 39.61077844311377, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1268, "step": 26460 }, { "epoch": 39.612275449101794, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1197, "step": 26461 }, { "epoch": 39.61377245508982, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1213, "step": 26462 }, { "epoch": 39.615269461077844, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1254, "step": 26463 }, { "epoch": 39.616766467065865, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1211, "step": 26464 }, { "epoch": 39.618263473053894, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.122, "step": 26465 }, { "epoch": 39.619760479041915, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1256, "step": 26466 }, { "epoch": 39.62125748502994, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1245, "step": 26467 }, { "epoch": 39.622754491017965, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1206, "step": 26468 }, { "epoch": 39.624251497005986, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1198, "step": 26469 }, { "epoch": 39.625748502994014, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.127, "step": 26470 }, { "epoch": 39.627245508982035, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1265, "step": 26471 }, { "epoch": 39.62874251497006, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.125, "step": 26472 }, { "epoch": 39.630239520958085, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1322, "step": 26473 }, { "epoch": 39.631736526946106, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1193, "step": 26474 }, { "epoch": 39.633233532934135, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1272, "step": 26475 }, { "epoch": 39.634730538922156, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1289, "step": 26476 }, { "epoch": 39.63622754491018, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1254, "step": 26477 }, { "epoch": 39.637724550898206, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1243, "step": 26478 }, { "epoch": 39.63922155688623, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 26479 }, { "epoch": 39.64071856287425, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1231, "step": 26480 }, { "epoch": 39.64221556886228, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1238, "step": 26481 }, { "epoch": 39.6437125748503, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1198, "step": 26482 }, { "epoch": 39.645209580838326, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1255, "step": 26483 }, { "epoch": 39.64670658682635, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1234, "step": 26484 }, { "epoch": 39.64820359281437, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1264, "step": 26485 }, { "epoch": 39.6497005988024, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1248, "step": 26486 }, { "epoch": 39.65119760479042, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1264, "step": 26487 }, { "epoch": 39.65269461077844, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1208, "step": 26488 }, { "epoch": 39.65419161676647, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1231, "step": 26489 }, { "epoch": 39.65568862275449, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1318, "step": 26490 }, { "epoch": 39.65718562874252, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1209, "step": 26491 }, { "epoch": 39.65868263473054, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1204, "step": 26492 }, { "epoch": 39.66017964071856, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1256, "step": 26493 }, { "epoch": 39.66167664670659, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1205, "step": 26494 }, { "epoch": 39.66317365269461, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1198, "step": 26495 }, { "epoch": 39.66467065868264, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1261, "step": 26496 }, { "epoch": 39.66616766467066, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1266, "step": 26497 }, { "epoch": 39.66766467065868, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1226, "step": 26498 }, { "epoch": 39.66916167664671, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1256, "step": 26499 }, { "epoch": 39.67065868263473, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.129, "step": 26500 }, { "epoch": 39.67215568862275, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1235, "step": 26501 }, { "epoch": 39.67365269461078, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.123, "step": 26502 }, { "epoch": 39.6751497005988, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.124, "step": 26503 }, { "epoch": 39.67664670658683, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1222, "step": 26504 }, { "epoch": 39.67814371257485, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1226, "step": 26505 }, { "epoch": 39.67964071856287, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1188, "step": 26506 }, { "epoch": 39.6811377245509, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1242, "step": 26507 }, { "epoch": 39.68263473053892, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1254, "step": 26508 }, { "epoch": 39.68413173652694, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1168, "step": 26509 }, { "epoch": 39.68562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1202, "step": 26510 }, { "epoch": 39.68712574850299, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1274, "step": 26511 }, { "epoch": 39.68862275449102, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1261, "step": 26512 }, { "epoch": 39.69011976047904, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1322, "step": 26513 }, { "epoch": 39.691616766467064, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1233, "step": 26514 }, { "epoch": 39.69311377245509, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1199, "step": 26515 }, { "epoch": 39.69461077844311, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.132, "step": 26516 }, { "epoch": 39.696107784431135, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1262, "step": 26517 }, { "epoch": 39.69760479041916, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1239, "step": 26518 }, { "epoch": 39.699101796407184, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.126, "step": 26519 }, { "epoch": 39.70059880239521, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1205, "step": 26520 }, { "epoch": 39.702095808383234, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1247, "step": 26521 }, { "epoch": 39.703592814371255, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1292, "step": 26522 }, { "epoch": 39.705089820359284, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1262, "step": 26523 }, { "epoch": 39.706586826347305, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.126, "step": 26524 }, { "epoch": 39.708083832335326, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1226, "step": 26525 }, { "epoch": 39.709580838323355, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.129, "step": 26526 }, { "epoch": 39.711077844311376, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1271, "step": 26527 }, { "epoch": 39.712574850299404, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1275, "step": 26528 }, { "epoch": 39.714071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.122, "step": 26529 }, { "epoch": 39.71556886227545, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1212, "step": 26530 }, { "epoch": 39.717065868263475, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1209, "step": 26531 }, { "epoch": 39.7185628742515, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1284, "step": 26532 }, { "epoch": 39.72005988023952, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1276, "step": 26533 }, { "epoch": 39.721556886227546, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1296, "step": 26534 }, { "epoch": 39.72305389221557, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1253, "step": 26535 }, { "epoch": 39.724550898203596, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1226, "step": 26536 }, { "epoch": 39.72604790419162, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1163, "step": 26537 }, { "epoch": 39.72754491017964, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1253, "step": 26538 }, { "epoch": 39.72904191616767, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.127, "step": 26539 }, { "epoch": 39.73053892215569, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1268, "step": 26540 }, { "epoch": 39.73203592814371, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1329, "step": 26541 }, { "epoch": 39.73353293413174, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1242, "step": 26542 }, { "epoch": 39.73502994011976, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1251, "step": 26543 }, { "epoch": 39.73652694610779, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1196, "step": 26544 }, { "epoch": 39.73802395209581, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 26545 }, { "epoch": 39.73952095808383, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.137, "step": 26546 }, { "epoch": 39.74101796407186, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1203, "step": 26547 }, { "epoch": 39.74251497005988, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1253, "step": 26548 }, { "epoch": 39.7440119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1277, "step": 26549 }, { "epoch": 39.74550898203593, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1242, "step": 26550 }, { "epoch": 39.74700598802395, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1219, "step": 26551 }, { "epoch": 39.74850299401198, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1271, "step": 26552 }, { "epoch": 39.75, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1314, "step": 26553 }, { "epoch": 39.75149700598802, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1255, "step": 26554 }, { "epoch": 39.75299401197605, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.133, "step": 26555 }, { "epoch": 39.75449101796407, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1237, "step": 26556 }, { "epoch": 39.7559880239521, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1235, "step": 26557 }, { "epoch": 39.75748502994012, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1178, "step": 26558 }, { "epoch": 39.75898203592814, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1206, "step": 26559 }, { "epoch": 39.76047904191617, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1242, "step": 26560 }, { "epoch": 39.76197604790419, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1256, "step": 26561 }, { "epoch": 39.76347305389221, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.121, "step": 26562 }, { "epoch": 39.76497005988024, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1265, "step": 26563 }, { "epoch": 39.76646706586826, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1229, "step": 26564 }, { "epoch": 39.76796407185629, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1221, "step": 26565 }, { "epoch": 39.76946107784431, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1206, "step": 26566 }, { "epoch": 39.77095808383233, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1226, "step": 26567 }, { "epoch": 39.77245508982036, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1235, "step": 26568 }, { "epoch": 39.77395209580838, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.122, "step": 26569 }, { "epoch": 39.775449101796404, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1262, "step": 26570 }, { "epoch": 39.77694610778443, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1274, "step": 26571 }, { "epoch": 39.778443113772454, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1271, "step": 26572 }, { "epoch": 39.77994011976048, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1222, "step": 26573 }, { "epoch": 39.7814371257485, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1212, "step": 26574 }, { "epoch": 39.782934131736525, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1253, "step": 26575 }, { "epoch": 39.78443113772455, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1309, "step": 26576 }, { "epoch": 39.785928143712574, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1284, "step": 26577 }, { "epoch": 39.787425149700596, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1287, "step": 26578 }, { "epoch": 39.788922155688624, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1234, "step": 26579 }, { "epoch": 39.790419161676645, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.124, "step": 26580 }, { "epoch": 39.791916167664674, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1261, "step": 26581 }, { "epoch": 39.793413173652695, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1274, "step": 26582 }, { "epoch": 39.794910179640716, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1198, "step": 26583 }, { "epoch": 39.796407185628745, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1259, "step": 26584 }, { "epoch": 39.797904191616766, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.12, "step": 26585 }, { "epoch": 39.79940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1301, "step": 26586 }, { "epoch": 39.800898203592816, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1214, "step": 26587 }, { "epoch": 39.80239520958084, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.127, "step": 26588 }, { "epoch": 39.803892215568865, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1264, "step": 26589 }, { "epoch": 39.80538922155689, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1298, "step": 26590 }, { "epoch": 39.80688622754491, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1205, "step": 26591 }, { "epoch": 39.808383233532936, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1203, "step": 26592 }, { "epoch": 39.80988023952096, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1271, "step": 26593 }, { "epoch": 39.81137724550898, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1271, "step": 26594 }, { "epoch": 39.81287425149701, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1257, "step": 26595 }, { "epoch": 39.81437125748503, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1296, "step": 26596 }, { "epoch": 39.81586826347306, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1265, "step": 26597 }, { "epoch": 39.81736526946108, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1285, "step": 26598 }, { "epoch": 39.8188622754491, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1243, "step": 26599 }, { "epoch": 39.82035928143713, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 26600 }, { "epoch": 39.82185628742515, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1213, "step": 26601 }, { "epoch": 39.82335329341317, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1245, "step": 26602 }, { "epoch": 39.8248502994012, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1259, "step": 26603 }, { "epoch": 39.82634730538922, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1231, "step": 26604 }, { "epoch": 39.82784431137725, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1175, "step": 26605 }, { "epoch": 39.82934131736527, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1287, "step": 26606 }, { "epoch": 39.83083832335329, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1231, "step": 26607 }, { "epoch": 39.83233532934132, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1211, "step": 26608 }, { "epoch": 39.83383233532934, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1267, "step": 26609 }, { "epoch": 39.83532934131736, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 26610 }, { "epoch": 39.83682634730539, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1199, "step": 26611 }, { "epoch": 39.83832335329341, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1298, "step": 26612 }, { "epoch": 39.83982035928144, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1285, "step": 26613 }, { "epoch": 39.84131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 26614 }, { "epoch": 39.84281437125748, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1281, "step": 26615 }, { "epoch": 39.84431137724551, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1246, "step": 26616 }, { "epoch": 39.84580838323353, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1267, "step": 26617 }, { "epoch": 39.84730538922156, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1232, "step": 26618 }, { "epoch": 39.84880239520958, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1242, "step": 26619 }, { "epoch": 39.8502994011976, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1268, "step": 26620 }, { "epoch": 39.85179640718563, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1227, "step": 26621 }, { "epoch": 39.85329341317365, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1219, "step": 26622 }, { "epoch": 39.854790419161674, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1252, "step": 26623 }, { "epoch": 39.8562874251497, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1239, "step": 26624 }, { "epoch": 39.85778443113772, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1256, "step": 26625 }, { "epoch": 39.85928143712575, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1272, "step": 26626 }, { "epoch": 39.86077844311377, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1252, "step": 26627 }, { "epoch": 39.862275449101794, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1221, "step": 26628 }, { "epoch": 39.86377245508982, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1325, "step": 26629 }, { "epoch": 39.865269461077844, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1256, "step": 26630 }, { "epoch": 39.866766467065865, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1231, "step": 26631 }, { "epoch": 39.868263473053894, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1292, "step": 26632 }, { "epoch": 39.869760479041915, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1237, "step": 26633 }, { "epoch": 39.87125748502994, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1204, "step": 26634 }, { "epoch": 39.872754491017965, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.1207, "step": 26635 }, { "epoch": 39.874251497005986, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1309, "step": 26636 }, { "epoch": 39.875748502994014, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1233, "step": 26637 }, { "epoch": 39.877245508982035, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1226, "step": 26638 }, { "epoch": 39.87874251497006, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1241, "step": 26639 }, { "epoch": 39.880239520958085, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1223, "step": 26640 }, { "epoch": 39.881736526946106, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1186, "step": 26641 }, { "epoch": 39.883233532934135, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1282, "step": 26642 }, { "epoch": 39.884730538922156, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1272, "step": 26643 }, { "epoch": 39.88622754491018, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1168, "step": 26644 }, { "epoch": 39.887724550898206, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.123, "step": 26645 }, { "epoch": 39.88922155688623, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1208, "step": 26646 }, { "epoch": 39.89071856287425, "grad_norm": 0.050537109375, "learning_rate": 0.0008, "loss": 1.1252, "step": 26647 }, { "epoch": 39.89221556886228, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1228, "step": 26648 }, { "epoch": 39.8937125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.126, "step": 26649 }, { "epoch": 39.895209580838326, "grad_norm": 0.05029296875, "learning_rate": 0.0008, "loss": 1.119, "step": 26650 }, { "epoch": 39.89670658682635, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1217, "step": 26651 }, { "epoch": 39.89820359281437, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.126, "step": 26652 }, { "epoch": 39.8997005988024, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1247, "step": 26653 }, { "epoch": 39.90119760479042, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1289, "step": 26654 }, { "epoch": 39.90269461077844, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1317, "step": 26655 }, { "epoch": 39.90419161676647, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1262, "step": 26656 }, { "epoch": 39.90568862275449, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1258, "step": 26657 }, { "epoch": 39.90718562874252, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1255, "step": 26658 }, { "epoch": 39.90868263473054, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.123, "step": 26659 }, { "epoch": 39.91017964071856, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 26660 }, { "epoch": 39.91167664670659, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1235, "step": 26661 }, { "epoch": 39.91317365269461, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1259, "step": 26662 }, { "epoch": 39.91467065868264, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.124, "step": 26663 }, { "epoch": 39.91616766467066, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1261, "step": 26664 }, { "epoch": 39.91766467065868, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1275, "step": 26665 }, { "epoch": 39.91916167664671, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.13, "step": 26666 }, { "epoch": 39.92065868263473, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1283, "step": 26667 }, { "epoch": 39.92215568862275, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1219, "step": 26668 }, { "epoch": 39.92365269461078, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1258, "step": 26669 }, { "epoch": 39.9251497005988, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 26670 }, { "epoch": 39.92664670658683, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1354, "step": 26671 }, { "epoch": 39.92814371257485, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.125, "step": 26672 }, { "epoch": 39.92964071856287, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1255, "step": 26673 }, { "epoch": 39.9311377245509, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1223, "step": 26674 }, { "epoch": 39.93263473053892, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1242, "step": 26675 }, { "epoch": 39.93413173652694, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1273, "step": 26676 }, { "epoch": 39.93562874251497, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1257, "step": 26677 }, { "epoch": 39.93712574850299, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1198, "step": 26678 }, { "epoch": 39.93862275449102, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 26679 }, { "epoch": 39.94011976047904, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1199, "step": 26680 }, { "epoch": 39.941616766467064, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1254, "step": 26681 }, { "epoch": 39.94311377245509, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.122, "step": 26682 }, { "epoch": 39.94461077844311, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1301, "step": 26683 }, { "epoch": 39.946107784431135, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1241, "step": 26684 }, { "epoch": 39.94760479041916, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1228, "step": 26685 }, { "epoch": 39.949101796407184, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1228, "step": 26686 }, { "epoch": 39.95059880239521, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1231, "step": 26687 }, { "epoch": 39.952095808383234, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1286, "step": 26688 }, { "epoch": 39.953592814371255, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1177, "step": 26689 }, { "epoch": 39.955089820359284, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1249, "step": 26690 }, { "epoch": 39.956586826347305, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1252, "step": 26691 }, { "epoch": 39.958083832335326, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1225, "step": 26692 }, { "epoch": 39.959580838323355, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1318, "step": 26693 }, { "epoch": 39.961077844311376, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1269, "step": 26694 }, { "epoch": 39.962574850299404, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1252, "step": 26695 }, { "epoch": 39.964071856287426, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1227, "step": 26696 }, { "epoch": 39.96556886227545, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1278, "step": 26697 }, { "epoch": 39.967065868263475, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.127, "step": 26698 }, { "epoch": 39.9685628742515, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1196, "step": 26699 }, { "epoch": 39.97005988023952, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1239, "step": 26700 }, { "epoch": 39.971556886227546, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1257, "step": 26701 }, { "epoch": 39.97305389221557, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1245, "step": 26702 }, { "epoch": 39.974550898203596, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1229, "step": 26703 }, { "epoch": 39.97604790419162, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1258, "step": 26704 }, { "epoch": 39.97754491017964, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1197, "step": 26705 }, { "epoch": 39.97904191616767, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 26706 }, { "epoch": 39.98053892215569, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1314, "step": 26707 }, { "epoch": 39.98203592814371, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1158, "step": 26708 }, { "epoch": 39.98353293413174, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1241, "step": 26709 }, { "epoch": 39.98502994011976, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1201, "step": 26710 }, { "epoch": 39.98652694610779, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1198, "step": 26711 }, { "epoch": 39.98802395209581, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1247, "step": 26712 }, { "epoch": 39.98952095808383, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1245, "step": 26713 }, { "epoch": 39.99101796407186, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1253, "step": 26714 }, { "epoch": 39.99251497005988, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1206, "step": 26715 }, { "epoch": 39.9940119760479, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1211, "step": 26716 }, { "epoch": 39.99550898203593, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1323, "step": 26717 }, { "epoch": 39.99700598802395, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.123, "step": 26718 }, { "epoch": 39.99850299401198, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1229, "step": 26719 }, { "epoch": 40.0, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1191, "step": 26720 }, { "epoch": 40.00149700598802, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1229, "step": 26721 }, { "epoch": 40.00299401197605, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1236, "step": 26722 }, { "epoch": 40.00449101796407, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1205, "step": 26723 }, { "epoch": 40.0059880239521, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1201, "step": 26724 }, { "epoch": 40.00748502994012, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1242, "step": 26725 }, { "epoch": 40.00898203592814, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1205, "step": 26726 }, { "epoch": 40.01047904191617, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1246, "step": 26727 }, { "epoch": 40.01197604790419, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.127, "step": 26728 }, { "epoch": 40.01347305389221, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1249, "step": 26729 }, { "epoch": 40.01497005988024, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1254, "step": 26730 }, { "epoch": 40.01646706586826, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1222, "step": 26731 }, { "epoch": 40.01796407185629, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1209, "step": 26732 }, { "epoch": 40.01946107784431, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1228, "step": 26733 }, { "epoch": 40.02095808383233, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1252, "step": 26734 }, { "epoch": 40.02245508982036, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1329, "step": 26735 }, { "epoch": 40.02395209580838, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1254, "step": 26736 }, { "epoch": 40.025449101796404, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1201, "step": 26737 }, { "epoch": 40.02694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1225, "step": 26738 }, { "epoch": 40.028443113772454, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1234, "step": 26739 }, { "epoch": 40.02994011976048, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1295, "step": 26740 }, { "epoch": 40.0314371257485, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 26741 }, { "epoch": 40.032934131736525, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1231, "step": 26742 }, { "epoch": 40.03443113772455, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.113, "step": 26743 }, { "epoch": 40.035928143712574, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 26744 }, { "epoch": 40.037425149700596, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1183, "step": 26745 }, { "epoch": 40.038922155688624, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 26746 }, { "epoch": 40.040419161676645, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1233, "step": 26747 }, { "epoch": 40.041916167664674, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1225, "step": 26748 }, { "epoch": 40.043413173652695, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1271, "step": 26749 }, { "epoch": 40.044910179640716, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1261, "step": 26750 }, { "epoch": 40.046407185628745, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1305, "step": 26751 }, { "epoch": 40.047904191616766, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1219, "step": 26752 }, { "epoch": 40.04940119760479, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1218, "step": 26753 }, { "epoch": 40.050898203592816, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.128, "step": 26754 }, { "epoch": 40.05239520958084, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1214, "step": 26755 }, { "epoch": 40.053892215568865, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1335, "step": 26756 }, { "epoch": 40.05538922155689, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1212, "step": 26757 }, { "epoch": 40.05688622754491, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1259, "step": 26758 }, { "epoch": 40.058383233532936, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1232, "step": 26759 }, { "epoch": 40.05988023952096, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1257, "step": 26760 }, { "epoch": 40.06137724550898, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1209, "step": 26761 }, { "epoch": 40.06287425149701, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1161, "step": 26762 }, { "epoch": 40.06437125748503, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1321, "step": 26763 }, { "epoch": 40.06586826347306, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1135, "step": 26764 }, { "epoch": 40.06736526946108, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1211, "step": 26765 }, { "epoch": 40.0688622754491, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1258, "step": 26766 }, { "epoch": 40.07035928143713, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1304, "step": 26767 }, { "epoch": 40.07185628742515, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1305, "step": 26768 }, { "epoch": 40.07335329341317, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1285, "step": 26769 }, { "epoch": 40.0748502994012, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1169, "step": 26770 }, { "epoch": 40.07634730538922, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1275, "step": 26771 }, { "epoch": 40.07784431137725, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1253, "step": 26772 }, { "epoch": 40.07934131736527, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1265, "step": 26773 }, { "epoch": 40.08083832335329, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1206, "step": 26774 }, { "epoch": 40.08233532934132, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1257, "step": 26775 }, { "epoch": 40.08383233532934, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.128, "step": 26776 }, { "epoch": 40.08532934131737, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1286, "step": 26777 }, { "epoch": 40.08682634730539, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.122, "step": 26778 }, { "epoch": 40.08832335329341, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1206, "step": 26779 }, { "epoch": 40.08982035928144, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1286, "step": 26780 }, { "epoch": 40.09131736526946, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1251, "step": 26781 }, { "epoch": 40.09281437125748, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1222, "step": 26782 }, { "epoch": 40.09431137724551, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1233, "step": 26783 }, { "epoch": 40.09580838323353, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1255, "step": 26784 }, { "epoch": 40.09730538922156, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1289, "step": 26785 }, { "epoch": 40.09880239520958, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.12, "step": 26786 }, { "epoch": 40.1002994011976, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1261, "step": 26787 }, { "epoch": 40.10179640718563, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1242, "step": 26788 }, { "epoch": 40.10329341317365, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1226, "step": 26789 }, { "epoch": 40.104790419161674, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1232, "step": 26790 }, { "epoch": 40.1062874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.12, "step": 26791 }, { "epoch": 40.10778443113772, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1278, "step": 26792 }, { "epoch": 40.10928143712575, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1216, "step": 26793 }, { "epoch": 40.11077844311377, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1252, "step": 26794 }, { "epoch": 40.112275449101794, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1239, "step": 26795 }, { "epoch": 40.11377245508982, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1219, "step": 26796 }, { "epoch": 40.115269461077844, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1271, "step": 26797 }, { "epoch": 40.116766467065865, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1235, "step": 26798 }, { "epoch": 40.118263473053894, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1312, "step": 26799 }, { "epoch": 40.119760479041915, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.126, "step": 26800 }, { "epoch": 40.12125748502994, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1231, "step": 26801 }, { "epoch": 40.122754491017965, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1249, "step": 26802 }, { "epoch": 40.124251497005986, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1276, "step": 26803 }, { "epoch": 40.125748502994014, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1284, "step": 26804 }, { "epoch": 40.127245508982035, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1253, "step": 26805 }, { "epoch": 40.12874251497006, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.121, "step": 26806 }, { "epoch": 40.130239520958085, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1277, "step": 26807 }, { "epoch": 40.131736526946106, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.124, "step": 26808 }, { "epoch": 40.133233532934135, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1272, "step": 26809 }, { "epoch": 40.134730538922156, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1242, "step": 26810 }, { "epoch": 40.13622754491018, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.123, "step": 26811 }, { "epoch": 40.137724550898206, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1301, "step": 26812 }, { "epoch": 40.13922155688623, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1239, "step": 26813 }, { "epoch": 40.14071856287425, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1199, "step": 26814 }, { "epoch": 40.14221556886228, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.13, "step": 26815 }, { "epoch": 40.1437125748503, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1215, "step": 26816 }, { "epoch": 40.145209580838326, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1234, "step": 26817 }, { "epoch": 40.14670658682635, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1261, "step": 26818 }, { "epoch": 40.14820359281437, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1283, "step": 26819 }, { "epoch": 40.1497005988024, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1297, "step": 26820 }, { "epoch": 40.15119760479042, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1286, "step": 26821 }, { "epoch": 40.15269461077844, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1246, "step": 26822 }, { "epoch": 40.15419161676647, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1288, "step": 26823 }, { "epoch": 40.15568862275449, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1211, "step": 26824 }, { "epoch": 40.15718562874252, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1206, "step": 26825 }, { "epoch": 40.15868263473054, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1173, "step": 26826 }, { "epoch": 40.16017964071856, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1229, "step": 26827 }, { "epoch": 40.16167664670659, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1192, "step": 26828 }, { "epoch": 40.16317365269461, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1253, "step": 26829 }, { "epoch": 40.16467065868263, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1321, "step": 26830 }, { "epoch": 40.16616766467066, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1195, "step": 26831 }, { "epoch": 40.16766467065868, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1204, "step": 26832 }, { "epoch": 40.16916167664671, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1204, "step": 26833 }, { "epoch": 40.17065868263473, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1244, "step": 26834 }, { "epoch": 40.17215568862275, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1245, "step": 26835 }, { "epoch": 40.17365269461078, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1182, "step": 26836 }, { "epoch": 40.1751497005988, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.127, "step": 26837 }, { "epoch": 40.17664670658683, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1217, "step": 26838 }, { "epoch": 40.17814371257485, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1221, "step": 26839 }, { "epoch": 40.17964071856287, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.129, "step": 26840 }, { "epoch": 40.1811377245509, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1231, "step": 26841 }, { "epoch": 40.18263473053892, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1218, "step": 26842 }, { "epoch": 40.18413173652694, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.118, "step": 26843 }, { "epoch": 40.18562874251497, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.124, "step": 26844 }, { "epoch": 40.18712574850299, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1205, "step": 26845 }, { "epoch": 40.18862275449102, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.123, "step": 26846 }, { "epoch": 40.19011976047904, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1213, "step": 26847 }, { "epoch": 40.191616766467064, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1214, "step": 26848 }, { "epoch": 40.19311377245509, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.124, "step": 26849 }, { "epoch": 40.19461077844311, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1189, "step": 26850 }, { "epoch": 40.196107784431135, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.13, "step": 26851 }, { "epoch": 40.19760479041916, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1181, "step": 26852 }, { "epoch": 40.199101796407184, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1268, "step": 26853 }, { "epoch": 40.20059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1251, "step": 26854 }, { "epoch": 40.202095808383234, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1244, "step": 26855 }, { "epoch": 40.203592814371255, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.128, "step": 26856 }, { "epoch": 40.205089820359284, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1179, "step": 26857 }, { "epoch": 40.206586826347305, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1182, "step": 26858 }, { "epoch": 40.208083832335326, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1185, "step": 26859 }, { "epoch": 40.209580838323355, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.127, "step": 26860 }, { "epoch": 40.211077844311376, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1187, "step": 26861 }, { "epoch": 40.212574850299404, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 26862 }, { "epoch": 40.214071856287426, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1275, "step": 26863 }, { "epoch": 40.21556886227545, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1247, "step": 26864 }, { "epoch": 40.217065868263475, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.122, "step": 26865 }, { "epoch": 40.2185628742515, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 26866 }, { "epoch": 40.22005988023952, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1255, "step": 26867 }, { "epoch": 40.221556886227546, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1224, "step": 26868 }, { "epoch": 40.22305389221557, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.129, "step": 26869 }, { "epoch": 40.224550898203596, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1261, "step": 26870 }, { "epoch": 40.22604790419162, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1213, "step": 26871 }, { "epoch": 40.22754491017964, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1187, "step": 26872 }, { "epoch": 40.22904191616767, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1218, "step": 26873 }, { "epoch": 40.23053892215569, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1272, "step": 26874 }, { "epoch": 40.23203592814371, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1157, "step": 26875 }, { "epoch": 40.23353293413174, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1196, "step": 26876 }, { "epoch": 40.23502994011976, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1203, "step": 26877 }, { "epoch": 40.23652694610779, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1284, "step": 26878 }, { "epoch": 40.23802395209581, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1245, "step": 26879 }, { "epoch": 40.23952095808383, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1281, "step": 26880 }, { "epoch": 40.24101796407186, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1277, "step": 26881 }, { "epoch": 40.24251497005988, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1146, "step": 26882 }, { "epoch": 40.2440119760479, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1232, "step": 26883 }, { "epoch": 40.24550898203593, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1245, "step": 26884 }, { "epoch": 40.24700598802395, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.12, "step": 26885 }, { "epoch": 40.24850299401198, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.125, "step": 26886 }, { "epoch": 40.25, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1202, "step": 26887 }, { "epoch": 40.25149700598802, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1228, "step": 26888 }, { "epoch": 40.25299401197605, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1309, "step": 26889 }, { "epoch": 40.25449101796407, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1241, "step": 26890 }, { "epoch": 40.2559880239521, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1216, "step": 26891 }, { "epoch": 40.25748502994012, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1243, "step": 26892 }, { "epoch": 40.25898203592814, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1255, "step": 26893 }, { "epoch": 40.26047904191617, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1275, "step": 26894 }, { "epoch": 40.26197604790419, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1225, "step": 26895 }, { "epoch": 40.26347305389221, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1247, "step": 26896 }, { "epoch": 40.26497005988024, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1233, "step": 26897 }, { "epoch": 40.26646706586826, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.125, "step": 26898 }, { "epoch": 40.26796407185629, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1239, "step": 26899 }, { "epoch": 40.26946107784431, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1165, "step": 26900 }, { "epoch": 40.27095808383233, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1268, "step": 26901 }, { "epoch": 40.27245508982036, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1226, "step": 26902 }, { "epoch": 40.27395209580838, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.122, "step": 26903 }, { "epoch": 40.275449101796404, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1246, "step": 26904 }, { "epoch": 40.27694610778443, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1262, "step": 26905 }, { "epoch": 40.278443113772454, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1267, "step": 26906 }, { "epoch": 40.27994011976048, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1226, "step": 26907 }, { "epoch": 40.2814371257485, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1228, "step": 26908 }, { "epoch": 40.282934131736525, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1221, "step": 26909 }, { "epoch": 40.28443113772455, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1213, "step": 26910 }, { "epoch": 40.285928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1232, "step": 26911 }, { "epoch": 40.287425149700596, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1208, "step": 26912 }, { "epoch": 40.288922155688624, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1214, "step": 26913 }, { "epoch": 40.290419161676645, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1195, "step": 26914 }, { "epoch": 40.291916167664674, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1239, "step": 26915 }, { "epoch": 40.293413173652695, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1314, "step": 26916 }, { "epoch": 40.294910179640716, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1282, "step": 26917 }, { "epoch": 40.296407185628745, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1237, "step": 26918 }, { "epoch": 40.297904191616766, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1199, "step": 26919 }, { "epoch": 40.29940119760479, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1245, "step": 26920 }, { "epoch": 40.300898203592816, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1215, "step": 26921 }, { "epoch": 40.30239520958084, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1239, "step": 26922 }, { "epoch": 40.303892215568865, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.125, "step": 26923 }, { "epoch": 40.30538922155689, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1234, "step": 26924 }, { "epoch": 40.30688622754491, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1243, "step": 26925 }, { "epoch": 40.308383233532936, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.12, "step": 26926 }, { "epoch": 40.30988023952096, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1261, "step": 26927 }, { "epoch": 40.31137724550898, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1216, "step": 26928 }, { "epoch": 40.31287425149701, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1262, "step": 26929 }, { "epoch": 40.31437125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 26930 }, { "epoch": 40.31586826347306, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1285, "step": 26931 }, { "epoch": 40.31736526946108, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.125, "step": 26932 }, { "epoch": 40.3188622754491, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1205, "step": 26933 }, { "epoch": 40.32035928143713, "grad_norm": 0.0498046875, "learning_rate": 0.0008, "loss": 1.1302, "step": 26934 }, { "epoch": 40.32185628742515, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1244, "step": 26935 }, { "epoch": 40.32335329341317, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1239, "step": 26936 }, { "epoch": 40.3248502994012, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1179, "step": 26937 }, { "epoch": 40.32634730538922, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1229, "step": 26938 }, { "epoch": 40.32784431137725, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.124, "step": 26939 }, { "epoch": 40.32934131736527, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1239, "step": 26940 }, { "epoch": 40.33083832335329, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1271, "step": 26941 }, { "epoch": 40.33233532934132, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1256, "step": 26942 }, { "epoch": 40.33383233532934, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1255, "step": 26943 }, { "epoch": 40.33532934131736, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1213, "step": 26944 }, { "epoch": 40.33682634730539, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.122, "step": 26945 }, { "epoch": 40.33832335329341, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1257, "step": 26946 }, { "epoch": 40.33982035928144, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1242, "step": 26947 }, { "epoch": 40.34131736526946, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1253, "step": 26948 }, { "epoch": 40.34281437125748, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1223, "step": 26949 }, { "epoch": 40.34431137724551, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1248, "step": 26950 }, { "epoch": 40.34580838323353, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1173, "step": 26951 }, { "epoch": 40.34730538922156, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1261, "step": 26952 }, { "epoch": 40.34880239520958, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1263, "step": 26953 }, { "epoch": 40.3502994011976, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1214, "step": 26954 }, { "epoch": 40.35179640718563, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1221, "step": 26955 }, { "epoch": 40.35329341317365, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1201, "step": 26956 }, { "epoch": 40.354790419161674, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1255, "step": 26957 }, { "epoch": 40.3562874251497, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1327, "step": 26958 }, { "epoch": 40.35778443113772, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1241, "step": 26959 }, { "epoch": 40.35928143712575, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1272, "step": 26960 }, { "epoch": 40.36077844311377, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1276, "step": 26961 }, { "epoch": 40.362275449101794, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1245, "step": 26962 }, { "epoch": 40.36377245508982, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1266, "step": 26963 }, { "epoch": 40.365269461077844, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 26964 }, { "epoch": 40.366766467065865, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.125, "step": 26965 }, { "epoch": 40.368263473053894, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1224, "step": 26966 }, { "epoch": 40.369760479041915, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1308, "step": 26967 }, { "epoch": 40.37125748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.117, "step": 26968 }, { "epoch": 40.372754491017965, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1295, "step": 26969 }, { "epoch": 40.374251497005986, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1202, "step": 26970 }, { "epoch": 40.375748502994014, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1191, "step": 26971 }, { "epoch": 40.377245508982035, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1245, "step": 26972 }, { "epoch": 40.37874251497006, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1204, "step": 26973 }, { "epoch": 40.380239520958085, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1286, "step": 26974 }, { "epoch": 40.381736526946106, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1193, "step": 26975 }, { "epoch": 40.383233532934135, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1194, "step": 26976 }, { "epoch": 40.384730538922156, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1235, "step": 26977 }, { "epoch": 40.38622754491018, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1267, "step": 26978 }, { "epoch": 40.387724550898206, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1327, "step": 26979 }, { "epoch": 40.38922155688623, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1223, "step": 26980 }, { "epoch": 40.39071856287425, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1231, "step": 26981 }, { "epoch": 40.39221556886228, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1224, "step": 26982 }, { "epoch": 40.3937125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1222, "step": 26983 }, { "epoch": 40.395209580838326, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1273, "step": 26984 }, { "epoch": 40.39670658682635, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1285, "step": 26985 }, { "epoch": 40.39820359281437, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1232, "step": 26986 }, { "epoch": 40.3997005988024, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1248, "step": 26987 }, { "epoch": 40.40119760479042, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1235, "step": 26988 }, { "epoch": 40.40269461077844, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.124, "step": 26989 }, { "epoch": 40.40419161676647, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1189, "step": 26990 }, { "epoch": 40.40568862275449, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1193, "step": 26991 }, { "epoch": 40.40718562874252, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1292, "step": 26992 }, { "epoch": 40.40868263473054, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1175, "step": 26993 }, { "epoch": 40.41017964071856, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1291, "step": 26994 }, { "epoch": 40.41167664670659, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1156, "step": 26995 }, { "epoch": 40.41317365269461, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1154, "step": 26996 }, { "epoch": 40.41467065868264, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1242, "step": 26997 }, { "epoch": 40.41616766467066, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1225, "step": 26998 }, { "epoch": 40.41766467065868, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1209, "step": 26999 }, { "epoch": 40.41916167664671, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1151, "step": 27000 }, { "epoch": 40.42065868263473, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1192, "step": 27001 }, { "epoch": 40.42215568862275, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1245, "step": 27002 }, { "epoch": 40.42365269461078, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1276, "step": 27003 }, { "epoch": 40.4251497005988, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1198, "step": 27004 }, { "epoch": 40.42664670658683, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1239, "step": 27005 }, { "epoch": 40.42814371257485, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1195, "step": 27006 }, { "epoch": 40.42964071856287, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.119, "step": 27007 }, { "epoch": 40.4311377245509, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1242, "step": 27008 }, { "epoch": 40.43263473053892, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1278, "step": 27009 }, { "epoch": 40.43413173652694, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1352, "step": 27010 }, { "epoch": 40.43562874251497, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1231, "step": 27011 }, { "epoch": 40.43712574850299, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1263, "step": 27012 }, { "epoch": 40.43862275449102, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.137, "step": 27013 }, { "epoch": 40.44011976047904, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.125, "step": 27014 }, { "epoch": 40.441616766467064, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1309, "step": 27015 }, { "epoch": 40.44311377245509, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1279, "step": 27016 }, { "epoch": 40.44461077844311, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1307, "step": 27017 }, { "epoch": 40.446107784431135, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1222, "step": 27018 }, { "epoch": 40.44760479041916, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1233, "step": 27019 }, { "epoch": 40.449101796407184, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.123, "step": 27020 }, { "epoch": 40.45059880239521, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1258, "step": 27021 }, { "epoch": 40.452095808383234, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.124, "step": 27022 }, { "epoch": 40.453592814371255, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1262, "step": 27023 }, { "epoch": 40.455089820359284, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1263, "step": 27024 }, { "epoch": 40.456586826347305, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1216, "step": 27025 }, { "epoch": 40.458083832335326, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1301, "step": 27026 }, { "epoch": 40.459580838323355, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1332, "step": 27027 }, { "epoch": 40.461077844311376, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1191, "step": 27028 }, { "epoch": 40.462574850299404, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1207, "step": 27029 }, { "epoch": 40.464071856287426, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1247, "step": 27030 }, { "epoch": 40.46556886227545, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1199, "step": 27031 }, { "epoch": 40.467065868263475, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1167, "step": 27032 }, { "epoch": 40.4685628742515, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1318, "step": 27033 }, { "epoch": 40.47005988023952, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1268, "step": 27034 }, { "epoch": 40.471556886227546, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1256, "step": 27035 }, { "epoch": 40.47305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1185, "step": 27036 }, { "epoch": 40.474550898203596, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1249, "step": 27037 }, { "epoch": 40.47604790419162, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1195, "step": 27038 }, { "epoch": 40.47754491017964, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1222, "step": 27039 }, { "epoch": 40.47904191616767, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.125, "step": 27040 }, { "epoch": 40.48053892215569, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1291, "step": 27041 }, { "epoch": 40.48203592814371, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1275, "step": 27042 }, { "epoch": 40.48353293413174, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1282, "step": 27043 }, { "epoch": 40.48502994011976, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1303, "step": 27044 }, { "epoch": 40.48652694610779, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1281, "step": 27045 }, { "epoch": 40.48802395209581, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1213, "step": 27046 }, { "epoch": 40.48952095808383, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1222, "step": 27047 }, { "epoch": 40.49101796407186, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1278, "step": 27048 }, { "epoch": 40.49251497005988, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1283, "step": 27049 }, { "epoch": 40.4940119760479, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1163, "step": 27050 }, { "epoch": 40.49550898203593, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1202, "step": 27051 }, { "epoch": 40.49700598802395, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.129, "step": 27052 }, { "epoch": 40.49850299401198, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1206, "step": 27053 }, { "epoch": 40.5, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1229, "step": 27054 }, { "epoch": 40.50149700598802, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1302, "step": 27055 }, { "epoch": 40.50299401197605, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1198, "step": 27056 }, { "epoch": 40.50449101796407, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1267, "step": 27057 }, { "epoch": 40.5059880239521, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1307, "step": 27058 }, { "epoch": 40.50748502994012, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1208, "step": 27059 }, { "epoch": 40.50898203592814, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.119, "step": 27060 }, { "epoch": 40.51047904191617, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1318, "step": 27061 }, { "epoch": 40.51197604790419, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1163, "step": 27062 }, { "epoch": 40.51347305389221, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 27063 }, { "epoch": 40.51497005988024, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1223, "step": 27064 }, { "epoch": 40.51646706586826, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1271, "step": 27065 }, { "epoch": 40.51796407185629, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1187, "step": 27066 }, { "epoch": 40.51946107784431, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1282, "step": 27067 }, { "epoch": 40.52095808383233, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.123, "step": 27068 }, { "epoch": 40.52245508982036, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1234, "step": 27069 }, { "epoch": 40.52395209580838, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1201, "step": 27070 }, { "epoch": 40.525449101796404, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1248, "step": 27071 }, { "epoch": 40.52694610778443, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1247, "step": 27072 }, { "epoch": 40.528443113772454, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1281, "step": 27073 }, { "epoch": 40.52994011976048, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1244, "step": 27074 }, { "epoch": 40.5314371257485, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1301, "step": 27075 }, { "epoch": 40.532934131736525, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1263, "step": 27076 }, { "epoch": 40.53443113772455, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1204, "step": 27077 }, { "epoch": 40.535928143712574, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1263, "step": 27078 }, { "epoch": 40.537425149700596, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1195, "step": 27079 }, { "epoch": 40.538922155688624, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1174, "step": 27080 }, { "epoch": 40.540419161676645, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1205, "step": 27081 }, { "epoch": 40.541916167664674, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1236, "step": 27082 }, { "epoch": 40.543413173652695, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1241, "step": 27083 }, { "epoch": 40.544910179640716, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1262, "step": 27084 }, { "epoch": 40.546407185628745, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1223, "step": 27085 }, { "epoch": 40.547904191616766, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1255, "step": 27086 }, { "epoch": 40.54940119760479, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1181, "step": 27087 }, { "epoch": 40.550898203592816, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1221, "step": 27088 }, { "epoch": 40.55239520958084, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1276, "step": 27089 }, { "epoch": 40.553892215568865, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1202, "step": 27090 }, { "epoch": 40.55538922155689, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1265, "step": 27091 }, { "epoch": 40.55688622754491, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1225, "step": 27092 }, { "epoch": 40.558383233532936, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1262, "step": 27093 }, { "epoch": 40.55988023952096, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1299, "step": 27094 }, { "epoch": 40.56137724550898, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1251, "step": 27095 }, { "epoch": 40.56287425149701, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1239, "step": 27096 }, { "epoch": 40.56437125748503, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1211, "step": 27097 }, { "epoch": 40.56586826347306, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1282, "step": 27098 }, { "epoch": 40.56736526946108, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.124, "step": 27099 }, { "epoch": 40.5688622754491, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1306, "step": 27100 }, { "epoch": 40.57035928143713, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1215, "step": 27101 }, { "epoch": 40.57185628742515, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1232, "step": 27102 }, { "epoch": 40.57335329341317, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.127, "step": 27103 }, { "epoch": 40.5748502994012, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1189, "step": 27104 }, { "epoch": 40.57634730538922, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1267, "step": 27105 }, { "epoch": 40.57784431137725, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.128, "step": 27106 }, { "epoch": 40.57934131736527, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1254, "step": 27107 }, { "epoch": 40.58083832335329, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1254, "step": 27108 }, { "epoch": 40.58233532934132, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1263, "step": 27109 }, { "epoch": 40.58383233532934, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.12, "step": 27110 }, { "epoch": 40.58532934131736, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1216, "step": 27111 }, { "epoch": 40.58682634730539, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1289, "step": 27112 }, { "epoch": 40.58832335329341, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1227, "step": 27113 }, { "epoch": 40.58982035928144, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1187, "step": 27114 }, { "epoch": 40.59131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.13, "step": 27115 }, { "epoch": 40.59281437125748, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1251, "step": 27116 }, { "epoch": 40.59431137724551, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1174, "step": 27117 }, { "epoch": 40.59580838323353, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1223, "step": 27118 }, { "epoch": 40.59730538922156, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1214, "step": 27119 }, { "epoch": 40.59880239520958, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1217, "step": 27120 }, { "epoch": 40.6002994011976, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1263, "step": 27121 }, { "epoch": 40.60179640718563, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1254, "step": 27122 }, { "epoch": 40.60329341317365, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1232, "step": 27123 }, { "epoch": 40.604790419161674, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1311, "step": 27124 }, { "epoch": 40.6062874251497, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.127, "step": 27125 }, { "epoch": 40.60778443113772, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1275, "step": 27126 }, { "epoch": 40.60928143712575, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1278, "step": 27127 }, { "epoch": 40.61077844311377, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1218, "step": 27128 }, { "epoch": 40.612275449101794, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1196, "step": 27129 }, { "epoch": 40.61377245508982, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1286, "step": 27130 }, { "epoch": 40.615269461077844, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1237, "step": 27131 }, { "epoch": 40.616766467065865, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1299, "step": 27132 }, { "epoch": 40.618263473053894, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 27133 }, { "epoch": 40.619760479041915, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1221, "step": 27134 }, { "epoch": 40.62125748502994, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1211, "step": 27135 }, { "epoch": 40.622754491017965, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1278, "step": 27136 }, { "epoch": 40.624251497005986, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1243, "step": 27137 }, { "epoch": 40.625748502994014, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1305, "step": 27138 }, { "epoch": 40.627245508982035, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1222, "step": 27139 }, { "epoch": 40.62874251497006, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 27140 }, { "epoch": 40.630239520958085, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1289, "step": 27141 }, { "epoch": 40.631736526946106, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1235, "step": 27142 }, { "epoch": 40.633233532934135, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1237, "step": 27143 }, { "epoch": 40.634730538922156, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 27144 }, { "epoch": 40.63622754491018, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 27145 }, { "epoch": 40.637724550898206, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1321, "step": 27146 }, { "epoch": 40.63922155688623, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1194, "step": 27147 }, { "epoch": 40.64071856287425, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.122, "step": 27148 }, { "epoch": 40.64221556886228, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1227, "step": 27149 }, { "epoch": 40.6437125748503, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1266, "step": 27150 }, { "epoch": 40.645209580838326, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.121, "step": 27151 }, { "epoch": 40.64670658682635, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1206, "step": 27152 }, { "epoch": 40.64820359281437, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1237, "step": 27153 }, { "epoch": 40.6497005988024, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1234, "step": 27154 }, { "epoch": 40.65119760479042, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1192, "step": 27155 }, { "epoch": 40.65269461077844, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1211, "step": 27156 }, { "epoch": 40.65419161676647, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1238, "step": 27157 }, { "epoch": 40.65568862275449, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1262, "step": 27158 }, { "epoch": 40.65718562874252, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1285, "step": 27159 }, { "epoch": 40.65868263473054, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1223, "step": 27160 }, { "epoch": 40.66017964071856, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1206, "step": 27161 }, { "epoch": 40.66167664670659, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1153, "step": 27162 }, { "epoch": 40.66317365269461, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.121, "step": 27163 }, { "epoch": 40.66467065868264, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1254, "step": 27164 }, { "epoch": 40.66616766467066, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1241, "step": 27165 }, { "epoch": 40.66766467065868, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1214, "step": 27166 }, { "epoch": 40.66916167664671, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1267, "step": 27167 }, { "epoch": 40.67065868263473, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1218, "step": 27168 }, { "epoch": 40.67215568862275, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1294, "step": 27169 }, { "epoch": 40.67365269461078, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.129, "step": 27170 }, { "epoch": 40.6751497005988, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1246, "step": 27171 }, { "epoch": 40.67664670658683, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1255, "step": 27172 }, { "epoch": 40.67814371257485, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1257, "step": 27173 }, { "epoch": 40.67964071856287, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1256, "step": 27174 }, { "epoch": 40.6811377245509, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1178, "step": 27175 }, { "epoch": 40.68263473053892, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1246, "step": 27176 }, { "epoch": 40.68413173652694, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1217, "step": 27177 }, { "epoch": 40.68562874251497, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1272, "step": 27178 }, { "epoch": 40.68712574850299, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1208, "step": 27179 }, { "epoch": 40.68862275449102, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1224, "step": 27180 }, { "epoch": 40.69011976047904, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1147, "step": 27181 }, { "epoch": 40.691616766467064, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1264, "step": 27182 }, { "epoch": 40.69311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1247, "step": 27183 }, { "epoch": 40.69461077844311, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1247, "step": 27184 }, { "epoch": 40.696107784431135, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1306, "step": 27185 }, { "epoch": 40.69760479041916, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1269, "step": 27186 }, { "epoch": 40.699101796407184, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.128, "step": 27187 }, { "epoch": 40.70059880239521, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1211, "step": 27188 }, { "epoch": 40.702095808383234, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1225, "step": 27189 }, { "epoch": 40.703592814371255, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1239, "step": 27190 }, { "epoch": 40.705089820359284, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1266, "step": 27191 }, { "epoch": 40.706586826347305, "grad_norm": 0.050048828125, "learning_rate": 0.0008, "loss": 1.1249, "step": 27192 }, { "epoch": 40.708083832335326, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.12, "step": 27193 }, { "epoch": 40.709580838323355, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 27194 }, { "epoch": 40.711077844311376, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1222, "step": 27195 }, { "epoch": 40.712574850299404, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1249, "step": 27196 }, { "epoch": 40.714071856287426, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1234, "step": 27197 }, { "epoch": 40.71556886227545, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1276, "step": 27198 }, { "epoch": 40.717065868263475, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1228, "step": 27199 }, { "epoch": 40.7185628742515, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1243, "step": 27200 }, { "epoch": 40.72005988023952, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1231, "step": 27201 }, { "epoch": 40.721556886227546, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1234, "step": 27202 }, { "epoch": 40.72305389221557, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1229, "step": 27203 }, { "epoch": 40.724550898203596, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1197, "step": 27204 }, { "epoch": 40.72604790419162, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1221, "step": 27205 }, { "epoch": 40.72754491017964, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1237, "step": 27206 }, { "epoch": 40.72904191616767, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1262, "step": 27207 }, { "epoch": 40.73053892215569, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1263, "step": 27208 }, { "epoch": 40.73203592814371, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1228, "step": 27209 }, { "epoch": 40.73353293413174, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1272, "step": 27210 }, { "epoch": 40.73502994011976, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.121, "step": 27211 }, { "epoch": 40.73652694610779, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1225, "step": 27212 }, { "epoch": 40.73802395209581, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1205, "step": 27213 }, { "epoch": 40.73952095808383, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1242, "step": 27214 }, { "epoch": 40.74101796407186, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1248, "step": 27215 }, { "epoch": 40.74251497005988, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1206, "step": 27216 }, { "epoch": 40.7440119760479, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1266, "step": 27217 }, { "epoch": 40.74550898203593, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1304, "step": 27218 }, { "epoch": 40.74700598802395, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1193, "step": 27219 }, { "epoch": 40.74850299401198, "grad_norm": 0.048095703125, "learning_rate": 0.0008, "loss": 1.119, "step": 27220 }, { "epoch": 40.75, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1273, "step": 27221 }, { "epoch": 40.75149700598802, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1147, "step": 27222 }, { "epoch": 40.75299401197605, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1268, "step": 27223 }, { "epoch": 40.75449101796407, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1282, "step": 27224 }, { "epoch": 40.7559880239521, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1299, "step": 27225 }, { "epoch": 40.75748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1224, "step": 27226 }, { "epoch": 40.75898203592814, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1279, "step": 27227 }, { "epoch": 40.76047904191617, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1215, "step": 27228 }, { "epoch": 40.76197604790419, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1202, "step": 27229 }, { "epoch": 40.76347305389221, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1273, "step": 27230 }, { "epoch": 40.76497005988024, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1264, "step": 27231 }, { "epoch": 40.76646706586826, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1185, "step": 27232 }, { "epoch": 40.76796407185629, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1228, "step": 27233 }, { "epoch": 40.76946107784431, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1189, "step": 27234 }, { "epoch": 40.77095808383233, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1216, "step": 27235 }, { "epoch": 40.77245508982036, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1268, "step": 27236 }, { "epoch": 40.77395209580838, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1238, "step": 27237 }, { "epoch": 40.775449101796404, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1213, "step": 27238 }, { "epoch": 40.77694610778443, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1272, "step": 27239 }, { "epoch": 40.778443113772454, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1225, "step": 27240 }, { "epoch": 40.77994011976048, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1198, "step": 27241 }, { "epoch": 40.7814371257485, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1256, "step": 27242 }, { "epoch": 40.782934131736525, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 27243 }, { "epoch": 40.78443113772455, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1246, "step": 27244 }, { "epoch": 40.785928143712574, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1201, "step": 27245 }, { "epoch": 40.787425149700596, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1288, "step": 27246 }, { "epoch": 40.788922155688624, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1209, "step": 27247 }, { "epoch": 40.790419161676645, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1238, "step": 27248 }, { "epoch": 40.791916167664674, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1236, "step": 27249 }, { "epoch": 40.793413173652695, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1251, "step": 27250 }, { "epoch": 40.794910179640716, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1164, "step": 27251 }, { "epoch": 40.796407185628745, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1243, "step": 27252 }, { "epoch": 40.797904191616766, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.118, "step": 27253 }, { "epoch": 40.79940119760479, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1237, "step": 27254 }, { "epoch": 40.800898203592816, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1228, "step": 27255 }, { "epoch": 40.80239520958084, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1212, "step": 27256 }, { "epoch": 40.803892215568865, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1238, "step": 27257 }, { "epoch": 40.80538922155689, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1251, "step": 27258 }, { "epoch": 40.80688622754491, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1224, "step": 27259 }, { "epoch": 40.808383233532936, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1294, "step": 27260 }, { "epoch": 40.80988023952096, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1193, "step": 27261 }, { "epoch": 40.81137724550898, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.126, "step": 27262 }, { "epoch": 40.81287425149701, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1228, "step": 27263 }, { "epoch": 40.81437125748503, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1273, "step": 27264 }, { "epoch": 40.81586826347306, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1277, "step": 27265 }, { "epoch": 40.81736526946108, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1227, "step": 27266 }, { "epoch": 40.8188622754491, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1187, "step": 27267 }, { "epoch": 40.82035928143713, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1235, "step": 27268 }, { "epoch": 40.82185628742515, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.125, "step": 27269 }, { "epoch": 40.82335329341317, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1231, "step": 27270 }, { "epoch": 40.8248502994012, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1176, "step": 27271 }, { "epoch": 40.82634730538922, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1218, "step": 27272 }, { "epoch": 40.82784431137725, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1215, "step": 27273 }, { "epoch": 40.82934131736527, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1215, "step": 27274 }, { "epoch": 40.83083832335329, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1217, "step": 27275 }, { "epoch": 40.83233532934132, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1274, "step": 27276 }, { "epoch": 40.83383233532934, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1278, "step": 27277 }, { "epoch": 40.83532934131736, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1205, "step": 27278 }, { "epoch": 40.83682634730539, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.118, "step": 27279 }, { "epoch": 40.83832335329341, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1192, "step": 27280 }, { "epoch": 40.83982035928144, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1272, "step": 27281 }, { "epoch": 40.84131736526946, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1217, "step": 27282 }, { "epoch": 40.84281437125748, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1263, "step": 27283 }, { "epoch": 40.84431137724551, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1255, "step": 27284 }, { "epoch": 40.84580838323353, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1199, "step": 27285 }, { "epoch": 40.84730538922156, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 27286 }, { "epoch": 40.84880239520958, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1254, "step": 27287 }, { "epoch": 40.8502994011976, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1293, "step": 27288 }, { "epoch": 40.85179640718563, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1211, "step": 27289 }, { "epoch": 40.85329341317365, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 27290 }, { "epoch": 40.854790419161674, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1244, "step": 27291 }, { "epoch": 40.8562874251497, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1181, "step": 27292 }, { "epoch": 40.85778443113772, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1188, "step": 27293 }, { "epoch": 40.85928143712575, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1196, "step": 27294 }, { "epoch": 40.86077844311377, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1276, "step": 27295 }, { "epoch": 40.862275449101794, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1203, "step": 27296 }, { "epoch": 40.86377245508982, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1281, "step": 27297 }, { "epoch": 40.865269461077844, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1247, "step": 27298 }, { "epoch": 40.866766467065865, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1252, "step": 27299 }, { "epoch": 40.868263473053894, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1257, "step": 27300 }, { "epoch": 40.869760479041915, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1237, "step": 27301 }, { "epoch": 40.87125748502994, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 27302 }, { "epoch": 40.872754491017965, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1227, "step": 27303 }, { "epoch": 40.874251497005986, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1246, "step": 27304 }, { "epoch": 40.875748502994014, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1301, "step": 27305 }, { "epoch": 40.877245508982035, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1255, "step": 27306 }, { "epoch": 40.87874251497006, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1239, "step": 27307 }, { "epoch": 40.880239520958085, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.118, "step": 27308 }, { "epoch": 40.881736526946106, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1271, "step": 27309 }, { "epoch": 40.883233532934135, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1259, "step": 27310 }, { "epoch": 40.884730538922156, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1221, "step": 27311 }, { "epoch": 40.88622754491018, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1249, "step": 27312 }, { "epoch": 40.887724550898206, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1278, "step": 27313 }, { "epoch": 40.88922155688623, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1206, "step": 27314 }, { "epoch": 40.89071856287425, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1331, "step": 27315 }, { "epoch": 40.89221556886228, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1251, "step": 27316 }, { "epoch": 40.8937125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1209, "step": 27317 }, { "epoch": 40.895209580838326, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1224, "step": 27318 }, { "epoch": 40.89670658682635, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.12, "step": 27319 }, { "epoch": 40.89820359281437, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1184, "step": 27320 }, { "epoch": 40.8997005988024, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1248, "step": 27321 }, { "epoch": 40.90119760479042, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1213, "step": 27322 }, { "epoch": 40.90269461077844, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1254, "step": 27323 }, { "epoch": 40.90419161676647, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1218, "step": 27324 }, { "epoch": 40.90568862275449, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1258, "step": 27325 }, { "epoch": 40.90718562874252, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1308, "step": 27326 }, { "epoch": 40.90868263473054, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1215, "step": 27327 }, { "epoch": 40.91017964071856, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1242, "step": 27328 }, { "epoch": 40.91167664670659, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.117, "step": 27329 }, { "epoch": 40.91317365269461, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1327, "step": 27330 }, { "epoch": 40.91467065868264, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1282, "step": 27331 }, { "epoch": 40.91616766467066, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1166, "step": 27332 }, { "epoch": 40.91766467065868, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1256, "step": 27333 }, { "epoch": 40.91916167664671, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1343, "step": 27334 }, { "epoch": 40.92065868263473, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1208, "step": 27335 }, { "epoch": 40.92215568862275, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1295, "step": 27336 }, { "epoch": 40.92365269461078, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1252, "step": 27337 }, { "epoch": 40.9251497005988, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1279, "step": 27338 }, { "epoch": 40.92664670658683, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1266, "step": 27339 }, { "epoch": 40.92814371257485, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1212, "step": 27340 }, { "epoch": 40.92964071856287, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1229, "step": 27341 }, { "epoch": 40.9311377245509, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1225, "step": 27342 }, { "epoch": 40.93263473053892, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1243, "step": 27343 }, { "epoch": 40.93413173652694, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1163, "step": 27344 }, { "epoch": 40.93562874251497, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1239, "step": 27345 }, { "epoch": 40.93712574850299, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1278, "step": 27346 }, { "epoch": 40.93862275449102, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1242, "step": 27347 }, { "epoch": 40.94011976047904, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1275, "step": 27348 }, { "epoch": 40.941616766467064, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1227, "step": 27349 }, { "epoch": 40.94311377245509, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1224, "step": 27350 }, { "epoch": 40.94461077844311, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.121, "step": 27351 }, { "epoch": 40.946107784431135, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1196, "step": 27352 }, { "epoch": 40.94760479041916, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1234, "step": 27353 }, { "epoch": 40.949101796407184, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1224, "step": 27354 }, { "epoch": 40.95059880239521, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1243, "step": 27355 }, { "epoch": 40.952095808383234, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1259, "step": 27356 }, { "epoch": 40.953592814371255, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1274, "step": 27357 }, { "epoch": 40.955089820359284, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.126, "step": 27358 }, { "epoch": 40.956586826347305, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1258, "step": 27359 }, { "epoch": 40.958083832335326, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1253, "step": 27360 }, { "epoch": 40.959580838323355, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1246, "step": 27361 }, { "epoch": 40.961077844311376, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.123, "step": 27362 }, { "epoch": 40.962574850299404, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1229, "step": 27363 }, { "epoch": 40.964071856287426, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1185, "step": 27364 }, { "epoch": 40.96556886227545, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1151, "step": 27365 }, { "epoch": 40.967065868263475, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1264, "step": 27366 }, { "epoch": 40.9685628742515, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.115, "step": 27367 }, { "epoch": 40.97005988023952, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1233, "step": 27368 }, { "epoch": 40.971556886227546, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1249, "step": 27369 }, { "epoch": 40.97305389221557, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1248, "step": 27370 }, { "epoch": 40.974550898203596, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.121, "step": 27371 }, { "epoch": 40.97604790419162, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1252, "step": 27372 }, { "epoch": 40.97754491017964, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1222, "step": 27373 }, { "epoch": 40.97904191616767, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1269, "step": 27374 }, { "epoch": 40.98053892215569, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1227, "step": 27375 }, { "epoch": 40.98203592814371, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1216, "step": 27376 }, { "epoch": 40.98353293413174, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.128, "step": 27377 }, { "epoch": 40.98502994011976, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1189, "step": 27378 }, { "epoch": 40.98652694610779, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.1212, "step": 27379 }, { "epoch": 40.98802395209581, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1335, "step": 27380 }, { "epoch": 40.98952095808383, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1261, "step": 27381 }, { "epoch": 40.99101796407186, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.122, "step": 27382 }, { "epoch": 40.99251497005988, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1211, "step": 27383 }, { "epoch": 40.9940119760479, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1229, "step": 27384 }, { "epoch": 40.99550898203593, "grad_norm": 0.046630859375, "learning_rate": 0.0008, "loss": 1.1193, "step": 27385 }, { "epoch": 40.99700598802395, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1209, "step": 27386 }, { "epoch": 40.99850299401198, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1289, "step": 27387 }, { "epoch": 41.0, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.119, "step": 27388 }, { "epoch": 41.00149700598802, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1215, "step": 27389 }, { "epoch": 41.00299401197605, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1215, "step": 27390 }, { "epoch": 41.00449101796407, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.118, "step": 27391 }, { "epoch": 41.0059880239521, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1245, "step": 27392 }, { "epoch": 41.00748502994012, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1218, "step": 27393 }, { "epoch": 41.00898203592814, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1246, "step": 27394 }, { "epoch": 41.01047904191617, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1208, "step": 27395 }, { "epoch": 41.01197604790419, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 27396 }, { "epoch": 41.01347305389221, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1229, "step": 27397 }, { "epoch": 41.01497005988024, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1224, "step": 27398 }, { "epoch": 41.01646706586826, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1256, "step": 27399 }, { "epoch": 41.01796407185629, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1272, "step": 27400 }, { "epoch": 41.01946107784431, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1227, "step": 27401 }, { "epoch": 41.02095808383233, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1242, "step": 27402 }, { "epoch": 41.02245508982036, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1216, "step": 27403 }, { "epoch": 41.02395209580838, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1243, "step": 27404 }, { "epoch": 41.025449101796404, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1245, "step": 27405 }, { "epoch": 41.02694610778443, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.1256, "step": 27406 }, { "epoch": 41.028443113772454, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1286, "step": 27407 }, { "epoch": 41.02994011976048, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1262, "step": 27408 }, { "epoch": 41.0314371257485, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1147, "step": 27409 }, { "epoch": 41.032934131736525, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1241, "step": 27410 }, { "epoch": 41.03443113772455, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1185, "step": 27411 }, { "epoch": 41.035928143712574, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1243, "step": 27412 }, { "epoch": 41.037425149700596, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1195, "step": 27413 }, { "epoch": 41.038922155688624, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1211, "step": 27414 }, { "epoch": 41.040419161676645, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.1264, "step": 27415 }, { "epoch": 41.041916167664674, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1243, "step": 27416 }, { "epoch": 41.043413173652695, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.125, "step": 27417 }, { "epoch": 41.044910179640716, "grad_norm": 0.193359375, "learning_rate": 0.0008, "loss": 1.1249, "step": 27418 }, { "epoch": 41.046407185628745, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1222, "step": 27419 }, { "epoch": 41.047904191616766, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1212, "step": 27420 }, { "epoch": 41.04940119760479, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1226, "step": 27421 }, { "epoch": 41.050898203592816, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1235, "step": 27422 }, { "epoch": 41.05239520958084, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.1272, "step": 27423 }, { "epoch": 41.053892215568865, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1237, "step": 27424 }, { "epoch": 41.05538922155689, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1241, "step": 27425 }, { "epoch": 41.05688622754491, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1218, "step": 27426 }, { "epoch": 41.058383233532936, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1259, "step": 27427 }, { "epoch": 41.05988023952096, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1237, "step": 27428 }, { "epoch": 41.06137724550898, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1257, "step": 27429 }, { "epoch": 41.06287425149701, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1236, "step": 27430 }, { "epoch": 41.06437125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1272, "step": 27431 }, { "epoch": 41.06586826347306, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 27432 }, { "epoch": 41.06736526946108, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 27433 }, { "epoch": 41.0688622754491, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1262, "step": 27434 }, { "epoch": 41.07035928143713, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1195, "step": 27435 }, { "epoch": 41.07185628742515, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1257, "step": 27436 }, { "epoch": 41.07335329341317, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1241, "step": 27437 }, { "epoch": 41.0748502994012, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1212, "step": 27438 }, { "epoch": 41.07634730538922, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 27439 }, { "epoch": 41.07784431137725, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 27440 }, { "epoch": 41.07934131736527, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1277, "step": 27441 }, { "epoch": 41.08083832335329, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1297, "step": 27442 }, { "epoch": 41.08233532934132, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1268, "step": 27443 }, { "epoch": 41.08383233532934, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1225, "step": 27444 }, { "epoch": 41.08532934131737, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1218, "step": 27445 }, { "epoch": 41.08682634730539, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1233, "step": 27446 }, { "epoch": 41.08832335329341, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 27447 }, { "epoch": 41.08982035928144, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1202, "step": 27448 }, { "epoch": 41.09131736526946, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1235, "step": 27449 }, { "epoch": 41.09281437125748, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1158, "step": 27450 }, { "epoch": 41.09431137724551, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1274, "step": 27451 }, { "epoch": 41.09580838323353, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1265, "step": 27452 }, { "epoch": 41.09730538922156, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1283, "step": 27453 }, { "epoch": 41.09880239520958, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1203, "step": 27454 }, { "epoch": 41.1002994011976, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1213, "step": 27455 }, { "epoch": 41.10179640718563, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1228, "step": 27456 }, { "epoch": 41.10329341317365, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.127, "step": 27457 }, { "epoch": 41.104790419161674, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1287, "step": 27458 }, { "epoch": 41.1062874251497, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.122, "step": 27459 }, { "epoch": 41.10778443113772, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 27460 }, { "epoch": 41.10928143712575, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1192, "step": 27461 }, { "epoch": 41.11077844311377, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.126, "step": 27462 }, { "epoch": 41.112275449101794, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1221, "step": 27463 }, { "epoch": 41.11377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1264, "step": 27464 }, { "epoch": 41.115269461077844, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1279, "step": 27465 }, { "epoch": 41.116766467065865, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1235, "step": 27466 }, { "epoch": 41.118263473053894, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1235, "step": 27467 }, { "epoch": 41.119760479041915, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.125, "step": 27468 }, { "epoch": 41.12125748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1286, "step": 27469 }, { "epoch": 41.122754491017965, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1248, "step": 27470 }, { "epoch": 41.124251497005986, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1249, "step": 27471 }, { "epoch": 41.125748502994014, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1245, "step": 27472 }, { "epoch": 41.127245508982035, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1226, "step": 27473 }, { "epoch": 41.12874251497006, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1249, "step": 27474 }, { "epoch": 41.130239520958085, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1216, "step": 27475 }, { "epoch": 41.131736526946106, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 27476 }, { "epoch": 41.133233532934135, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1229, "step": 27477 }, { "epoch": 41.134730538922156, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1232, "step": 27478 }, { "epoch": 41.13622754491018, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 27479 }, { "epoch": 41.137724550898206, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1292, "step": 27480 }, { "epoch": 41.13922155688623, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1214, "step": 27481 }, { "epoch": 41.14071856287425, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1299, "step": 27482 }, { "epoch": 41.14221556886228, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1281, "step": 27483 }, { "epoch": 41.1437125748503, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1228, "step": 27484 }, { "epoch": 41.145209580838326, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.123, "step": 27485 }, { "epoch": 41.14670658682635, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1201, "step": 27486 }, { "epoch": 41.14820359281437, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1235, "step": 27487 }, { "epoch": 41.1497005988024, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1223, "step": 27488 }, { "epoch": 41.15119760479042, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1261, "step": 27489 }, { "epoch": 41.15269461077844, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1247, "step": 27490 }, { "epoch": 41.15419161676647, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1282, "step": 27491 }, { "epoch": 41.15568862275449, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1279, "step": 27492 }, { "epoch": 41.15718562874252, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1287, "step": 27493 }, { "epoch": 41.15868263473054, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1257, "step": 27494 }, { "epoch": 41.16017964071856, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1222, "step": 27495 }, { "epoch": 41.16167664670659, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1209, "step": 27496 }, { "epoch": 41.16317365269461, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1205, "step": 27497 }, { "epoch": 41.16467065868263, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1261, "step": 27498 }, { "epoch": 41.16616766467066, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1217, "step": 27499 }, { "epoch": 41.16766467065868, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.121, "step": 27500 }, { "epoch": 41.16916167664671, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1222, "step": 27501 }, { "epoch": 41.17065868263473, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1223, "step": 27502 }, { "epoch": 41.17215568862275, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1279, "step": 27503 }, { "epoch": 41.17365269461078, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1217, "step": 27504 }, { "epoch": 41.1751497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1219, "step": 27505 }, { "epoch": 41.17664670658683, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1226, "step": 27506 }, { "epoch": 41.17814371257485, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1242, "step": 27507 }, { "epoch": 41.17964071856287, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1218, "step": 27508 }, { "epoch": 41.1811377245509, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1312, "step": 27509 }, { "epoch": 41.18263473053892, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1211, "step": 27510 }, { "epoch": 41.18413173652694, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1244, "step": 27511 }, { "epoch": 41.18562874251497, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1308, "step": 27512 }, { "epoch": 41.18712574850299, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1222, "step": 27513 }, { "epoch": 41.18862275449102, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1257, "step": 27514 }, { "epoch": 41.19011976047904, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1259, "step": 27515 }, { "epoch": 41.191616766467064, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1198, "step": 27516 }, { "epoch": 41.19311377245509, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 27517 }, { "epoch": 41.19461077844311, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1225, "step": 27518 }, { "epoch": 41.196107784431135, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1227, "step": 27519 }, { "epoch": 41.19760479041916, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1207, "step": 27520 }, { "epoch": 41.199101796407184, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1277, "step": 27521 }, { "epoch": 41.20059880239521, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1221, "step": 27522 }, { "epoch": 41.202095808383234, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1262, "step": 27523 }, { "epoch": 41.203592814371255, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1202, "step": 27524 }, { "epoch": 41.205089820359284, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1263, "step": 27525 }, { "epoch": 41.206586826347305, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1212, "step": 27526 }, { "epoch": 41.208083832335326, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1188, "step": 27527 }, { "epoch": 41.209580838323355, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1285, "step": 27528 }, { "epoch": 41.211077844311376, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1258, "step": 27529 }, { "epoch": 41.212574850299404, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1214, "step": 27530 }, { "epoch": 41.214071856287426, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1323, "step": 27531 }, { "epoch": 41.21556886227545, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.118, "step": 27532 }, { "epoch": 41.217065868263475, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1199, "step": 27533 }, { "epoch": 41.2185628742515, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1213, "step": 27534 }, { "epoch": 41.22005988023952, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.122, "step": 27535 }, { "epoch": 41.221556886227546, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1279, "step": 27536 }, { "epoch": 41.22305389221557, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1241, "step": 27537 }, { "epoch": 41.224550898203596, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1241, "step": 27538 }, { "epoch": 41.22604790419162, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1243, "step": 27539 }, { "epoch": 41.22754491017964, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1213, "step": 27540 }, { "epoch": 41.22904191616767, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.12, "step": 27541 }, { "epoch": 41.23053892215569, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1235, "step": 27542 }, { "epoch": 41.23203592814371, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1265, "step": 27543 }, { "epoch": 41.23353293413174, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1258, "step": 27544 }, { "epoch": 41.23502994011976, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1211, "step": 27545 }, { "epoch": 41.23652694610779, "grad_norm": 0.051025390625, "learning_rate": 0.0008, "loss": 1.1229, "step": 27546 }, { "epoch": 41.23802395209581, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1237, "step": 27547 }, { "epoch": 41.23952095808383, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1215, "step": 27548 }, { "epoch": 41.24101796407186, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1272, "step": 27549 }, { "epoch": 41.24251497005988, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1291, "step": 27550 }, { "epoch": 41.2440119760479, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1299, "step": 27551 }, { "epoch": 41.24550898203593, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.119, "step": 27552 }, { "epoch": 41.24700598802395, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1251, "step": 27553 }, { "epoch": 41.24850299401198, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1178, "step": 27554 }, { "epoch": 41.25, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1262, "step": 27555 }, { "epoch": 41.25149700598802, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1224, "step": 27556 }, { "epoch": 41.25299401197605, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1248, "step": 27557 }, { "epoch": 41.25449101796407, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1211, "step": 27558 }, { "epoch": 41.2559880239521, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1253, "step": 27559 }, { "epoch": 41.25748502994012, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 27560 }, { "epoch": 41.25898203592814, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1243, "step": 27561 }, { "epoch": 41.26047904191617, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1235, "step": 27562 }, { "epoch": 41.26197604790419, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 27563 }, { "epoch": 41.26347305389221, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1175, "step": 27564 }, { "epoch": 41.26497005988024, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1231, "step": 27565 }, { "epoch": 41.26646706586826, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 27566 }, { "epoch": 41.26796407185629, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1171, "step": 27567 }, { "epoch": 41.26946107784431, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1311, "step": 27568 }, { "epoch": 41.27095808383233, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1228, "step": 27569 }, { "epoch": 41.27245508982036, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.123, "step": 27570 }, { "epoch": 41.27395209580838, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1197, "step": 27571 }, { "epoch": 41.275449101796404, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1171, "step": 27572 }, { "epoch": 41.27694610778443, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1251, "step": 27573 }, { "epoch": 41.278443113772454, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1255, "step": 27574 }, { "epoch": 41.27994011976048, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.123, "step": 27575 }, { "epoch": 41.2814371257485, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1295, "step": 27576 }, { "epoch": 41.282934131736525, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1241, "step": 27577 }, { "epoch": 41.28443113772455, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1261, "step": 27578 }, { "epoch": 41.285928143712574, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.123, "step": 27579 }, { "epoch": 41.287425149700596, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1223, "step": 27580 }, { "epoch": 41.288922155688624, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1213, "step": 27581 }, { "epoch": 41.290419161676645, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1204, "step": 27582 }, { "epoch": 41.291916167664674, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1259, "step": 27583 }, { "epoch": 41.293413173652695, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1239, "step": 27584 }, { "epoch": 41.294910179640716, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1189, "step": 27585 }, { "epoch": 41.296407185628745, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1279, "step": 27586 }, { "epoch": 41.297904191616766, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1216, "step": 27587 }, { "epoch": 41.29940119760479, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1212, "step": 27588 }, { "epoch": 41.300898203592816, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1176, "step": 27589 }, { "epoch": 41.30239520958084, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1277, "step": 27590 }, { "epoch": 41.303892215568865, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1282, "step": 27591 }, { "epoch": 41.30538922155689, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1219, "step": 27592 }, { "epoch": 41.30688622754491, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1231, "step": 27593 }, { "epoch": 41.308383233532936, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.124, "step": 27594 }, { "epoch": 41.30988023952096, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1232, "step": 27595 }, { "epoch": 41.31137724550898, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1221, "step": 27596 }, { "epoch": 41.31287425149701, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1272, "step": 27597 }, { "epoch": 41.31437125748503, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1241, "step": 27598 }, { "epoch": 41.31586826347306, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1194, "step": 27599 }, { "epoch": 41.31736526946108, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1221, "step": 27600 }, { "epoch": 41.3188622754491, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1202, "step": 27601 }, { "epoch": 41.32035928143713, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1214, "step": 27602 }, { "epoch": 41.32185628742515, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.124, "step": 27603 }, { "epoch": 41.32335329341317, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1196, "step": 27604 }, { "epoch": 41.3248502994012, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1278, "step": 27605 }, { "epoch": 41.32634730538922, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1198, "step": 27606 }, { "epoch": 41.32784431137725, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.123, "step": 27607 }, { "epoch": 41.32934131736527, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.117, "step": 27608 }, { "epoch": 41.33083832335329, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1212, "step": 27609 }, { "epoch": 41.33233532934132, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1269, "step": 27610 }, { "epoch": 41.33383233532934, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.125, "step": 27611 }, { "epoch": 41.33532934131736, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1242, "step": 27612 }, { "epoch": 41.33682634730539, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1228, "step": 27613 }, { "epoch": 41.33832335329341, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1193, "step": 27614 }, { "epoch": 41.33982035928144, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1247, "step": 27615 }, { "epoch": 41.34131736526946, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1264, "step": 27616 }, { "epoch": 41.34281437125748, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1308, "step": 27617 }, { "epoch": 41.34431137724551, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1201, "step": 27618 }, { "epoch": 41.34580838323353, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1224, "step": 27619 }, { "epoch": 41.34730538922156, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1249, "step": 27620 }, { "epoch": 41.34880239520958, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1166, "step": 27621 }, { "epoch": 41.3502994011976, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1223, "step": 27622 }, { "epoch": 41.35179640718563, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1273, "step": 27623 }, { "epoch": 41.35329341317365, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1162, "step": 27624 }, { "epoch": 41.354790419161674, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1268, "step": 27625 }, { "epoch": 41.3562874251497, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1295, "step": 27626 }, { "epoch": 41.35778443113772, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1265, "step": 27627 }, { "epoch": 41.35928143712575, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1204, "step": 27628 }, { "epoch": 41.36077844311377, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 27629 }, { "epoch": 41.362275449101794, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1222, "step": 27630 }, { "epoch": 41.36377245508982, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1199, "step": 27631 }, { "epoch": 41.365269461077844, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1259, "step": 27632 }, { "epoch": 41.366766467065865, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1201, "step": 27633 }, { "epoch": 41.368263473053894, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1179, "step": 27634 }, { "epoch": 41.369760479041915, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1285, "step": 27635 }, { "epoch": 41.37125748502994, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1267, "step": 27636 }, { "epoch": 41.372754491017965, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1231, "step": 27637 }, { "epoch": 41.374251497005986, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1294, "step": 27638 }, { "epoch": 41.375748502994014, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1305, "step": 27639 }, { "epoch": 41.377245508982035, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 27640 }, { "epoch": 41.37874251497006, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1271, "step": 27641 }, { "epoch": 41.380239520958085, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1209, "step": 27642 }, { "epoch": 41.381736526946106, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1225, "step": 27643 }, { "epoch": 41.383233532934135, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1251, "step": 27644 }, { "epoch": 41.384730538922156, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1228, "step": 27645 }, { "epoch": 41.38622754491018, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1185, "step": 27646 }, { "epoch": 41.387724550898206, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1136, "step": 27647 }, { "epoch": 41.38922155688623, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1299, "step": 27648 }, { "epoch": 41.39071856287425, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1233, "step": 27649 }, { "epoch": 41.39221556886228, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1233, "step": 27650 }, { "epoch": 41.3937125748503, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1301, "step": 27651 }, { "epoch": 41.395209580838326, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1243, "step": 27652 }, { "epoch": 41.39670658682635, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 27653 }, { "epoch": 41.39820359281437, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1196, "step": 27654 }, { "epoch": 41.3997005988024, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1265, "step": 27655 }, { "epoch": 41.40119760479042, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1214, "step": 27656 }, { "epoch": 41.40269461077844, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1154, "step": 27657 }, { "epoch": 41.40419161676647, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1223, "step": 27658 }, { "epoch": 41.40568862275449, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1247, "step": 27659 }, { "epoch": 41.40718562874252, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.121, "step": 27660 }, { "epoch": 41.40868263473054, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1276, "step": 27661 }, { "epoch": 41.41017964071856, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1173, "step": 27662 }, { "epoch": 41.41167664670659, "grad_norm": 0.050048828125, "learning_rate": 0.0008, "loss": 1.1211, "step": 27663 }, { "epoch": 41.41317365269461, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1281, "step": 27664 }, { "epoch": 41.41467065868264, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1209, "step": 27665 }, { "epoch": 41.41616766467066, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1173, "step": 27666 }, { "epoch": 41.41766467065868, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1279, "step": 27667 }, { "epoch": 41.41916167664671, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1269, "step": 27668 }, { "epoch": 41.42065868263473, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1225, "step": 27669 }, { "epoch": 41.42215568862275, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1221, "step": 27670 }, { "epoch": 41.42365269461078, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1254, "step": 27671 }, { "epoch": 41.4251497005988, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1216, "step": 27672 }, { "epoch": 41.42664670658683, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1195, "step": 27673 }, { "epoch": 41.42814371257485, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1195, "step": 27674 }, { "epoch": 41.42964071856287, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.126, "step": 27675 }, { "epoch": 41.4311377245509, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1258, "step": 27676 }, { "epoch": 41.43263473053892, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1211, "step": 27677 }, { "epoch": 41.43413173652694, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1229, "step": 27678 }, { "epoch": 41.43562874251497, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1179, "step": 27679 }, { "epoch": 41.43712574850299, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1283, "step": 27680 }, { "epoch": 41.43862275449102, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.121, "step": 27681 }, { "epoch": 41.44011976047904, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 27682 }, { "epoch": 41.441616766467064, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1209, "step": 27683 }, { "epoch": 41.44311377245509, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1244, "step": 27684 }, { "epoch": 41.44461077844311, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1232, "step": 27685 }, { "epoch": 41.446107784431135, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1246, "step": 27686 }, { "epoch": 41.44760479041916, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1247, "step": 27687 }, { "epoch": 41.449101796407184, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1254, "step": 27688 }, { "epoch": 41.45059880239521, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1219, "step": 27689 }, { "epoch": 41.452095808383234, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1302, "step": 27690 }, { "epoch": 41.453592814371255, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1221, "step": 27691 }, { "epoch": 41.455089820359284, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.115, "step": 27692 }, { "epoch": 41.456586826347305, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1223, "step": 27693 }, { "epoch": 41.458083832335326, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1251, "step": 27694 }, { "epoch": 41.459580838323355, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1226, "step": 27695 }, { "epoch": 41.461077844311376, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1224, "step": 27696 }, { "epoch": 41.462574850299404, "grad_norm": 0.046630859375, "learning_rate": 0.0008, "loss": 1.1234, "step": 27697 }, { "epoch": 41.464071856287426, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1226, "step": 27698 }, { "epoch": 41.46556886227545, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1246, "step": 27699 }, { "epoch": 41.467065868263475, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1294, "step": 27700 }, { "epoch": 41.4685628742515, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1168, "step": 27701 }, { "epoch": 41.47005988023952, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1211, "step": 27702 }, { "epoch": 41.471556886227546, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1268, "step": 27703 }, { "epoch": 41.47305389221557, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1277, "step": 27704 }, { "epoch": 41.474550898203596, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1294, "step": 27705 }, { "epoch": 41.47604790419162, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.122, "step": 27706 }, { "epoch": 41.47754491017964, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1196, "step": 27707 }, { "epoch": 41.47904191616767, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1225, "step": 27708 }, { "epoch": 41.48053892215569, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1271, "step": 27709 }, { "epoch": 41.48203592814371, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1235, "step": 27710 }, { "epoch": 41.48353293413174, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1282, "step": 27711 }, { "epoch": 41.48502994011976, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1209, "step": 27712 }, { "epoch": 41.48652694610779, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1257, "step": 27713 }, { "epoch": 41.48802395209581, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1197, "step": 27714 }, { "epoch": 41.48952095808383, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1233, "step": 27715 }, { "epoch": 41.49101796407186, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.117, "step": 27716 }, { "epoch": 41.49251497005988, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 27717 }, { "epoch": 41.4940119760479, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1297, "step": 27718 }, { "epoch": 41.49550898203593, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1291, "step": 27719 }, { "epoch": 41.49700598802395, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1228, "step": 27720 }, { "epoch": 41.49850299401198, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1254, "step": 27721 }, { "epoch": 41.5, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1233, "step": 27722 }, { "epoch": 41.50149700598802, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1187, "step": 27723 }, { "epoch": 41.50299401197605, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1268, "step": 27724 }, { "epoch": 41.50449101796407, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1259, "step": 27725 }, { "epoch": 41.5059880239521, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.127, "step": 27726 }, { "epoch": 41.50748502994012, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1234, "step": 27727 }, { "epoch": 41.50898203592814, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1257, "step": 27728 }, { "epoch": 41.51047904191617, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1278, "step": 27729 }, { "epoch": 41.51197604790419, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1219, "step": 27730 }, { "epoch": 41.51347305389221, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1216, "step": 27731 }, { "epoch": 41.51497005988024, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1264, "step": 27732 }, { "epoch": 41.51646706586826, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1187, "step": 27733 }, { "epoch": 41.51796407185629, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1267, "step": 27734 }, { "epoch": 41.51946107784431, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1284, "step": 27735 }, { "epoch": 41.52095808383233, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1244, "step": 27736 }, { "epoch": 41.52245508982036, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1259, "step": 27737 }, { "epoch": 41.52395209580838, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1325, "step": 27738 }, { "epoch": 41.525449101796404, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1265, "step": 27739 }, { "epoch": 41.52694610778443, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1286, "step": 27740 }, { "epoch": 41.528443113772454, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1225, "step": 27741 }, { "epoch": 41.52994011976048, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.123, "step": 27742 }, { "epoch": 41.5314371257485, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1238, "step": 27743 }, { "epoch": 41.532934131736525, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1216, "step": 27744 }, { "epoch": 41.53443113772455, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1265, "step": 27745 }, { "epoch": 41.535928143712574, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1265, "step": 27746 }, { "epoch": 41.537425149700596, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1217, "step": 27747 }, { "epoch": 41.538922155688624, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1272, "step": 27748 }, { "epoch": 41.540419161676645, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1225, "step": 27749 }, { "epoch": 41.541916167664674, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1216, "step": 27750 }, { "epoch": 41.543413173652695, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1237, "step": 27751 }, { "epoch": 41.544910179640716, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.132, "step": 27752 }, { "epoch": 41.546407185628745, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1198, "step": 27753 }, { "epoch": 41.547904191616766, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1202, "step": 27754 }, { "epoch": 41.54940119760479, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 27755 }, { "epoch": 41.550898203592816, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1237, "step": 27756 }, { "epoch": 41.55239520958084, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1204, "step": 27757 }, { "epoch": 41.553892215568865, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1232, "step": 27758 }, { "epoch": 41.55538922155689, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1182, "step": 27759 }, { "epoch": 41.55688622754491, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1233, "step": 27760 }, { "epoch": 41.558383233532936, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1236, "step": 27761 }, { "epoch": 41.55988023952096, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1223, "step": 27762 }, { "epoch": 41.56137724550898, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.129, "step": 27763 }, { "epoch": 41.56287425149701, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1271, "step": 27764 }, { "epoch": 41.56437125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1193, "step": 27765 }, { "epoch": 41.56586826347306, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 27766 }, { "epoch": 41.56736526946108, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1186, "step": 27767 }, { "epoch": 41.5688622754491, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1267, "step": 27768 }, { "epoch": 41.57035928143713, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1206, "step": 27769 }, { "epoch": 41.57185628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1237, "step": 27770 }, { "epoch": 41.57335329341317, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1215, "step": 27771 }, { "epoch": 41.5748502994012, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1189, "step": 27772 }, { "epoch": 41.57634730538922, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1265, "step": 27773 }, { "epoch": 41.57784431137725, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1267, "step": 27774 }, { "epoch": 41.57934131736527, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1207, "step": 27775 }, { "epoch": 41.58083832335329, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1267, "step": 27776 }, { "epoch": 41.58233532934132, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 27777 }, { "epoch": 41.58383233532934, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1192, "step": 27778 }, { "epoch": 41.58532934131736, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1213, "step": 27779 }, { "epoch": 41.58682634730539, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1264, "step": 27780 }, { "epoch": 41.58832335329341, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1207, "step": 27781 }, { "epoch": 41.58982035928144, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1237, "step": 27782 }, { "epoch": 41.59131736526946, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1222, "step": 27783 }, { "epoch": 41.59281437125748, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1181, "step": 27784 }, { "epoch": 41.59431137724551, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1293, "step": 27785 }, { "epoch": 41.59580838323353, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1266, "step": 27786 }, { "epoch": 41.59730538922156, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1242, "step": 27787 }, { "epoch": 41.59880239520958, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1193, "step": 27788 }, { "epoch": 41.6002994011976, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1214, "step": 27789 }, { "epoch": 41.60179640718563, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1232, "step": 27790 }, { "epoch": 41.60329341317365, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1199, "step": 27791 }, { "epoch": 41.604790419161674, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1283, "step": 27792 }, { "epoch": 41.6062874251497, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1287, "step": 27793 }, { "epoch": 41.60778443113772, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1249, "step": 27794 }, { "epoch": 41.60928143712575, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1272, "step": 27795 }, { "epoch": 41.61077844311377, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1258, "step": 27796 }, { "epoch": 41.612275449101794, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1189, "step": 27797 }, { "epoch": 41.61377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1293, "step": 27798 }, { "epoch": 41.615269461077844, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1231, "step": 27799 }, { "epoch": 41.616766467065865, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1244, "step": 27800 }, { "epoch": 41.618263473053894, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1304, "step": 27801 }, { "epoch": 41.619760479041915, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1206, "step": 27802 }, { "epoch": 41.62125748502994, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1171, "step": 27803 }, { "epoch": 41.622754491017965, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1317, "step": 27804 }, { "epoch": 41.624251497005986, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.126, "step": 27805 }, { "epoch": 41.625748502994014, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1237, "step": 27806 }, { "epoch": 41.627245508982035, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1298, "step": 27807 }, { "epoch": 41.62874251497006, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1247, "step": 27808 }, { "epoch": 41.630239520958085, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1233, "step": 27809 }, { "epoch": 41.631736526946106, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1184, "step": 27810 }, { "epoch": 41.633233532934135, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 27811 }, { "epoch": 41.634730538922156, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1266, "step": 27812 }, { "epoch": 41.63622754491018, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1234, "step": 27813 }, { "epoch": 41.637724550898206, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1233, "step": 27814 }, { "epoch": 41.63922155688623, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1197, "step": 27815 }, { "epoch": 41.64071856287425, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1229, "step": 27816 }, { "epoch": 41.64221556886228, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1268, "step": 27817 }, { "epoch": 41.6437125748503, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1222, "step": 27818 }, { "epoch": 41.645209580838326, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1248, "step": 27819 }, { "epoch": 41.64670658682635, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1209, "step": 27820 }, { "epoch": 41.64820359281437, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1276, "step": 27821 }, { "epoch": 41.6497005988024, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1219, "step": 27822 }, { "epoch": 41.65119760479042, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1243, "step": 27823 }, { "epoch": 41.65269461077844, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1246, "step": 27824 }, { "epoch": 41.65419161676647, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1192, "step": 27825 }, { "epoch": 41.65568862275449, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.126, "step": 27826 }, { "epoch": 41.65718562874252, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1208, "step": 27827 }, { "epoch": 41.65868263473054, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1227, "step": 27828 }, { "epoch": 41.66017964071856, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1216, "step": 27829 }, { "epoch": 41.66167664670659, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1251, "step": 27830 }, { "epoch": 41.66317365269461, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1226, "step": 27831 }, { "epoch": 41.66467065868264, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1228, "step": 27832 }, { "epoch": 41.66616766467066, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1237, "step": 27833 }, { "epoch": 41.66766467065868, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1217, "step": 27834 }, { "epoch": 41.66916167664671, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1219, "step": 27835 }, { "epoch": 41.67065868263473, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1199, "step": 27836 }, { "epoch": 41.67215568862275, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1221, "step": 27837 }, { "epoch": 41.67365269461078, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1243, "step": 27838 }, { "epoch": 41.6751497005988, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1149, "step": 27839 }, { "epoch": 41.67664670658683, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1262, "step": 27840 }, { "epoch": 41.67814371257485, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1241, "step": 27841 }, { "epoch": 41.67964071856287, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1195, "step": 27842 }, { "epoch": 41.6811377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1246, "step": 27843 }, { "epoch": 41.68263473053892, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1292, "step": 27844 }, { "epoch": 41.68413173652694, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1217, "step": 27845 }, { "epoch": 41.68562874251497, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1259, "step": 27846 }, { "epoch": 41.68712574850299, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1255, "step": 27847 }, { "epoch": 41.68862275449102, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.131, "step": 27848 }, { "epoch": 41.69011976047904, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1202, "step": 27849 }, { "epoch": 41.691616766467064, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 27850 }, { "epoch": 41.69311377245509, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1218, "step": 27851 }, { "epoch": 41.69461077844311, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1235, "step": 27852 }, { "epoch": 41.696107784431135, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1304, "step": 27853 }, { "epoch": 41.69760479041916, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.131, "step": 27854 }, { "epoch": 41.699101796407184, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1236, "step": 27855 }, { "epoch": 41.70059880239521, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1253, "step": 27856 }, { "epoch": 41.702095808383234, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1261, "step": 27857 }, { "epoch": 41.703592814371255, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1238, "step": 27858 }, { "epoch": 41.705089820359284, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1239, "step": 27859 }, { "epoch": 41.706586826347305, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 27860 }, { "epoch": 41.708083832335326, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1217, "step": 27861 }, { "epoch": 41.709580838323355, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 27862 }, { "epoch": 41.711077844311376, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1232, "step": 27863 }, { "epoch": 41.712574850299404, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.125, "step": 27864 }, { "epoch": 41.714071856287426, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1208, "step": 27865 }, { "epoch": 41.71556886227545, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1224, "step": 27866 }, { "epoch": 41.717065868263475, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1191, "step": 27867 }, { "epoch": 41.7185628742515, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1225, "step": 27868 }, { "epoch": 41.72005988023952, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1267, "step": 27869 }, { "epoch": 41.721556886227546, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1237, "step": 27870 }, { "epoch": 41.72305389221557, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.123, "step": 27871 }, { "epoch": 41.724550898203596, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1244, "step": 27872 }, { "epoch": 41.72604790419162, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.126, "step": 27873 }, { "epoch": 41.72754491017964, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.127, "step": 27874 }, { "epoch": 41.72904191616767, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1176, "step": 27875 }, { "epoch": 41.73053892215569, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1254, "step": 27876 }, { "epoch": 41.73203592814371, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.128, "step": 27877 }, { "epoch": 41.73353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1219, "step": 27878 }, { "epoch": 41.73502994011976, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1243, "step": 27879 }, { "epoch": 41.73652694610779, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1258, "step": 27880 }, { "epoch": 41.73802395209581, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1174, "step": 27881 }, { "epoch": 41.73952095808383, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1235, "step": 27882 }, { "epoch": 41.74101796407186, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1231, "step": 27883 }, { "epoch": 41.74251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1222, "step": 27884 }, { "epoch": 41.7440119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1202, "step": 27885 }, { "epoch": 41.74550898203593, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1329, "step": 27886 }, { "epoch": 41.74700598802395, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1249, "step": 27887 }, { "epoch": 41.74850299401198, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1265, "step": 27888 }, { "epoch": 41.75, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1203, "step": 27889 }, { "epoch": 41.75149700598802, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1221, "step": 27890 }, { "epoch": 41.75299401197605, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.122, "step": 27891 }, { "epoch": 41.75449101796407, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1256, "step": 27892 }, { "epoch": 41.7559880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1229, "step": 27893 }, { "epoch": 41.75748502994012, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1221, "step": 27894 }, { "epoch": 41.75898203592814, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1235, "step": 27895 }, { "epoch": 41.76047904191617, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1231, "step": 27896 }, { "epoch": 41.76197604790419, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1231, "step": 27897 }, { "epoch": 41.76347305389221, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1266, "step": 27898 }, { "epoch": 41.76497005988024, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1208, "step": 27899 }, { "epoch": 41.76646706586826, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1267, "step": 27900 }, { "epoch": 41.76796407185629, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1255, "step": 27901 }, { "epoch": 41.76946107784431, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1285, "step": 27902 }, { "epoch": 41.77095808383233, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.131, "step": 27903 }, { "epoch": 41.77245508982036, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1246, "step": 27904 }, { "epoch": 41.77395209580838, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1229, "step": 27905 }, { "epoch": 41.775449101796404, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1247, "step": 27906 }, { "epoch": 41.77694610778443, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1245, "step": 27907 }, { "epoch": 41.778443113772454, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1235, "step": 27908 }, { "epoch": 41.77994011976048, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1177, "step": 27909 }, { "epoch": 41.7814371257485, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1226, "step": 27910 }, { "epoch": 41.782934131736525, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1177, "step": 27911 }, { "epoch": 41.78443113772455, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 27912 }, { "epoch": 41.785928143712574, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.126, "step": 27913 }, { "epoch": 41.787425149700596, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1234, "step": 27914 }, { "epoch": 41.788922155688624, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1221, "step": 27915 }, { "epoch": 41.790419161676645, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.126, "step": 27916 }, { "epoch": 41.791916167664674, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1229, "step": 27917 }, { "epoch": 41.793413173652695, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1272, "step": 27918 }, { "epoch": 41.794910179640716, "grad_norm": 0.047607421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 27919 }, { "epoch": 41.796407185628745, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1261, "step": 27920 }, { "epoch": 41.797904191616766, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1236, "step": 27921 }, { "epoch": 41.79940119760479, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1272, "step": 27922 }, { "epoch": 41.800898203592816, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1225, "step": 27923 }, { "epoch": 41.80239520958084, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1211, "step": 27924 }, { "epoch": 41.803892215568865, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1238, "step": 27925 }, { "epoch": 41.80538922155689, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1293, "step": 27926 }, { "epoch": 41.80688622754491, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1254, "step": 27927 }, { "epoch": 41.808383233532936, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1239, "step": 27928 }, { "epoch": 41.80988023952096, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1214, "step": 27929 }, { "epoch": 41.81137724550898, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1242, "step": 27930 }, { "epoch": 41.81287425149701, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1247, "step": 27931 }, { "epoch": 41.81437125748503, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1225, "step": 27932 }, { "epoch": 41.81586826347306, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.125, "step": 27933 }, { "epoch": 41.81736526946108, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1192, "step": 27934 }, { "epoch": 41.8188622754491, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1247, "step": 27935 }, { "epoch": 41.82035928143713, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1272, "step": 27936 }, { "epoch": 41.82185628742515, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1253, "step": 27937 }, { "epoch": 41.82335329341317, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1252, "step": 27938 }, { "epoch": 41.8248502994012, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1214, "step": 27939 }, { "epoch": 41.82634730538922, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1274, "step": 27940 }, { "epoch": 41.82784431137725, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.123, "step": 27941 }, { "epoch": 41.82934131736527, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1213, "step": 27942 }, { "epoch": 41.83083832335329, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1236, "step": 27943 }, { "epoch": 41.83233532934132, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1222, "step": 27944 }, { "epoch": 41.83383233532934, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1272, "step": 27945 }, { "epoch": 41.83532934131736, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1247, "step": 27946 }, { "epoch": 41.83682634730539, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1247, "step": 27947 }, { "epoch": 41.83832335329341, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1186, "step": 27948 }, { "epoch": 41.83982035928144, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1251, "step": 27949 }, { "epoch": 41.84131736526946, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1238, "step": 27950 }, { "epoch": 41.84281437125748, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1142, "step": 27951 }, { "epoch": 41.84431137724551, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1276, "step": 27952 }, { "epoch": 41.84580838323353, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1238, "step": 27953 }, { "epoch": 41.84730538922156, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1254, "step": 27954 }, { "epoch": 41.84880239520958, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1189, "step": 27955 }, { "epoch": 41.8502994011976, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1232, "step": 27956 }, { "epoch": 41.85179640718563, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1243, "step": 27957 }, { "epoch": 41.85329341317365, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1245, "step": 27958 }, { "epoch": 41.854790419161674, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1205, "step": 27959 }, { "epoch": 41.8562874251497, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1235, "step": 27960 }, { "epoch": 41.85778443113772, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.124, "step": 27961 }, { "epoch": 41.85928143712575, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.118, "step": 27962 }, { "epoch": 41.86077844311377, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1239, "step": 27963 }, { "epoch": 41.862275449101794, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1229, "step": 27964 }, { "epoch": 41.86377245508982, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1308, "step": 27965 }, { "epoch": 41.865269461077844, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1187, "step": 27966 }, { "epoch": 41.866766467065865, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1233, "step": 27967 }, { "epoch": 41.868263473053894, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1241, "step": 27968 }, { "epoch": 41.869760479041915, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1195, "step": 27969 }, { "epoch": 41.87125748502994, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1201, "step": 27970 }, { "epoch": 41.872754491017965, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1175, "step": 27971 }, { "epoch": 41.874251497005986, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1264, "step": 27972 }, { "epoch": 41.875748502994014, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1251, "step": 27973 }, { "epoch": 41.877245508982035, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1217, "step": 27974 }, { "epoch": 41.87874251497006, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1257, "step": 27975 }, { "epoch": 41.880239520958085, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.125, "step": 27976 }, { "epoch": 41.881736526946106, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.125, "step": 27977 }, { "epoch": 41.883233532934135, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1252, "step": 27978 }, { "epoch": 41.884730538922156, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1207, "step": 27979 }, { "epoch": 41.88622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1186, "step": 27980 }, { "epoch": 41.887724550898206, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1277, "step": 27981 }, { "epoch": 41.88922155688623, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.122, "step": 27982 }, { "epoch": 41.89071856287425, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 27983 }, { "epoch": 41.89221556886228, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1249, "step": 27984 }, { "epoch": 41.8937125748503, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.126, "step": 27985 }, { "epoch": 41.895209580838326, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1174, "step": 27986 }, { "epoch": 41.89670658682635, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.124, "step": 27987 }, { "epoch": 41.89820359281437, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1187, "step": 27988 }, { "epoch": 41.8997005988024, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1299, "step": 27989 }, { "epoch": 41.90119760479042, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1249, "step": 27990 }, { "epoch": 41.90269461077844, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1222, "step": 27991 }, { "epoch": 41.90419161676647, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.117, "step": 27992 }, { "epoch": 41.90568862275449, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1178, "step": 27993 }, { "epoch": 41.90718562874252, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1227, "step": 27994 }, { "epoch": 41.90868263473054, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1249, "step": 27995 }, { "epoch": 41.91017964071856, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1153, "step": 27996 }, { "epoch": 41.91167664670659, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1219, "step": 27997 }, { "epoch": 41.91317365269461, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1202, "step": 27998 }, { "epoch": 41.91467065868264, "grad_norm": 0.0498046875, "learning_rate": 0.0008, "loss": 1.1216, "step": 27999 }, { "epoch": 41.91616766467066, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1256, "step": 28000 }, { "epoch": 41.91766467065868, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1253, "step": 28001 }, { "epoch": 41.91916167664671, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1264, "step": 28002 }, { "epoch": 41.92065868263473, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1209, "step": 28003 }, { "epoch": 41.92215568862275, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1247, "step": 28004 }, { "epoch": 41.92365269461078, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1216, "step": 28005 }, { "epoch": 41.9251497005988, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1265, "step": 28006 }, { "epoch": 41.92664670658683, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1245, "step": 28007 }, { "epoch": 41.92814371257485, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1215, "step": 28008 }, { "epoch": 41.92964071856287, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1257, "step": 28009 }, { "epoch": 41.9311377245509, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.122, "step": 28010 }, { "epoch": 41.93263473053892, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1231, "step": 28011 }, { "epoch": 41.93413173652694, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1209, "step": 28012 }, { "epoch": 41.93562874251497, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1257, "step": 28013 }, { "epoch": 41.93712574850299, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1213, "step": 28014 }, { "epoch": 41.93862275449102, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1265, "step": 28015 }, { "epoch": 41.94011976047904, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1176, "step": 28016 }, { "epoch": 41.941616766467064, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1235, "step": 28017 }, { "epoch": 41.94311377245509, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.12, "step": 28018 }, { "epoch": 41.94461077844311, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.119, "step": 28019 }, { "epoch": 41.946107784431135, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1244, "step": 28020 }, { "epoch": 41.94760479041916, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1222, "step": 28021 }, { "epoch": 41.949101796407184, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.121, "step": 28022 }, { "epoch": 41.95059880239521, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1253, "step": 28023 }, { "epoch": 41.952095808383234, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1227, "step": 28024 }, { "epoch": 41.953592814371255, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1221, "step": 28025 }, { "epoch": 41.955089820359284, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1261, "step": 28026 }, { "epoch": 41.956586826347305, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1148, "step": 28027 }, { "epoch": 41.958083832335326, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1275, "step": 28028 }, { "epoch": 41.959580838323355, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1282, "step": 28029 }, { "epoch": 41.961077844311376, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1241, "step": 28030 }, { "epoch": 41.962574850299404, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1284, "step": 28031 }, { "epoch": 41.964071856287426, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1229, "step": 28032 }, { "epoch": 41.96556886227545, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1216, "step": 28033 }, { "epoch": 41.967065868263475, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1255, "step": 28034 }, { "epoch": 41.9685628742515, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1173, "step": 28035 }, { "epoch": 41.97005988023952, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1173, "step": 28036 }, { "epoch": 41.971556886227546, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1252, "step": 28037 }, { "epoch": 41.97305389221557, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1257, "step": 28038 }, { "epoch": 41.974550898203596, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1201, "step": 28039 }, { "epoch": 41.97604790419162, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1184, "step": 28040 }, { "epoch": 41.97754491017964, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1323, "step": 28041 }, { "epoch": 41.97904191616767, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1231, "step": 28042 }, { "epoch": 41.98053892215569, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1174, "step": 28043 }, { "epoch": 41.98203592814371, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1268, "step": 28044 }, { "epoch": 41.98353293413174, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1226, "step": 28045 }, { "epoch": 41.98502994011976, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1206, "step": 28046 }, { "epoch": 41.98652694610779, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.12, "step": 28047 }, { "epoch": 41.98802395209581, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1172, "step": 28048 }, { "epoch": 41.98952095808383, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1169, "step": 28049 }, { "epoch": 41.99101796407186, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1212, "step": 28050 }, { "epoch": 41.99251497005988, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1217, "step": 28051 }, { "epoch": 41.9940119760479, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1193, "step": 28052 }, { "epoch": 41.99550898203593, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1245, "step": 28053 }, { "epoch": 41.99700598802395, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1226, "step": 28054 }, { "epoch": 41.99850299401198, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1221, "step": 28055 }, { "epoch": 42.0, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1214, "step": 28056 }, { "epoch": 42.00149700598802, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1221, "step": 28057 }, { "epoch": 42.00299401197605, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1253, "step": 28058 }, { "epoch": 42.00449101796407, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1184, "step": 28059 }, { "epoch": 42.0059880239521, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1163, "step": 28060 }, { "epoch": 42.00748502994012, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1243, "step": 28061 }, { "epoch": 42.00898203592814, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1193, "step": 28062 }, { "epoch": 42.01047904191617, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1217, "step": 28063 }, { "epoch": 42.01197604790419, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1204, "step": 28064 }, { "epoch": 42.01347305389221, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.121, "step": 28065 }, { "epoch": 42.01497005988024, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1263, "step": 28066 }, { "epoch": 42.01646706586826, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1186, "step": 28067 }, { "epoch": 42.01796407185629, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1211, "step": 28068 }, { "epoch": 42.01946107784431, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1248, "step": 28069 }, { "epoch": 42.02095808383233, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1219, "step": 28070 }, { "epoch": 42.02245508982036, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1322, "step": 28071 }, { "epoch": 42.02395209580838, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1297, "step": 28072 }, { "epoch": 42.025449101796404, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1163, "step": 28073 }, { "epoch": 42.02694610778443, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1242, "step": 28074 }, { "epoch": 42.028443113772454, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1239, "step": 28075 }, { "epoch": 42.02994011976048, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1228, "step": 28076 }, { "epoch": 42.0314371257485, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1199, "step": 28077 }, { "epoch": 42.032934131736525, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1236, "step": 28078 }, { "epoch": 42.03443113772455, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1256, "step": 28079 }, { "epoch": 42.035928143712574, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1175, "step": 28080 }, { "epoch": 42.037425149700596, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.121, "step": 28081 }, { "epoch": 42.038922155688624, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1228, "step": 28082 }, { "epoch": 42.040419161676645, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1186, "step": 28083 }, { "epoch": 42.041916167664674, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 28084 }, { "epoch": 42.043413173652695, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1183, "step": 28085 }, { "epoch": 42.044910179640716, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.117, "step": 28086 }, { "epoch": 42.046407185628745, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1222, "step": 28087 }, { "epoch": 42.047904191616766, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1204, "step": 28088 }, { "epoch": 42.04940119760479, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1217, "step": 28089 }, { "epoch": 42.050898203592816, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1241, "step": 28090 }, { "epoch": 42.05239520958084, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1153, "step": 28091 }, { "epoch": 42.053892215568865, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1172, "step": 28092 }, { "epoch": 42.05538922155689, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1267, "step": 28093 }, { "epoch": 42.05688622754491, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1215, "step": 28094 }, { "epoch": 42.058383233532936, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1239, "step": 28095 }, { "epoch": 42.05988023952096, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1256, "step": 28096 }, { "epoch": 42.06137724550898, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1239, "step": 28097 }, { "epoch": 42.06287425149701, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1235, "step": 28098 }, { "epoch": 42.06437125748503, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1236, "step": 28099 }, { "epoch": 42.06586826347306, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.123, "step": 28100 }, { "epoch": 42.06736526946108, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1246, "step": 28101 }, { "epoch": 42.0688622754491, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1266, "step": 28102 }, { "epoch": 42.07035928143713, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1238, "step": 28103 }, { "epoch": 42.07185628742515, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1193, "step": 28104 }, { "epoch": 42.07335329341317, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1217, "step": 28105 }, { "epoch": 42.0748502994012, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1292, "step": 28106 }, { "epoch": 42.07634730538922, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1219, "step": 28107 }, { "epoch": 42.07784431137725, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1154, "step": 28108 }, { "epoch": 42.07934131736527, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1219, "step": 28109 }, { "epoch": 42.08083832335329, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1246, "step": 28110 }, { "epoch": 42.08233532934132, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.121, "step": 28111 }, { "epoch": 42.08383233532934, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1207, "step": 28112 }, { "epoch": 42.08532934131737, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.128, "step": 28113 }, { "epoch": 42.08682634730539, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1223, "step": 28114 }, { "epoch": 42.08832335329341, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.117, "step": 28115 }, { "epoch": 42.08982035928144, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1201, "step": 28116 }, { "epoch": 42.09131736526946, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1235, "step": 28117 }, { "epoch": 42.09281437125748, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1244, "step": 28118 }, { "epoch": 42.09431137724551, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1237, "step": 28119 }, { "epoch": 42.09580838323353, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1271, "step": 28120 }, { "epoch": 42.09730538922156, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1289, "step": 28121 }, { "epoch": 42.09880239520958, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1221, "step": 28122 }, { "epoch": 42.1002994011976, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1319, "step": 28123 }, { "epoch": 42.10179640718563, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1217, "step": 28124 }, { "epoch": 42.10329341317365, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1201, "step": 28125 }, { "epoch": 42.104790419161674, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1241, "step": 28126 }, { "epoch": 42.1062874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1292, "step": 28127 }, { "epoch": 42.10778443113772, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1258, "step": 28128 }, { "epoch": 42.10928143712575, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1283, "step": 28129 }, { "epoch": 42.11077844311377, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 28130 }, { "epoch": 42.112275449101794, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1188, "step": 28131 }, { "epoch": 42.11377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 28132 }, { "epoch": 42.115269461077844, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1207, "step": 28133 }, { "epoch": 42.116766467065865, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1261, "step": 28134 }, { "epoch": 42.118263473053894, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1296, "step": 28135 }, { "epoch": 42.119760479041915, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1253, "step": 28136 }, { "epoch": 42.12125748502994, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1204, "step": 28137 }, { "epoch": 42.122754491017965, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1221, "step": 28138 }, { "epoch": 42.124251497005986, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1215, "step": 28139 }, { "epoch": 42.125748502994014, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.124, "step": 28140 }, { "epoch": 42.127245508982035, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1222, "step": 28141 }, { "epoch": 42.12874251497006, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1257, "step": 28142 }, { "epoch": 42.130239520958085, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1168, "step": 28143 }, { "epoch": 42.131736526946106, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.124, "step": 28144 }, { "epoch": 42.133233532934135, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1204, "step": 28145 }, { "epoch": 42.134730538922156, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1239, "step": 28146 }, { "epoch": 42.13622754491018, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.122, "step": 28147 }, { "epoch": 42.137724550898206, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1276, "step": 28148 }, { "epoch": 42.13922155688623, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1209, "step": 28149 }, { "epoch": 42.14071856287425, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1315, "step": 28150 }, { "epoch": 42.14221556886228, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1249, "step": 28151 }, { "epoch": 42.1437125748503, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1223, "step": 28152 }, { "epoch": 42.145209580838326, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.125, "step": 28153 }, { "epoch": 42.14670658682635, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1181, "step": 28154 }, { "epoch": 42.14820359281437, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1197, "step": 28155 }, { "epoch": 42.1497005988024, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1278, "step": 28156 }, { "epoch": 42.15119760479042, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1308, "step": 28157 }, { "epoch": 42.15269461077844, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1245, "step": 28158 }, { "epoch": 42.15419161676647, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1241, "step": 28159 }, { "epoch": 42.15568862275449, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1136, "step": 28160 }, { "epoch": 42.15718562874252, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 28161 }, { "epoch": 42.15868263473054, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1202, "step": 28162 }, { "epoch": 42.16017964071856, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1164, "step": 28163 }, { "epoch": 42.16167664670659, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1181, "step": 28164 }, { "epoch": 42.16317365269461, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1212, "step": 28165 }, { "epoch": 42.16467065868263, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1256, "step": 28166 }, { "epoch": 42.16616766467066, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1244, "step": 28167 }, { "epoch": 42.16766467065868, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1251, "step": 28168 }, { "epoch": 42.16916167664671, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1277, "step": 28169 }, { "epoch": 42.17065868263473, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1268, "step": 28170 }, { "epoch": 42.17215568862275, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1213, "step": 28171 }, { "epoch": 42.17365269461078, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1251, "step": 28172 }, { "epoch": 42.1751497005988, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1271, "step": 28173 }, { "epoch": 42.17664670658683, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1248, "step": 28174 }, { "epoch": 42.17814371257485, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1222, "step": 28175 }, { "epoch": 42.17964071856287, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1167, "step": 28176 }, { "epoch": 42.1811377245509, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1224, "step": 28177 }, { "epoch": 42.18263473053892, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1212, "step": 28178 }, { "epoch": 42.18413173652694, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1244, "step": 28179 }, { "epoch": 42.18562874251497, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1262, "step": 28180 }, { "epoch": 42.18712574850299, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1294, "step": 28181 }, { "epoch": 42.18862275449102, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1233, "step": 28182 }, { "epoch": 42.19011976047904, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.12, "step": 28183 }, { "epoch": 42.191616766467064, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1266, "step": 28184 }, { "epoch": 42.19311377245509, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1251, "step": 28185 }, { "epoch": 42.19461077844311, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1221, "step": 28186 }, { "epoch": 42.196107784431135, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.123, "step": 28187 }, { "epoch": 42.19760479041916, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1251, "step": 28188 }, { "epoch": 42.199101796407184, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1201, "step": 28189 }, { "epoch": 42.20059880239521, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1208, "step": 28190 }, { "epoch": 42.202095808383234, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.129, "step": 28191 }, { "epoch": 42.203592814371255, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1282, "step": 28192 }, { "epoch": 42.205089820359284, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1249, "step": 28193 }, { "epoch": 42.206586826347305, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1255, "step": 28194 }, { "epoch": 42.208083832335326, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1196, "step": 28195 }, { "epoch": 42.209580838323355, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1268, "step": 28196 }, { "epoch": 42.211077844311376, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1214, "step": 28197 }, { "epoch": 42.212574850299404, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 28198 }, { "epoch": 42.214071856287426, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1194, "step": 28199 }, { "epoch": 42.21556886227545, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1242, "step": 28200 }, { "epoch": 42.217065868263475, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1239, "step": 28201 }, { "epoch": 42.2185628742515, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1201, "step": 28202 }, { "epoch": 42.22005988023952, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1228, "step": 28203 }, { "epoch": 42.221556886227546, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1242, "step": 28204 }, { "epoch": 42.22305389221557, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1229, "step": 28205 }, { "epoch": 42.224550898203596, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1276, "step": 28206 }, { "epoch": 42.22604790419162, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1233, "step": 28207 }, { "epoch": 42.22754491017964, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1167, "step": 28208 }, { "epoch": 42.22904191616767, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1169, "step": 28209 }, { "epoch": 42.23053892215569, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1219, "step": 28210 }, { "epoch": 42.23203592814371, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1193, "step": 28211 }, { "epoch": 42.23353293413174, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1254, "step": 28212 }, { "epoch": 42.23502994011976, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1251, "step": 28213 }, { "epoch": 42.23652694610779, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1272, "step": 28214 }, { "epoch": 42.23802395209581, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1237, "step": 28215 }, { "epoch": 42.23952095808383, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1324, "step": 28216 }, { "epoch": 42.24101796407186, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1229, "step": 28217 }, { "epoch": 42.24251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1189, "step": 28218 }, { "epoch": 42.2440119760479, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1182, "step": 28219 }, { "epoch": 42.24550898203593, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1243, "step": 28220 }, { "epoch": 42.24700598802395, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.121, "step": 28221 }, { "epoch": 42.24850299401198, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1226, "step": 28222 }, { "epoch": 42.25, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.126, "step": 28223 }, { "epoch": 42.25149700598802, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1265, "step": 28224 }, { "epoch": 42.25299401197605, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1197, "step": 28225 }, { "epoch": 42.25449101796407, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.128, "step": 28226 }, { "epoch": 42.2559880239521, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1212, "step": 28227 }, { "epoch": 42.25748502994012, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1191, "step": 28228 }, { "epoch": 42.25898203592814, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1227, "step": 28229 }, { "epoch": 42.26047904191617, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1163, "step": 28230 }, { "epoch": 42.26197604790419, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1179, "step": 28231 }, { "epoch": 42.26347305389221, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1301, "step": 28232 }, { "epoch": 42.26497005988024, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1221, "step": 28233 }, { "epoch": 42.26646706586826, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1316, "step": 28234 }, { "epoch": 42.26796407185629, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1232, "step": 28235 }, { "epoch": 42.26946107784431, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1269, "step": 28236 }, { "epoch": 42.27095808383233, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.133, "step": 28237 }, { "epoch": 42.27245508982036, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1255, "step": 28238 }, { "epoch": 42.27395209580838, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1211, "step": 28239 }, { "epoch": 42.275449101796404, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.122, "step": 28240 }, { "epoch": 42.27694610778443, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.123, "step": 28241 }, { "epoch": 42.278443113772454, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1197, "step": 28242 }, { "epoch": 42.27994011976048, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1216, "step": 28243 }, { "epoch": 42.2814371257485, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.123, "step": 28244 }, { "epoch": 42.282934131736525, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1304, "step": 28245 }, { "epoch": 42.28443113772455, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1235, "step": 28246 }, { "epoch": 42.285928143712574, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1223, "step": 28247 }, { "epoch": 42.287425149700596, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1228, "step": 28248 }, { "epoch": 42.288922155688624, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1208, "step": 28249 }, { "epoch": 42.290419161676645, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1222, "step": 28250 }, { "epoch": 42.291916167664674, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1203, "step": 28251 }, { "epoch": 42.293413173652695, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1251, "step": 28252 }, { "epoch": 42.294910179640716, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1231, "step": 28253 }, { "epoch": 42.296407185628745, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1234, "step": 28254 }, { "epoch": 42.297904191616766, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1244, "step": 28255 }, { "epoch": 42.29940119760479, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1233, "step": 28256 }, { "epoch": 42.300898203592816, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1293, "step": 28257 }, { "epoch": 42.30239520958084, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1245, "step": 28258 }, { "epoch": 42.303892215568865, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1206, "step": 28259 }, { "epoch": 42.30538922155689, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1139, "step": 28260 }, { "epoch": 42.30688622754491, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 28261 }, { "epoch": 42.308383233532936, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1224, "step": 28262 }, { "epoch": 42.30988023952096, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1278, "step": 28263 }, { "epoch": 42.31137724550898, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1206, "step": 28264 }, { "epoch": 42.31287425149701, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.12, "step": 28265 }, { "epoch": 42.31437125748503, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1248, "step": 28266 }, { "epoch": 42.31586826347306, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1196, "step": 28267 }, { "epoch": 42.31736526946108, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1236, "step": 28268 }, { "epoch": 42.3188622754491, "grad_norm": 0.04833984375, "learning_rate": 0.0008, "loss": 1.1261, "step": 28269 }, { "epoch": 42.32035928143713, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1224, "step": 28270 }, { "epoch": 42.32185628742515, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1245, "step": 28271 }, { "epoch": 42.32335329341317, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1254, "step": 28272 }, { "epoch": 42.3248502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1225, "step": 28273 }, { "epoch": 42.32634730538922, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 28274 }, { "epoch": 42.32784431137725, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1281, "step": 28275 }, { "epoch": 42.32934131736527, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1172, "step": 28276 }, { "epoch": 42.33083832335329, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1215, "step": 28277 }, { "epoch": 42.33233532934132, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.112, "step": 28278 }, { "epoch": 42.33383233532934, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1281, "step": 28279 }, { "epoch": 42.33532934131736, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1272, "step": 28280 }, { "epoch": 42.33682634730539, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1179, "step": 28281 }, { "epoch": 42.33832335329341, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1279, "step": 28282 }, { "epoch": 42.33982035928144, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1282, "step": 28283 }, { "epoch": 42.34131736526946, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1265, "step": 28284 }, { "epoch": 42.34281437125748, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1209, "step": 28285 }, { "epoch": 42.34431137724551, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1256, "step": 28286 }, { "epoch": 42.34580838323353, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1207, "step": 28287 }, { "epoch": 42.34730538922156, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1344, "step": 28288 }, { "epoch": 42.34880239520958, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1187, "step": 28289 }, { "epoch": 42.3502994011976, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1248, "step": 28290 }, { "epoch": 42.35179640718563, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1222, "step": 28291 }, { "epoch": 42.35329341317365, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1104, "step": 28292 }, { "epoch": 42.354790419161674, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 28293 }, { "epoch": 42.3562874251497, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1278, "step": 28294 }, { "epoch": 42.35778443113772, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1263, "step": 28295 }, { "epoch": 42.35928143712575, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1211, "step": 28296 }, { "epoch": 42.36077844311377, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1226, "step": 28297 }, { "epoch": 42.362275449101794, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1269, "step": 28298 }, { "epoch": 42.36377245508982, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1249, "step": 28299 }, { "epoch": 42.365269461077844, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1233, "step": 28300 }, { "epoch": 42.366766467065865, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1284, "step": 28301 }, { "epoch": 42.368263473053894, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1298, "step": 28302 }, { "epoch": 42.369760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1246, "step": 28303 }, { "epoch": 42.37125748502994, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1237, "step": 28304 }, { "epoch": 42.372754491017965, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1272, "step": 28305 }, { "epoch": 42.374251497005986, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1242, "step": 28306 }, { "epoch": 42.375748502994014, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1249, "step": 28307 }, { "epoch": 42.377245508982035, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1246, "step": 28308 }, { "epoch": 42.37874251497006, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1187, "step": 28309 }, { "epoch": 42.380239520958085, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1328, "step": 28310 }, { "epoch": 42.381736526946106, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1173, "step": 28311 }, { "epoch": 42.383233532934135, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1243, "step": 28312 }, { "epoch": 42.384730538922156, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1279, "step": 28313 }, { "epoch": 42.38622754491018, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1186, "step": 28314 }, { "epoch": 42.387724550898206, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1147, "step": 28315 }, { "epoch": 42.38922155688623, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.121, "step": 28316 }, { "epoch": 42.39071856287425, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 28317 }, { "epoch": 42.39221556886228, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1181, "step": 28318 }, { "epoch": 42.3937125748503, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1212, "step": 28319 }, { "epoch": 42.395209580838326, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1293, "step": 28320 }, { "epoch": 42.39670658682635, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1253, "step": 28321 }, { "epoch": 42.39820359281437, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1226, "step": 28322 }, { "epoch": 42.3997005988024, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1211, "step": 28323 }, { "epoch": 42.40119760479042, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1208, "step": 28324 }, { "epoch": 42.40269461077844, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1219, "step": 28325 }, { "epoch": 42.40419161676647, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.13, "step": 28326 }, { "epoch": 42.40568862275449, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1225, "step": 28327 }, { "epoch": 42.40718562874252, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1238, "step": 28328 }, { "epoch": 42.40868263473054, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1203, "step": 28329 }, { "epoch": 42.41017964071856, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1282, "step": 28330 }, { "epoch": 42.41167664670659, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1189, "step": 28331 }, { "epoch": 42.41317365269461, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.126, "step": 28332 }, { "epoch": 42.41467065868264, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1252, "step": 28333 }, { "epoch": 42.41616766467066, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1239, "step": 28334 }, { "epoch": 42.41766467065868, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1246, "step": 28335 }, { "epoch": 42.41916167664671, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1224, "step": 28336 }, { "epoch": 42.42065868263473, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.119, "step": 28337 }, { "epoch": 42.42215568862275, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1197, "step": 28338 }, { "epoch": 42.42365269461078, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1262, "step": 28339 }, { "epoch": 42.4251497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1234, "step": 28340 }, { "epoch": 42.42664670658683, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1273, "step": 28341 }, { "epoch": 42.42814371257485, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1187, "step": 28342 }, { "epoch": 42.42964071856287, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1268, "step": 28343 }, { "epoch": 42.4311377245509, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1242, "step": 28344 }, { "epoch": 42.43263473053892, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1199, "step": 28345 }, { "epoch": 42.43413173652694, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1219, "step": 28346 }, { "epoch": 42.43562874251497, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1214, "step": 28347 }, { "epoch": 42.43712574850299, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1242, "step": 28348 }, { "epoch": 42.43862275449102, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1184, "step": 28349 }, { "epoch": 42.44011976047904, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1136, "step": 28350 }, { "epoch": 42.441616766467064, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1244, "step": 28351 }, { "epoch": 42.44311377245509, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.124, "step": 28352 }, { "epoch": 42.44461077844311, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1171, "step": 28353 }, { "epoch": 42.446107784431135, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1244, "step": 28354 }, { "epoch": 42.44760479041916, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1243, "step": 28355 }, { "epoch": 42.449101796407184, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1241, "step": 28356 }, { "epoch": 42.45059880239521, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1197, "step": 28357 }, { "epoch": 42.452095808383234, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 28358 }, { "epoch": 42.453592814371255, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1209, "step": 28359 }, { "epoch": 42.455089820359284, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1182, "step": 28360 }, { "epoch": 42.456586826347305, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1271, "step": 28361 }, { "epoch": 42.458083832335326, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.119, "step": 28362 }, { "epoch": 42.459580838323355, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1189, "step": 28363 }, { "epoch": 42.461077844311376, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1219, "step": 28364 }, { "epoch": 42.462574850299404, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1247, "step": 28365 }, { "epoch": 42.464071856287426, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1206, "step": 28366 }, { "epoch": 42.46556886227545, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1214, "step": 28367 }, { "epoch": 42.467065868263475, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1199, "step": 28368 }, { "epoch": 42.4685628742515, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1344, "step": 28369 }, { "epoch": 42.47005988023952, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.118, "step": 28370 }, { "epoch": 42.471556886227546, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1224, "step": 28371 }, { "epoch": 42.47305389221557, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1216, "step": 28372 }, { "epoch": 42.474550898203596, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1275, "step": 28373 }, { "epoch": 42.47604790419162, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1229, "step": 28374 }, { "epoch": 42.47754491017964, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.124, "step": 28375 }, { "epoch": 42.47904191616767, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1238, "step": 28376 }, { "epoch": 42.48053892215569, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.129, "step": 28377 }, { "epoch": 42.48203592814371, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1221, "step": 28378 }, { "epoch": 42.48353293413174, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1187, "step": 28379 }, { "epoch": 42.48502994011976, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1223, "step": 28380 }, { "epoch": 42.48652694610779, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1217, "step": 28381 }, { "epoch": 42.48802395209581, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1321, "step": 28382 }, { "epoch": 42.48952095808383, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1227, "step": 28383 }, { "epoch": 42.49101796407186, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1247, "step": 28384 }, { "epoch": 42.49251497005988, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1195, "step": 28385 }, { "epoch": 42.4940119760479, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1259, "step": 28386 }, { "epoch": 42.49550898203593, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.126, "step": 28387 }, { "epoch": 42.49700598802395, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1266, "step": 28388 }, { "epoch": 42.49850299401198, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1247, "step": 28389 }, { "epoch": 42.5, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1255, "step": 28390 }, { "epoch": 42.50149700598802, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1267, "step": 28391 }, { "epoch": 42.50299401197605, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1161, "step": 28392 }, { "epoch": 42.50449101796407, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1242, "step": 28393 }, { "epoch": 42.5059880239521, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1219, "step": 28394 }, { "epoch": 42.50748502994012, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1263, "step": 28395 }, { "epoch": 42.50898203592814, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1176, "step": 28396 }, { "epoch": 42.51047904191617, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1243, "step": 28397 }, { "epoch": 42.51197604790419, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1268, "step": 28398 }, { "epoch": 42.51347305389221, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1222, "step": 28399 }, { "epoch": 42.51497005988024, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1216, "step": 28400 }, { "epoch": 42.51646706586826, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1218, "step": 28401 }, { "epoch": 42.51796407185629, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1187, "step": 28402 }, { "epoch": 42.51946107784431, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.122, "step": 28403 }, { "epoch": 42.52095808383233, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1153, "step": 28404 }, { "epoch": 42.52245508982036, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1206, "step": 28405 }, { "epoch": 42.52395209580838, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1268, "step": 28406 }, { "epoch": 42.525449101796404, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1172, "step": 28407 }, { "epoch": 42.52694610778443, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1258, "step": 28408 }, { "epoch": 42.528443113772454, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1236, "step": 28409 }, { "epoch": 42.52994011976048, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1206, "step": 28410 }, { "epoch": 42.5314371257485, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1221, "step": 28411 }, { "epoch": 42.532934131736525, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1225, "step": 28412 }, { "epoch": 42.53443113772455, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1274, "step": 28413 }, { "epoch": 42.535928143712574, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1344, "step": 28414 }, { "epoch": 42.537425149700596, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1227, "step": 28415 }, { "epoch": 42.538922155688624, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1223, "step": 28416 }, { "epoch": 42.540419161676645, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1211, "step": 28417 }, { "epoch": 42.541916167664674, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1225, "step": 28418 }, { "epoch": 42.543413173652695, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1299, "step": 28419 }, { "epoch": 42.544910179640716, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1186, "step": 28420 }, { "epoch": 42.546407185628745, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1254, "step": 28421 }, { "epoch": 42.547904191616766, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1237, "step": 28422 }, { "epoch": 42.54940119760479, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1282, "step": 28423 }, { "epoch": 42.550898203592816, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1287, "step": 28424 }, { "epoch": 42.55239520958084, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1206, "step": 28425 }, { "epoch": 42.553892215568865, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1179, "step": 28426 }, { "epoch": 42.55538922155689, "grad_norm": 0.04443359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 28427 }, { "epoch": 42.55688622754491, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.121, "step": 28428 }, { "epoch": 42.558383233532936, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1196, "step": 28429 }, { "epoch": 42.55988023952096, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1189, "step": 28430 }, { "epoch": 42.56137724550898, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.12, "step": 28431 }, { "epoch": 42.56287425149701, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1221, "step": 28432 }, { "epoch": 42.56437125748503, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1162, "step": 28433 }, { "epoch": 42.56586826347306, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1235, "step": 28434 }, { "epoch": 42.56736526946108, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1205, "step": 28435 }, { "epoch": 42.5688622754491, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1232, "step": 28436 }, { "epoch": 42.57035928143713, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1254, "step": 28437 }, { "epoch": 42.57185628742515, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.12, "step": 28438 }, { "epoch": 42.57335329341317, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1248, "step": 28439 }, { "epoch": 42.5748502994012, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1224, "step": 28440 }, { "epoch": 42.57634730538922, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1222, "step": 28441 }, { "epoch": 42.57784431137725, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1227, "step": 28442 }, { "epoch": 42.57934131736527, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1261, "step": 28443 }, { "epoch": 42.58083832335329, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1267, "step": 28444 }, { "epoch": 42.58233532934132, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1219, "step": 28445 }, { "epoch": 42.58383233532934, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1231, "step": 28446 }, { "epoch": 42.58532934131736, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1238, "step": 28447 }, { "epoch": 42.58682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1311, "step": 28448 }, { "epoch": 42.58832335329341, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1187, "step": 28449 }, { "epoch": 42.58982035928144, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1212, "step": 28450 }, { "epoch": 42.59131736526946, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1325, "step": 28451 }, { "epoch": 42.59281437125748, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1147, "step": 28452 }, { "epoch": 42.59431137724551, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1239, "step": 28453 }, { "epoch": 42.59580838323353, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1216, "step": 28454 }, { "epoch": 42.59730538922156, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1272, "step": 28455 }, { "epoch": 42.59880239520958, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1248, "step": 28456 }, { "epoch": 42.6002994011976, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1233, "step": 28457 }, { "epoch": 42.60179640718563, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1261, "step": 28458 }, { "epoch": 42.60329341317365, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1215, "step": 28459 }, { "epoch": 42.604790419161674, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1156, "step": 28460 }, { "epoch": 42.6062874251497, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1225, "step": 28461 }, { "epoch": 42.60778443113772, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1234, "step": 28462 }, { "epoch": 42.60928143712575, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1221, "step": 28463 }, { "epoch": 42.61077844311377, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.114, "step": 28464 }, { "epoch": 42.612275449101794, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 28465 }, { "epoch": 42.61377245508982, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.121, "step": 28466 }, { "epoch": 42.615269461077844, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1263, "step": 28467 }, { "epoch": 42.616766467065865, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1219, "step": 28468 }, { "epoch": 42.618263473053894, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1232, "step": 28469 }, { "epoch": 42.619760479041915, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1282, "step": 28470 }, { "epoch": 42.62125748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1266, "step": 28471 }, { "epoch": 42.622754491017965, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1217, "step": 28472 }, { "epoch": 42.624251497005986, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1281, "step": 28473 }, { "epoch": 42.625748502994014, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1252, "step": 28474 }, { "epoch": 42.627245508982035, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1303, "step": 28475 }, { "epoch": 42.62874251497006, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1181, "step": 28476 }, { "epoch": 42.630239520958085, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1292, "step": 28477 }, { "epoch": 42.631736526946106, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1184, "step": 28478 }, { "epoch": 42.633233532934135, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1197, "step": 28479 }, { "epoch": 42.634730538922156, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1279, "step": 28480 }, { "epoch": 42.63622754491018, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 28481 }, { "epoch": 42.637724550898206, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1275, "step": 28482 }, { "epoch": 42.63922155688623, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1246, "step": 28483 }, { "epoch": 42.64071856287425, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1248, "step": 28484 }, { "epoch": 42.64221556886228, "grad_norm": 0.045166015625, "learning_rate": 0.0008, "loss": 1.1205, "step": 28485 }, { "epoch": 42.6437125748503, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.121, "step": 28486 }, { "epoch": 42.645209580838326, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1248, "step": 28487 }, { "epoch": 42.64670658682635, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1226, "step": 28488 }, { "epoch": 42.64820359281437, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1183, "step": 28489 }, { "epoch": 42.6497005988024, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1231, "step": 28490 }, { "epoch": 42.65119760479042, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.129, "step": 28491 }, { "epoch": 42.65269461077844, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1219, "step": 28492 }, { "epoch": 42.65419161676647, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1255, "step": 28493 }, { "epoch": 42.65568862275449, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.125, "step": 28494 }, { "epoch": 42.65718562874252, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1253, "step": 28495 }, { "epoch": 42.65868263473054, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1228, "step": 28496 }, { "epoch": 42.66017964071856, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1276, "step": 28497 }, { "epoch": 42.66167664670659, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1233, "step": 28498 }, { "epoch": 42.66317365269461, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1305, "step": 28499 }, { "epoch": 42.66467065868264, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1247, "step": 28500 }, { "epoch": 42.66616766467066, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1271, "step": 28501 }, { "epoch": 42.66766467065868, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1166, "step": 28502 }, { "epoch": 42.66916167664671, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.126, "step": 28503 }, { "epoch": 42.67065868263473, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1219, "step": 28504 }, { "epoch": 42.67215568862275, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1246, "step": 28505 }, { "epoch": 42.67365269461078, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1243, "step": 28506 }, { "epoch": 42.6751497005988, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1194, "step": 28507 }, { "epoch": 42.67664670658683, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1194, "step": 28508 }, { "epoch": 42.67814371257485, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1245, "step": 28509 }, { "epoch": 42.67964071856287, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1239, "step": 28510 }, { "epoch": 42.6811377245509, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1332, "step": 28511 }, { "epoch": 42.68263473053892, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1307, "step": 28512 }, { "epoch": 42.68413173652694, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1184, "step": 28513 }, { "epoch": 42.68562874251497, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1186, "step": 28514 }, { "epoch": 42.68712574850299, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1202, "step": 28515 }, { "epoch": 42.68862275449102, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1165, "step": 28516 }, { "epoch": 42.69011976047904, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1304, "step": 28517 }, { "epoch": 42.691616766467064, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1203, "step": 28518 }, { "epoch": 42.69311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1217, "step": 28519 }, { "epoch": 42.69461077844311, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1244, "step": 28520 }, { "epoch": 42.696107784431135, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.12, "step": 28521 }, { "epoch": 42.69760479041916, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1228, "step": 28522 }, { "epoch": 42.699101796407184, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1235, "step": 28523 }, { "epoch": 42.70059880239521, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1232, "step": 28524 }, { "epoch": 42.702095808383234, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1294, "step": 28525 }, { "epoch": 42.703592814371255, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1206, "step": 28526 }, { "epoch": 42.705089820359284, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1277, "step": 28527 }, { "epoch": 42.706586826347305, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1343, "step": 28528 }, { "epoch": 42.708083832335326, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1213, "step": 28529 }, { "epoch": 42.709580838323355, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.119, "step": 28530 }, { "epoch": 42.711077844311376, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1209, "step": 28531 }, { "epoch": 42.712574850299404, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1247, "step": 28532 }, { "epoch": 42.714071856287426, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 28533 }, { "epoch": 42.71556886227545, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1188, "step": 28534 }, { "epoch": 42.717065868263475, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1241, "step": 28535 }, { "epoch": 42.7185628742515, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1245, "step": 28536 }, { "epoch": 42.72005988023952, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1204, "step": 28537 }, { "epoch": 42.721556886227546, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1262, "step": 28538 }, { "epoch": 42.72305389221557, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1209, "step": 28539 }, { "epoch": 42.724550898203596, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1208, "step": 28540 }, { "epoch": 42.72604790419162, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1296, "step": 28541 }, { "epoch": 42.72754491017964, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1165, "step": 28542 }, { "epoch": 42.72904191616767, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1264, "step": 28543 }, { "epoch": 42.73053892215569, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1212, "step": 28544 }, { "epoch": 42.73203592814371, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1236, "step": 28545 }, { "epoch": 42.73353293413174, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1237, "step": 28546 }, { "epoch": 42.73502994011976, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1278, "step": 28547 }, { "epoch": 42.73652694610779, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1252, "step": 28548 }, { "epoch": 42.73802395209581, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1234, "step": 28549 }, { "epoch": 42.73952095808383, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1208, "step": 28550 }, { "epoch": 42.74101796407186, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1217, "step": 28551 }, { "epoch": 42.74251497005988, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1249, "step": 28552 }, { "epoch": 42.7440119760479, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1223, "step": 28553 }, { "epoch": 42.74550898203593, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1206, "step": 28554 }, { "epoch": 42.74700598802395, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1218, "step": 28555 }, { "epoch": 42.74850299401198, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1254, "step": 28556 }, { "epoch": 42.75, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1258, "step": 28557 }, { "epoch": 42.75149700598802, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1248, "step": 28558 }, { "epoch": 42.75299401197605, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1255, "step": 28559 }, { "epoch": 42.75449101796407, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1177, "step": 28560 }, { "epoch": 42.7559880239521, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.13, "step": 28561 }, { "epoch": 42.75748502994012, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.12, "step": 28562 }, { "epoch": 42.75898203592814, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1208, "step": 28563 }, { "epoch": 42.76047904191617, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1257, "step": 28564 }, { "epoch": 42.76197604790419, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1256, "step": 28565 }, { "epoch": 42.76347305389221, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1209, "step": 28566 }, { "epoch": 42.76497005988024, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1226, "step": 28567 }, { "epoch": 42.76646706586826, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1256, "step": 28568 }, { "epoch": 42.76796407185629, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1246, "step": 28569 }, { "epoch": 42.76946107784431, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.123, "step": 28570 }, { "epoch": 42.77095808383233, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1185, "step": 28571 }, { "epoch": 42.77245508982036, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1258, "step": 28572 }, { "epoch": 42.77395209580838, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1203, "step": 28573 }, { "epoch": 42.775449101796404, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1192, "step": 28574 }, { "epoch": 42.77694610778443, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1216, "step": 28575 }, { "epoch": 42.778443113772454, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1274, "step": 28576 }, { "epoch": 42.77994011976048, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1221, "step": 28577 }, { "epoch": 42.7814371257485, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1178, "step": 28578 }, { "epoch": 42.782934131736525, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1258, "step": 28579 }, { "epoch": 42.78443113772455, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1238, "step": 28580 }, { "epoch": 42.785928143712574, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1239, "step": 28581 }, { "epoch": 42.787425149700596, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1219, "step": 28582 }, { "epoch": 42.788922155688624, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1209, "step": 28583 }, { "epoch": 42.790419161676645, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1236, "step": 28584 }, { "epoch": 42.791916167664674, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1193, "step": 28585 }, { "epoch": 42.793413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1266, "step": 28586 }, { "epoch": 42.794910179640716, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1172, "step": 28587 }, { "epoch": 42.796407185628745, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1243, "step": 28588 }, { "epoch": 42.797904191616766, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.12, "step": 28589 }, { "epoch": 42.79940119760479, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1179, "step": 28590 }, { "epoch": 42.800898203592816, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1162, "step": 28591 }, { "epoch": 42.80239520958084, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1253, "step": 28592 }, { "epoch": 42.803892215568865, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1261, "step": 28593 }, { "epoch": 42.80538922155689, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.124, "step": 28594 }, { "epoch": 42.80688622754491, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1232, "step": 28595 }, { "epoch": 42.808383233532936, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1227, "step": 28596 }, { "epoch": 42.80988023952096, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1253, "step": 28597 }, { "epoch": 42.81137724550898, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.123, "step": 28598 }, { "epoch": 42.81287425149701, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1217, "step": 28599 }, { "epoch": 42.81437125748503, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1249, "step": 28600 }, { "epoch": 42.81586826347306, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1257, "step": 28601 }, { "epoch": 42.81736526946108, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1246, "step": 28602 }, { "epoch": 42.8188622754491, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1311, "step": 28603 }, { "epoch": 42.82035928143713, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1168, "step": 28604 }, { "epoch": 42.82185628742515, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 28605 }, { "epoch": 42.82335329341317, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1234, "step": 28606 }, { "epoch": 42.8248502994012, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1214, "step": 28607 }, { "epoch": 42.82634730538922, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1272, "step": 28608 }, { "epoch": 42.82784431137725, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1188, "step": 28609 }, { "epoch": 42.82934131736527, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1206, "step": 28610 }, { "epoch": 42.83083832335329, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1226, "step": 28611 }, { "epoch": 42.83233532934132, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1239, "step": 28612 }, { "epoch": 42.83383233532934, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1241, "step": 28613 }, { "epoch": 42.83532934131736, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1222, "step": 28614 }, { "epoch": 42.83682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1281, "step": 28615 }, { "epoch": 42.83832335329341, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1286, "step": 28616 }, { "epoch": 42.83982035928144, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1208, "step": 28617 }, { "epoch": 42.84131736526946, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1177, "step": 28618 }, { "epoch": 42.84281437125748, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1264, "step": 28619 }, { "epoch": 42.84431137724551, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1158, "step": 28620 }, { "epoch": 42.84580838323353, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1306, "step": 28621 }, { "epoch": 42.84730538922156, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1186, "step": 28622 }, { "epoch": 42.84880239520958, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1283, "step": 28623 }, { "epoch": 42.8502994011976, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.119, "step": 28624 }, { "epoch": 42.85179640718563, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1258, "step": 28625 }, { "epoch": 42.85329341317365, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1173, "step": 28626 }, { "epoch": 42.854790419161674, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1218, "step": 28627 }, { "epoch": 42.8562874251497, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1201, "step": 28628 }, { "epoch": 42.85778443113772, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1271, "step": 28629 }, { "epoch": 42.85928143712575, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1237, "step": 28630 }, { "epoch": 42.86077844311377, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1234, "step": 28631 }, { "epoch": 42.862275449101794, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1215, "step": 28632 }, { "epoch": 42.86377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1248, "step": 28633 }, { "epoch": 42.865269461077844, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1233, "step": 28634 }, { "epoch": 42.866766467065865, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1245, "step": 28635 }, { "epoch": 42.868263473053894, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1296, "step": 28636 }, { "epoch": 42.869760479041915, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1176, "step": 28637 }, { "epoch": 42.87125748502994, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1262, "step": 28638 }, { "epoch": 42.872754491017965, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1174, "step": 28639 }, { "epoch": 42.874251497005986, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1247, "step": 28640 }, { "epoch": 42.875748502994014, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1272, "step": 28641 }, { "epoch": 42.877245508982035, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1254, "step": 28642 }, { "epoch": 42.87874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1261, "step": 28643 }, { "epoch": 42.880239520958085, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1233, "step": 28644 }, { "epoch": 42.881736526946106, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1223, "step": 28645 }, { "epoch": 42.883233532934135, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1233, "step": 28646 }, { "epoch": 42.884730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 28647 }, { "epoch": 42.88622754491018, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1233, "step": 28648 }, { "epoch": 42.887724550898206, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1231, "step": 28649 }, { "epoch": 42.88922155688623, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1304, "step": 28650 }, { "epoch": 42.89071856287425, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1145, "step": 28651 }, { "epoch": 42.89221556886228, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1202, "step": 28652 }, { "epoch": 42.8937125748503, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1225, "step": 28653 }, { "epoch": 42.895209580838326, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1147, "step": 28654 }, { "epoch": 42.89670658682635, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1198, "step": 28655 }, { "epoch": 42.89820359281437, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1215, "step": 28656 }, { "epoch": 42.8997005988024, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.12, "step": 28657 }, { "epoch": 42.90119760479042, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1324, "step": 28658 }, { "epoch": 42.90269461077844, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.124, "step": 28659 }, { "epoch": 42.90419161676647, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1255, "step": 28660 }, { "epoch": 42.90568862275449, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1213, "step": 28661 }, { "epoch": 42.90718562874252, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1255, "step": 28662 }, { "epoch": 42.90868263473054, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1276, "step": 28663 }, { "epoch": 42.91017964071856, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1236, "step": 28664 }, { "epoch": 42.91167664670659, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1244, "step": 28665 }, { "epoch": 42.91317365269461, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1198, "step": 28666 }, { "epoch": 42.91467065868264, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1219, "step": 28667 }, { "epoch": 42.91616766467066, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1232, "step": 28668 }, { "epoch": 42.91766467065868, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1229, "step": 28669 }, { "epoch": 42.91916167664671, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1229, "step": 28670 }, { "epoch": 42.92065868263473, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1273, "step": 28671 }, { "epoch": 42.92215568862275, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1265, "step": 28672 }, { "epoch": 42.92365269461078, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1271, "step": 28673 }, { "epoch": 42.9251497005988, "grad_norm": 0.051025390625, "learning_rate": 0.0008, "loss": 1.1249, "step": 28674 }, { "epoch": 42.92664670658683, "grad_norm": 0.049072265625, "learning_rate": 0.0008, "loss": 1.1258, "step": 28675 }, { "epoch": 42.92814371257485, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1293, "step": 28676 }, { "epoch": 42.92964071856287, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1211, "step": 28677 }, { "epoch": 42.9311377245509, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.128, "step": 28678 }, { "epoch": 42.93263473053892, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1171, "step": 28679 }, { "epoch": 42.93413173652694, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1252, "step": 28680 }, { "epoch": 42.93562874251497, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1204, "step": 28681 }, { "epoch": 42.93712574850299, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.129, "step": 28682 }, { "epoch": 42.93862275449102, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.121, "step": 28683 }, { "epoch": 42.94011976047904, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1262, "step": 28684 }, { "epoch": 42.941616766467064, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1203, "step": 28685 }, { "epoch": 42.94311377245509, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.12, "step": 28686 }, { "epoch": 42.94461077844311, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1233, "step": 28687 }, { "epoch": 42.946107784431135, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1224, "step": 28688 }, { "epoch": 42.94760479041916, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1254, "step": 28689 }, { "epoch": 42.949101796407184, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1219, "step": 28690 }, { "epoch": 42.95059880239521, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1239, "step": 28691 }, { "epoch": 42.952095808383234, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1236, "step": 28692 }, { "epoch": 42.953592814371255, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1301, "step": 28693 }, { "epoch": 42.955089820359284, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1224, "step": 28694 }, { "epoch": 42.956586826347305, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1236, "step": 28695 }, { "epoch": 42.958083832335326, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1206, "step": 28696 }, { "epoch": 42.959580838323355, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1205, "step": 28697 }, { "epoch": 42.961077844311376, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1195, "step": 28698 }, { "epoch": 42.962574850299404, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1267, "step": 28699 }, { "epoch": 42.964071856287426, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 28700 }, { "epoch": 42.96556886227545, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.123, "step": 28701 }, { "epoch": 42.967065868263475, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1238, "step": 28702 }, { "epoch": 42.9685628742515, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1187, "step": 28703 }, { "epoch": 42.97005988023952, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1237, "step": 28704 }, { "epoch": 42.971556886227546, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1227, "step": 28705 }, { "epoch": 42.97305389221557, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1217, "step": 28706 }, { "epoch": 42.974550898203596, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1318, "step": 28707 }, { "epoch": 42.97604790419162, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1201, "step": 28708 }, { "epoch": 42.97754491017964, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1208, "step": 28709 }, { "epoch": 42.97904191616767, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1189, "step": 28710 }, { "epoch": 42.98053892215569, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1242, "step": 28711 }, { "epoch": 42.98203592814371, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1195, "step": 28712 }, { "epoch": 42.98353293413174, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1184, "step": 28713 }, { "epoch": 42.98502994011976, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1265, "step": 28714 }, { "epoch": 42.98652694610779, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1228, "step": 28715 }, { "epoch": 42.98802395209581, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1265, "step": 28716 }, { "epoch": 42.98952095808383, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1172, "step": 28717 }, { "epoch": 42.99101796407186, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1271, "step": 28718 }, { "epoch": 42.99251497005988, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.131, "step": 28719 }, { "epoch": 42.9940119760479, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1233, "step": 28720 }, { "epoch": 42.99550898203593, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1222, "step": 28721 }, { "epoch": 42.99700598802395, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1211, "step": 28722 }, { "epoch": 42.99850299401198, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1288, "step": 28723 }, { "epoch": 43.0, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1184, "step": 28724 }, { "epoch": 43.00149700598802, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1183, "step": 28725 }, { "epoch": 43.00299401197605, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1235, "step": 28726 }, { "epoch": 43.00449101796407, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1203, "step": 28727 }, { "epoch": 43.0059880239521, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1242, "step": 28728 }, { "epoch": 43.00748502994012, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1202, "step": 28729 }, { "epoch": 43.00898203592814, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1286, "step": 28730 }, { "epoch": 43.01047904191617, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 28731 }, { "epoch": 43.01197604790419, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1235, "step": 28732 }, { "epoch": 43.01347305389221, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1178, "step": 28733 }, { "epoch": 43.01497005988024, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1198, "step": 28734 }, { "epoch": 43.01646706586826, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1204, "step": 28735 }, { "epoch": 43.01796407185629, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1217, "step": 28736 }, { "epoch": 43.01946107784431, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1238, "step": 28737 }, { "epoch": 43.02095808383233, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1263, "step": 28738 }, { "epoch": 43.02245508982036, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1313, "step": 28739 }, { "epoch": 43.02395209580838, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1194, "step": 28740 }, { "epoch": 43.025449101796404, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1171, "step": 28741 }, { "epoch": 43.02694610778443, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1234, "step": 28742 }, { "epoch": 43.028443113772454, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1234, "step": 28743 }, { "epoch": 43.02994011976048, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1189, "step": 28744 }, { "epoch": 43.0314371257485, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1174, "step": 28745 }, { "epoch": 43.032934131736525, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.125, "step": 28746 }, { "epoch": 43.03443113772455, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1214, "step": 28747 }, { "epoch": 43.035928143712574, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1192, "step": 28748 }, { "epoch": 43.037425149700596, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1201, "step": 28749 }, { "epoch": 43.038922155688624, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1172, "step": 28750 }, { "epoch": 43.040419161676645, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1233, "step": 28751 }, { "epoch": 43.041916167664674, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1237, "step": 28752 }, { "epoch": 43.043413173652695, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1194, "step": 28753 }, { "epoch": 43.044910179640716, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.129, "step": 28754 }, { "epoch": 43.046407185628745, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1285, "step": 28755 }, { "epoch": 43.047904191616766, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.121, "step": 28756 }, { "epoch": 43.04940119760479, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1233, "step": 28757 }, { "epoch": 43.050898203592816, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1247, "step": 28758 }, { "epoch": 43.05239520958084, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1259, "step": 28759 }, { "epoch": 43.053892215568865, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1186, "step": 28760 }, { "epoch": 43.05538922155689, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1211, "step": 28761 }, { "epoch": 43.05688622754491, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1181, "step": 28762 }, { "epoch": 43.058383233532936, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1293, "step": 28763 }, { "epoch": 43.05988023952096, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1217, "step": 28764 }, { "epoch": 43.06137724550898, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1221, "step": 28765 }, { "epoch": 43.06287425149701, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.119, "step": 28766 }, { "epoch": 43.06437125748503, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1227, "step": 28767 }, { "epoch": 43.06586826347306, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1222, "step": 28768 }, { "epoch": 43.06736526946108, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1261, "step": 28769 }, { "epoch": 43.0688622754491, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1223, "step": 28770 }, { "epoch": 43.07035928143713, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1202, "step": 28771 }, { "epoch": 43.07185628742515, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1201, "step": 28772 }, { "epoch": 43.07335329341317, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1208, "step": 28773 }, { "epoch": 43.0748502994012, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1168, "step": 28774 }, { "epoch": 43.07634730538922, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1178, "step": 28775 }, { "epoch": 43.07784431137725, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1296, "step": 28776 }, { "epoch": 43.07934131736527, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1193, "step": 28777 }, { "epoch": 43.08083832335329, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1259, "step": 28778 }, { "epoch": 43.08233532934132, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1191, "step": 28779 }, { "epoch": 43.08383233532934, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1245, "step": 28780 }, { "epoch": 43.08532934131737, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1308, "step": 28781 }, { "epoch": 43.08682634730539, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.124, "step": 28782 }, { "epoch": 43.08832335329341, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.1247, "step": 28783 }, { "epoch": 43.08982035928144, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1197, "step": 28784 }, { "epoch": 43.09131736526946, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1254, "step": 28785 }, { "epoch": 43.09281437125748, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1293, "step": 28786 }, { "epoch": 43.09431137724551, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1261, "step": 28787 }, { "epoch": 43.09580838323353, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1207, "step": 28788 }, { "epoch": 43.09730538922156, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.118, "step": 28789 }, { "epoch": 43.09880239520958, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.125, "step": 28790 }, { "epoch": 43.1002994011976, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1177, "step": 28791 }, { "epoch": 43.10179640718563, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1218, "step": 28792 }, { "epoch": 43.10329341317365, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1176, "step": 28793 }, { "epoch": 43.104790419161674, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1316, "step": 28794 }, { "epoch": 43.1062874251497, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1272, "step": 28795 }, { "epoch": 43.10778443113772, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1231, "step": 28796 }, { "epoch": 43.10928143712575, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1227, "step": 28797 }, { "epoch": 43.11077844311377, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.12, "step": 28798 }, { "epoch": 43.112275449101794, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1168, "step": 28799 }, { "epoch": 43.11377245508982, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1264, "step": 28800 }, { "epoch": 43.115269461077844, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1173, "step": 28801 }, { "epoch": 43.116766467065865, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1267, "step": 28802 }, { "epoch": 43.118263473053894, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1258, "step": 28803 }, { "epoch": 43.119760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1256, "step": 28804 }, { "epoch": 43.12125748502994, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1259, "step": 28805 }, { "epoch": 43.122754491017965, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 28806 }, { "epoch": 43.124251497005986, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1259, "step": 28807 }, { "epoch": 43.125748502994014, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1174, "step": 28808 }, { "epoch": 43.127245508982035, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1222, "step": 28809 }, { "epoch": 43.12874251497006, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1274, "step": 28810 }, { "epoch": 43.130239520958085, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1211, "step": 28811 }, { "epoch": 43.131736526946106, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1208, "step": 28812 }, { "epoch": 43.133233532934135, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1214, "step": 28813 }, { "epoch": 43.134730538922156, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1167, "step": 28814 }, { "epoch": 43.13622754491018, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1274, "step": 28815 }, { "epoch": 43.137724550898206, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1197, "step": 28816 }, { "epoch": 43.13922155688623, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1218, "step": 28817 }, { "epoch": 43.14071856287425, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1227, "step": 28818 }, { "epoch": 43.14221556886228, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1231, "step": 28819 }, { "epoch": 43.1437125748503, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1182, "step": 28820 }, { "epoch": 43.145209580838326, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1159, "step": 28821 }, { "epoch": 43.14670658682635, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1204, "step": 28822 }, { "epoch": 43.14820359281437, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1271, "step": 28823 }, { "epoch": 43.1497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1239, "step": 28824 }, { "epoch": 43.15119760479042, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.119, "step": 28825 }, { "epoch": 43.15269461077844, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1228, "step": 28826 }, { "epoch": 43.15419161676647, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1256, "step": 28827 }, { "epoch": 43.15568862275449, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1216, "step": 28828 }, { "epoch": 43.15718562874252, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1279, "step": 28829 }, { "epoch": 43.15868263473054, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1202, "step": 28830 }, { "epoch": 43.16017964071856, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1206, "step": 28831 }, { "epoch": 43.16167664670659, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1177, "step": 28832 }, { "epoch": 43.16317365269461, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1236, "step": 28833 }, { "epoch": 43.16467065868263, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.126, "step": 28834 }, { "epoch": 43.16616766467066, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1113, "step": 28835 }, { "epoch": 43.16766467065868, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1269, "step": 28836 }, { "epoch": 43.16916167664671, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1226, "step": 28837 }, { "epoch": 43.17065868263473, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1259, "step": 28838 }, { "epoch": 43.17215568862275, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1181, "step": 28839 }, { "epoch": 43.17365269461078, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1192, "step": 28840 }, { "epoch": 43.1751497005988, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1157, "step": 28841 }, { "epoch": 43.17664670658683, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1228, "step": 28842 }, { "epoch": 43.17814371257485, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1202, "step": 28843 }, { "epoch": 43.17964071856287, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.122, "step": 28844 }, { "epoch": 43.1811377245509, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1252, "step": 28845 }, { "epoch": 43.18263473053892, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1214, "step": 28846 }, { "epoch": 43.18413173652694, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.121, "step": 28847 }, { "epoch": 43.18562874251497, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1229, "step": 28848 }, { "epoch": 43.18712574850299, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1297, "step": 28849 }, { "epoch": 43.18862275449102, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.125, "step": 28850 }, { "epoch": 43.19011976047904, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1263, "step": 28851 }, { "epoch": 43.191616766467064, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.126, "step": 28852 }, { "epoch": 43.19311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1237, "step": 28853 }, { "epoch": 43.19461077844311, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1206, "step": 28854 }, { "epoch": 43.196107784431135, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1255, "step": 28855 }, { "epoch": 43.19760479041916, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.125, "step": 28856 }, { "epoch": 43.199101796407184, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1218, "step": 28857 }, { "epoch": 43.20059880239521, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1298, "step": 28858 }, { "epoch": 43.202095808383234, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1208, "step": 28859 }, { "epoch": 43.203592814371255, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1255, "step": 28860 }, { "epoch": 43.205089820359284, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1241, "step": 28861 }, { "epoch": 43.206586826347305, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1269, "step": 28862 }, { "epoch": 43.208083832335326, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1234, "step": 28863 }, { "epoch": 43.209580838323355, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1206, "step": 28864 }, { "epoch": 43.211077844311376, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1219, "step": 28865 }, { "epoch": 43.212574850299404, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1266, "step": 28866 }, { "epoch": 43.214071856287426, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1266, "step": 28867 }, { "epoch": 43.21556886227545, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1219, "step": 28868 }, { "epoch": 43.217065868263475, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.12, "step": 28869 }, { "epoch": 43.2185628742515, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1262, "step": 28870 }, { "epoch": 43.22005988023952, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1206, "step": 28871 }, { "epoch": 43.221556886227546, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1245, "step": 28872 }, { "epoch": 43.22305389221557, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.124, "step": 28873 }, { "epoch": 43.224550898203596, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1189, "step": 28874 }, { "epoch": 43.22604790419162, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1247, "step": 28875 }, { "epoch": 43.22754491017964, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.126, "step": 28876 }, { "epoch": 43.22904191616767, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1247, "step": 28877 }, { "epoch": 43.23053892215569, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1257, "step": 28878 }, { "epoch": 43.23203592814371, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1211, "step": 28879 }, { "epoch": 43.23353293413174, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1259, "step": 28880 }, { "epoch": 43.23502994011976, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1232, "step": 28881 }, { "epoch": 43.23652694610779, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1247, "step": 28882 }, { "epoch": 43.23802395209581, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.124, "step": 28883 }, { "epoch": 43.23952095808383, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1252, "step": 28884 }, { "epoch": 43.24101796407186, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1278, "step": 28885 }, { "epoch": 43.24251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1184, "step": 28886 }, { "epoch": 43.2440119760479, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1205, "step": 28887 }, { "epoch": 43.24550898203593, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1231, "step": 28888 }, { "epoch": 43.24700598802395, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1215, "step": 28889 }, { "epoch": 43.24850299401198, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1249, "step": 28890 }, { "epoch": 43.25, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1273, "step": 28891 }, { "epoch": 43.25149700598802, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1268, "step": 28892 }, { "epoch": 43.25299401197605, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.125, "step": 28893 }, { "epoch": 43.25449101796407, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1266, "step": 28894 }, { "epoch": 43.2559880239521, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1195, "step": 28895 }, { "epoch": 43.25748502994012, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1226, "step": 28896 }, { "epoch": 43.25898203592814, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1196, "step": 28897 }, { "epoch": 43.26047904191617, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1242, "step": 28898 }, { "epoch": 43.26197604790419, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1226, "step": 28899 }, { "epoch": 43.26347305389221, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1166, "step": 28900 }, { "epoch": 43.26497005988024, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1235, "step": 28901 }, { "epoch": 43.26646706586826, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1227, "step": 28902 }, { "epoch": 43.26796407185629, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1247, "step": 28903 }, { "epoch": 43.26946107784431, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1243, "step": 28904 }, { "epoch": 43.27095808383233, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.121, "step": 28905 }, { "epoch": 43.27245508982036, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1225, "step": 28906 }, { "epoch": 43.27395209580838, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.129, "step": 28907 }, { "epoch": 43.275449101796404, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1282, "step": 28908 }, { "epoch": 43.27694610778443, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1181, "step": 28909 }, { "epoch": 43.278443113772454, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1206, "step": 28910 }, { "epoch": 43.27994011976048, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1207, "step": 28911 }, { "epoch": 43.2814371257485, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1198, "step": 28912 }, { "epoch": 43.282934131736525, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1218, "step": 28913 }, { "epoch": 43.28443113772455, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1243, "step": 28914 }, { "epoch": 43.285928143712574, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1193, "step": 28915 }, { "epoch": 43.287425149700596, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1284, "step": 28916 }, { "epoch": 43.288922155688624, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1227, "step": 28917 }, { "epoch": 43.290419161676645, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1261, "step": 28918 }, { "epoch": 43.291916167664674, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1226, "step": 28919 }, { "epoch": 43.293413173652695, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.1266, "step": 28920 }, { "epoch": 43.294910179640716, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1249, "step": 28921 }, { "epoch": 43.296407185628745, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1201, "step": 28922 }, { "epoch": 43.297904191616766, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.1291, "step": 28923 }, { "epoch": 43.29940119760479, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1212, "step": 28924 }, { "epoch": 43.300898203592816, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1255, "step": 28925 }, { "epoch": 43.30239520958084, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1219, "step": 28926 }, { "epoch": 43.303892215568865, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1221, "step": 28927 }, { "epoch": 43.30538922155689, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1215, "step": 28928 }, { "epoch": 43.30688622754491, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1179, "step": 28929 }, { "epoch": 43.308383233532936, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1226, "step": 28930 }, { "epoch": 43.30988023952096, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1229, "step": 28931 }, { "epoch": 43.31137724550898, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1227, "step": 28932 }, { "epoch": 43.31287425149701, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 28933 }, { "epoch": 43.31437125748503, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1213, "step": 28934 }, { "epoch": 43.31586826347306, "grad_norm": 0.049560546875, "learning_rate": 0.0008, "loss": 1.1224, "step": 28935 }, { "epoch": 43.31736526946108, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1147, "step": 28936 }, { "epoch": 43.3188622754491, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1165, "step": 28937 }, { "epoch": 43.32035928143713, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1245, "step": 28938 }, { "epoch": 43.32185628742515, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1178, "step": 28939 }, { "epoch": 43.32335329341317, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1258, "step": 28940 }, { "epoch": 43.3248502994012, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1263, "step": 28941 }, { "epoch": 43.32634730538922, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.128, "step": 28942 }, { "epoch": 43.32784431137725, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1225, "step": 28943 }, { "epoch": 43.32934131736527, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1231, "step": 28944 }, { "epoch": 43.33083832335329, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 28945 }, { "epoch": 43.33233532934132, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1314, "step": 28946 }, { "epoch": 43.33383233532934, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1216, "step": 28947 }, { "epoch": 43.33532934131736, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1244, "step": 28948 }, { "epoch": 43.33682634730539, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1294, "step": 28949 }, { "epoch": 43.33832335329341, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1295, "step": 28950 }, { "epoch": 43.33982035928144, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1248, "step": 28951 }, { "epoch": 43.34131736526946, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.124, "step": 28952 }, { "epoch": 43.34281437125748, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1235, "step": 28953 }, { "epoch": 43.34431137724551, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1219, "step": 28954 }, { "epoch": 43.34580838323353, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1211, "step": 28955 }, { "epoch": 43.34730538922156, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1245, "step": 28956 }, { "epoch": 43.34880239520958, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1157, "step": 28957 }, { "epoch": 43.3502994011976, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1259, "step": 28958 }, { "epoch": 43.35179640718563, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.12, "step": 28959 }, { "epoch": 43.35329341317365, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1229, "step": 28960 }, { "epoch": 43.354790419161674, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1183, "step": 28961 }, { "epoch": 43.3562874251497, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1225, "step": 28962 }, { "epoch": 43.35778443113772, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.122, "step": 28963 }, { "epoch": 43.35928143712575, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1131, "step": 28964 }, { "epoch": 43.36077844311377, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.121, "step": 28965 }, { "epoch": 43.362275449101794, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.126, "step": 28966 }, { "epoch": 43.36377245508982, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1228, "step": 28967 }, { "epoch": 43.365269461077844, "grad_norm": 0.051025390625, "learning_rate": 0.0008, "loss": 1.1199, "step": 28968 }, { "epoch": 43.366766467065865, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1258, "step": 28969 }, { "epoch": 43.368263473053894, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1197, "step": 28970 }, { "epoch": 43.369760479041915, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.128, "step": 28971 }, { "epoch": 43.37125748502994, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1243, "step": 28972 }, { "epoch": 43.372754491017965, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1211, "step": 28973 }, { "epoch": 43.374251497005986, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1226, "step": 28974 }, { "epoch": 43.375748502994014, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.119, "step": 28975 }, { "epoch": 43.377245508982035, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1262, "step": 28976 }, { "epoch": 43.37874251497006, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1272, "step": 28977 }, { "epoch": 43.380239520958085, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.123, "step": 28978 }, { "epoch": 43.381736526946106, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.117, "step": 28979 }, { "epoch": 43.383233532934135, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1213, "step": 28980 }, { "epoch": 43.384730538922156, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.127, "step": 28981 }, { "epoch": 43.38622754491018, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1285, "step": 28982 }, { "epoch": 43.387724550898206, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1234, "step": 28983 }, { "epoch": 43.38922155688623, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1276, "step": 28984 }, { "epoch": 43.39071856287425, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1215, "step": 28985 }, { "epoch": 43.39221556886228, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1192, "step": 28986 }, { "epoch": 43.3937125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1244, "step": 28987 }, { "epoch": 43.395209580838326, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1264, "step": 28988 }, { "epoch": 43.39670658682635, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1229, "step": 28989 }, { "epoch": 43.39820359281437, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1239, "step": 28990 }, { "epoch": 43.3997005988024, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1263, "step": 28991 }, { "epoch": 43.40119760479042, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1249, "step": 28992 }, { "epoch": 43.40269461077844, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1261, "step": 28993 }, { "epoch": 43.40419161676647, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.124, "step": 28994 }, { "epoch": 43.40568862275449, "grad_norm": 0.050537109375, "learning_rate": 0.0008, "loss": 1.122, "step": 28995 }, { "epoch": 43.40718562874252, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1224, "step": 28996 }, { "epoch": 43.40868263473054, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1235, "step": 28997 }, { "epoch": 43.41017964071856, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1286, "step": 28998 }, { "epoch": 43.41167664670659, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1257, "step": 28999 }, { "epoch": 43.41317365269461, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1223, "step": 29000 }, { "epoch": 43.41467065868264, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1259, "step": 29001 }, { "epoch": 43.41616766467066, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1234, "step": 29002 }, { "epoch": 43.41766467065868, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1137, "step": 29003 }, { "epoch": 43.41916167664671, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 29004 }, { "epoch": 43.42065868263473, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1224, "step": 29005 }, { "epoch": 43.42215568862275, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1195, "step": 29006 }, { "epoch": 43.42365269461078, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1218, "step": 29007 }, { "epoch": 43.4251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1276, "step": 29008 }, { "epoch": 43.42664670658683, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.124, "step": 29009 }, { "epoch": 43.42814371257485, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1216, "step": 29010 }, { "epoch": 43.42964071856287, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.119, "step": 29011 }, { "epoch": 43.4311377245509, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1234, "step": 29012 }, { "epoch": 43.43263473053892, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1268, "step": 29013 }, { "epoch": 43.43413173652694, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.124, "step": 29014 }, { "epoch": 43.43562874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1203, "step": 29015 }, { "epoch": 43.43712574850299, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.123, "step": 29016 }, { "epoch": 43.43862275449102, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.121, "step": 29017 }, { "epoch": 43.44011976047904, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1234, "step": 29018 }, { "epoch": 43.441616766467064, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1214, "step": 29019 }, { "epoch": 43.44311377245509, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1295, "step": 29020 }, { "epoch": 43.44461077844311, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1248, "step": 29021 }, { "epoch": 43.446107784431135, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1233, "step": 29022 }, { "epoch": 43.44760479041916, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1195, "step": 29023 }, { "epoch": 43.449101796407184, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1193, "step": 29024 }, { "epoch": 43.45059880239521, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1277, "step": 29025 }, { "epoch": 43.452095808383234, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1164, "step": 29026 }, { "epoch": 43.453592814371255, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.128, "step": 29027 }, { "epoch": 43.455089820359284, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1191, "step": 29028 }, { "epoch": 43.456586826347305, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1203, "step": 29029 }, { "epoch": 43.458083832335326, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1234, "step": 29030 }, { "epoch": 43.459580838323355, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1189, "step": 29031 }, { "epoch": 43.461077844311376, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1256, "step": 29032 }, { "epoch": 43.462574850299404, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1227, "step": 29033 }, { "epoch": 43.464071856287426, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1253, "step": 29034 }, { "epoch": 43.46556886227545, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1207, "step": 29035 }, { "epoch": 43.467065868263475, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1201, "step": 29036 }, { "epoch": 43.4685628742515, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1231, "step": 29037 }, { "epoch": 43.47005988023952, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1209, "step": 29038 }, { "epoch": 43.471556886227546, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1205, "step": 29039 }, { "epoch": 43.47305389221557, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1243, "step": 29040 }, { "epoch": 43.474550898203596, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1199, "step": 29041 }, { "epoch": 43.47604790419162, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1248, "step": 29042 }, { "epoch": 43.47754491017964, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1154, "step": 29043 }, { "epoch": 43.47904191616767, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1231, "step": 29044 }, { "epoch": 43.48053892215569, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1193, "step": 29045 }, { "epoch": 43.48203592814371, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1279, "step": 29046 }, { "epoch": 43.48353293413174, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1225, "step": 29047 }, { "epoch": 43.48502994011976, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1263, "step": 29048 }, { "epoch": 43.48652694610779, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1325, "step": 29049 }, { "epoch": 43.48802395209581, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1264, "step": 29050 }, { "epoch": 43.48952095808383, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1276, "step": 29051 }, { "epoch": 43.49101796407186, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1242, "step": 29052 }, { "epoch": 43.49251497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1254, "step": 29053 }, { "epoch": 43.4940119760479, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1166, "step": 29054 }, { "epoch": 43.49550898203593, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1303, "step": 29055 }, { "epoch": 43.49700598802395, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1221, "step": 29056 }, { "epoch": 43.49850299401198, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1195, "step": 29057 }, { "epoch": 43.5, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1254, "step": 29058 }, { "epoch": 43.50149700598802, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1232, "step": 29059 }, { "epoch": 43.50299401197605, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1221, "step": 29060 }, { "epoch": 43.50449101796407, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1218, "step": 29061 }, { "epoch": 43.5059880239521, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1149, "step": 29062 }, { "epoch": 43.50748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1214, "step": 29063 }, { "epoch": 43.50898203592814, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1253, "step": 29064 }, { "epoch": 43.51047904191617, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1209, "step": 29065 }, { "epoch": 43.51197604790419, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1194, "step": 29066 }, { "epoch": 43.51347305389221, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 29067 }, { "epoch": 43.51497005988024, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1216, "step": 29068 }, { "epoch": 43.51646706586826, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.125, "step": 29069 }, { "epoch": 43.51796407185629, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.124, "step": 29070 }, { "epoch": 43.51946107784431, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1213, "step": 29071 }, { "epoch": 43.52095808383233, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1227, "step": 29072 }, { "epoch": 43.52245508982036, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.118, "step": 29073 }, { "epoch": 43.52395209580838, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.118, "step": 29074 }, { "epoch": 43.525449101796404, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1224, "step": 29075 }, { "epoch": 43.52694610778443, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1227, "step": 29076 }, { "epoch": 43.528443113772454, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1219, "step": 29077 }, { "epoch": 43.52994011976048, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.122, "step": 29078 }, { "epoch": 43.5314371257485, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1256, "step": 29079 }, { "epoch": 43.532934131736525, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1224, "step": 29080 }, { "epoch": 43.53443113772455, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1235, "step": 29081 }, { "epoch": 43.535928143712574, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1266, "step": 29082 }, { "epoch": 43.537425149700596, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 29083 }, { "epoch": 43.538922155688624, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1303, "step": 29084 }, { "epoch": 43.540419161676645, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1179, "step": 29085 }, { "epoch": 43.541916167664674, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1261, "step": 29086 }, { "epoch": 43.543413173652695, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1261, "step": 29087 }, { "epoch": 43.544910179640716, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1235, "step": 29088 }, { "epoch": 43.546407185628745, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.121, "step": 29089 }, { "epoch": 43.547904191616766, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1214, "step": 29090 }, { "epoch": 43.54940119760479, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1242, "step": 29091 }, { "epoch": 43.550898203592816, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1236, "step": 29092 }, { "epoch": 43.55239520958084, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1258, "step": 29093 }, { "epoch": 43.553892215568865, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1261, "step": 29094 }, { "epoch": 43.55538922155689, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.124, "step": 29095 }, { "epoch": 43.55688622754491, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1169, "step": 29096 }, { "epoch": 43.558383233532936, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1234, "step": 29097 }, { "epoch": 43.55988023952096, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1244, "step": 29098 }, { "epoch": 43.56137724550898, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1234, "step": 29099 }, { "epoch": 43.56287425149701, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1202, "step": 29100 }, { "epoch": 43.56437125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1309, "step": 29101 }, { "epoch": 43.56586826347306, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1198, "step": 29102 }, { "epoch": 43.56736526946108, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1215, "step": 29103 }, { "epoch": 43.5688622754491, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1392, "step": 29104 }, { "epoch": 43.57035928143713, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1169, "step": 29105 }, { "epoch": 43.57185628742515, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.12, "step": 29106 }, { "epoch": 43.57335329341317, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1264, "step": 29107 }, { "epoch": 43.5748502994012, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1261, "step": 29108 }, { "epoch": 43.57634730538922, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1195, "step": 29109 }, { "epoch": 43.57784431137725, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1264, "step": 29110 }, { "epoch": 43.57934131736527, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1239, "step": 29111 }, { "epoch": 43.58083832335329, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1269, "step": 29112 }, { "epoch": 43.58233532934132, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1265, "step": 29113 }, { "epoch": 43.58383233532934, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1205, "step": 29114 }, { "epoch": 43.58532934131736, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1222, "step": 29115 }, { "epoch": 43.58682634730539, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1217, "step": 29116 }, { "epoch": 43.58832335329341, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1161, "step": 29117 }, { "epoch": 43.58982035928144, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1183, "step": 29118 }, { "epoch": 43.59131736526946, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.12, "step": 29119 }, { "epoch": 43.59281437125748, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1259, "step": 29120 }, { "epoch": 43.59431137724551, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1206, "step": 29121 }, { "epoch": 43.59580838323353, "grad_norm": 0.050537109375, "learning_rate": 0.0008, "loss": 1.1214, "step": 29122 }, { "epoch": 43.59730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1215, "step": 29123 }, { "epoch": 43.59880239520958, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1287, "step": 29124 }, { "epoch": 43.6002994011976, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1153, "step": 29125 }, { "epoch": 43.60179640718563, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1299, "step": 29126 }, { "epoch": 43.60329341317365, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1272, "step": 29127 }, { "epoch": 43.604790419161674, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1145, "step": 29128 }, { "epoch": 43.6062874251497, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1216, "step": 29129 }, { "epoch": 43.60778443113772, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1226, "step": 29130 }, { "epoch": 43.60928143712575, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 29131 }, { "epoch": 43.61077844311377, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1196, "step": 29132 }, { "epoch": 43.612275449101794, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1262, "step": 29133 }, { "epoch": 43.61377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1253, "step": 29134 }, { "epoch": 43.615269461077844, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1248, "step": 29135 }, { "epoch": 43.616766467065865, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.122, "step": 29136 }, { "epoch": 43.618263473053894, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 29137 }, { "epoch": 43.619760479041915, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1303, "step": 29138 }, { "epoch": 43.62125748502994, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1289, "step": 29139 }, { "epoch": 43.622754491017965, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1253, "step": 29140 }, { "epoch": 43.624251497005986, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1314, "step": 29141 }, { "epoch": 43.625748502994014, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1238, "step": 29142 }, { "epoch": 43.627245508982035, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1203, "step": 29143 }, { "epoch": 43.62874251497006, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1229, "step": 29144 }, { "epoch": 43.630239520958085, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1219, "step": 29145 }, { "epoch": 43.631736526946106, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1269, "step": 29146 }, { "epoch": 43.633233532934135, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 29147 }, { "epoch": 43.634730538922156, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1273, "step": 29148 }, { "epoch": 43.63622754491018, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.123, "step": 29149 }, { "epoch": 43.637724550898206, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1245, "step": 29150 }, { "epoch": 43.63922155688623, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1259, "step": 29151 }, { "epoch": 43.64071856287425, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1211, "step": 29152 }, { "epoch": 43.64221556886228, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1228, "step": 29153 }, { "epoch": 43.6437125748503, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1209, "step": 29154 }, { "epoch": 43.645209580838326, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1249, "step": 29155 }, { "epoch": 43.64670658682635, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1188, "step": 29156 }, { "epoch": 43.64820359281437, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1241, "step": 29157 }, { "epoch": 43.6497005988024, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1251, "step": 29158 }, { "epoch": 43.65119760479042, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1224, "step": 29159 }, { "epoch": 43.65269461077844, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1188, "step": 29160 }, { "epoch": 43.65419161676647, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.126, "step": 29161 }, { "epoch": 43.65568862275449, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1298, "step": 29162 }, { "epoch": 43.65718562874252, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1261, "step": 29163 }, { "epoch": 43.65868263473054, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1269, "step": 29164 }, { "epoch": 43.66017964071856, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1211, "step": 29165 }, { "epoch": 43.66167664670659, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1286, "step": 29166 }, { "epoch": 43.66317365269461, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1242, "step": 29167 }, { "epoch": 43.66467065868264, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1228, "step": 29168 }, { "epoch": 43.66616766467066, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 29169 }, { "epoch": 43.66766467065868, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1265, "step": 29170 }, { "epoch": 43.66916167664671, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1226, "step": 29171 }, { "epoch": 43.67065868263473, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1253, "step": 29172 }, { "epoch": 43.67215568862275, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1175, "step": 29173 }, { "epoch": 43.67365269461078, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1238, "step": 29174 }, { "epoch": 43.6751497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1259, "step": 29175 }, { "epoch": 43.67664670658683, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1262, "step": 29176 }, { "epoch": 43.67814371257485, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.116, "step": 29177 }, { "epoch": 43.67964071856287, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1239, "step": 29178 }, { "epoch": 43.6811377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.122, "step": 29179 }, { "epoch": 43.68263473053892, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1247, "step": 29180 }, { "epoch": 43.68413173652694, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1248, "step": 29181 }, { "epoch": 43.68562874251497, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 29182 }, { "epoch": 43.68712574850299, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1251, "step": 29183 }, { "epoch": 43.68862275449102, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1146, "step": 29184 }, { "epoch": 43.69011976047904, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1185, "step": 29185 }, { "epoch": 43.691616766467064, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1299, "step": 29186 }, { "epoch": 43.69311377245509, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 29187 }, { "epoch": 43.69461077844311, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1225, "step": 29188 }, { "epoch": 43.696107784431135, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1287, "step": 29189 }, { "epoch": 43.69760479041916, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1214, "step": 29190 }, { "epoch": 43.699101796407184, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1144, "step": 29191 }, { "epoch": 43.70059880239521, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1243, "step": 29192 }, { "epoch": 43.702095808383234, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1211, "step": 29193 }, { "epoch": 43.703592814371255, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1163, "step": 29194 }, { "epoch": 43.705089820359284, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1305, "step": 29195 }, { "epoch": 43.706586826347305, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1269, "step": 29196 }, { "epoch": 43.708083832335326, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1288, "step": 29197 }, { "epoch": 43.709580838323355, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1237, "step": 29198 }, { "epoch": 43.711077844311376, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1179, "step": 29199 }, { "epoch": 43.712574850299404, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1215, "step": 29200 }, { "epoch": 43.714071856287426, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.125, "step": 29201 }, { "epoch": 43.71556886227545, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1234, "step": 29202 }, { "epoch": 43.717065868263475, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1309, "step": 29203 }, { "epoch": 43.7185628742515, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1218, "step": 29204 }, { "epoch": 43.72005988023952, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.117, "step": 29205 }, { "epoch": 43.721556886227546, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1254, "step": 29206 }, { "epoch": 43.72305389221557, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.117, "step": 29207 }, { "epoch": 43.724550898203596, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 29208 }, { "epoch": 43.72604790419162, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1217, "step": 29209 }, { "epoch": 43.72754491017964, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1323, "step": 29210 }, { "epoch": 43.72904191616767, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1209, "step": 29211 }, { "epoch": 43.73053892215569, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1275, "step": 29212 }, { "epoch": 43.73203592814371, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1259, "step": 29213 }, { "epoch": 43.73353293413174, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1228, "step": 29214 }, { "epoch": 43.73502994011976, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1204, "step": 29215 }, { "epoch": 43.73652694610779, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1231, "step": 29216 }, { "epoch": 43.73802395209581, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1216, "step": 29217 }, { "epoch": 43.73952095808383, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1247, "step": 29218 }, { "epoch": 43.74101796407186, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1224, "step": 29219 }, { "epoch": 43.74251497005988, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1203, "step": 29220 }, { "epoch": 43.7440119760479, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1292, "step": 29221 }, { "epoch": 43.74550898203593, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1173, "step": 29222 }, { "epoch": 43.74700598802395, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1201, "step": 29223 }, { "epoch": 43.74850299401198, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1219, "step": 29224 }, { "epoch": 43.75, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1248, "step": 29225 }, { "epoch": 43.75149700598802, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1332, "step": 29226 }, { "epoch": 43.75299401197605, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1261, "step": 29227 }, { "epoch": 43.75449101796407, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1226, "step": 29228 }, { "epoch": 43.7559880239521, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1258, "step": 29229 }, { "epoch": 43.75748502994012, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1207, "step": 29230 }, { "epoch": 43.75898203592814, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1264, "step": 29231 }, { "epoch": 43.76047904191617, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1287, "step": 29232 }, { "epoch": 43.76197604790419, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1212, "step": 29233 }, { "epoch": 43.76347305389221, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1278, "step": 29234 }, { "epoch": 43.76497005988024, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.123, "step": 29235 }, { "epoch": 43.76646706586826, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1146, "step": 29236 }, { "epoch": 43.76796407185629, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1191, "step": 29237 }, { "epoch": 43.76946107784431, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1213, "step": 29238 }, { "epoch": 43.77095808383233, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1334, "step": 29239 }, { "epoch": 43.77245508982036, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1273, "step": 29240 }, { "epoch": 43.77395209580838, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1229, "step": 29241 }, { "epoch": 43.775449101796404, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1186, "step": 29242 }, { "epoch": 43.77694610778443, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1249, "step": 29243 }, { "epoch": 43.778443113772454, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1218, "step": 29244 }, { "epoch": 43.77994011976048, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1198, "step": 29245 }, { "epoch": 43.7814371257485, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1208, "step": 29246 }, { "epoch": 43.782934131736525, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1243, "step": 29247 }, { "epoch": 43.78443113772455, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1133, "step": 29248 }, { "epoch": 43.785928143712574, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1189, "step": 29249 }, { "epoch": 43.787425149700596, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1316, "step": 29250 }, { "epoch": 43.788922155688624, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1224, "step": 29251 }, { "epoch": 43.790419161676645, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1165, "step": 29252 }, { "epoch": 43.791916167664674, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1244, "step": 29253 }, { "epoch": 43.793413173652695, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1242, "step": 29254 }, { "epoch": 43.794910179640716, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1207, "step": 29255 }, { "epoch": 43.796407185628745, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1119, "step": 29256 }, { "epoch": 43.797904191616766, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1125, "step": 29257 }, { "epoch": 43.79940119760479, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1196, "step": 29258 }, { "epoch": 43.800898203592816, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1226, "step": 29259 }, { "epoch": 43.80239520958084, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1241, "step": 29260 }, { "epoch": 43.803892215568865, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1217, "step": 29261 }, { "epoch": 43.80538922155689, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1229, "step": 29262 }, { "epoch": 43.80688622754491, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1183, "step": 29263 }, { "epoch": 43.808383233532936, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1212, "step": 29264 }, { "epoch": 43.80988023952096, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1295, "step": 29265 }, { "epoch": 43.81137724550898, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1239, "step": 29266 }, { "epoch": 43.81287425149701, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1227, "step": 29267 }, { "epoch": 43.81437125748503, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1242, "step": 29268 }, { "epoch": 43.81586826347306, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1205, "step": 29269 }, { "epoch": 43.81736526946108, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1247, "step": 29270 }, { "epoch": 43.8188622754491, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1214, "step": 29271 }, { "epoch": 43.82035928143713, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1142, "step": 29272 }, { "epoch": 43.82185628742515, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1209, "step": 29273 }, { "epoch": 43.82335329341317, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1238, "step": 29274 }, { "epoch": 43.8248502994012, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1274, "step": 29275 }, { "epoch": 43.82634730538922, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.128, "step": 29276 }, { "epoch": 43.82784431137725, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1211, "step": 29277 }, { "epoch": 43.82934131736527, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1307, "step": 29278 }, { "epoch": 43.83083832335329, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1205, "step": 29279 }, { "epoch": 43.83233532934132, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1182, "step": 29280 }, { "epoch": 43.83383233532934, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1291, "step": 29281 }, { "epoch": 43.83532934131736, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1195, "step": 29282 }, { "epoch": 43.83682634730539, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.124, "step": 29283 }, { "epoch": 43.83832335329341, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1195, "step": 29284 }, { "epoch": 43.83982035928144, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1176, "step": 29285 }, { "epoch": 43.84131736526946, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1208, "step": 29286 }, { "epoch": 43.84281437125748, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1294, "step": 29287 }, { "epoch": 43.84431137724551, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1271, "step": 29288 }, { "epoch": 43.84580838323353, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.122, "step": 29289 }, { "epoch": 43.84730538922156, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.12, "step": 29290 }, { "epoch": 43.84880239520958, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1261, "step": 29291 }, { "epoch": 43.8502994011976, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1237, "step": 29292 }, { "epoch": 43.85179640718563, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 29293 }, { "epoch": 43.85329341317365, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1236, "step": 29294 }, { "epoch": 43.854790419161674, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1135, "step": 29295 }, { "epoch": 43.8562874251497, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1317, "step": 29296 }, { "epoch": 43.85778443113772, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1239, "step": 29297 }, { "epoch": 43.85928143712575, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1181, "step": 29298 }, { "epoch": 43.86077844311377, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1303, "step": 29299 }, { "epoch": 43.862275449101794, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1188, "step": 29300 }, { "epoch": 43.86377245508982, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1185, "step": 29301 }, { "epoch": 43.865269461077844, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1202, "step": 29302 }, { "epoch": 43.866766467065865, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1205, "step": 29303 }, { "epoch": 43.868263473053894, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1191, "step": 29304 }, { "epoch": 43.869760479041915, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1233, "step": 29305 }, { "epoch": 43.87125748502994, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1211, "step": 29306 }, { "epoch": 43.872754491017965, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1233, "step": 29307 }, { "epoch": 43.874251497005986, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1194, "step": 29308 }, { "epoch": 43.875748502994014, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1262, "step": 29309 }, { "epoch": 43.877245508982035, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1198, "step": 29310 }, { "epoch": 43.87874251497006, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1239, "step": 29311 }, { "epoch": 43.880239520958085, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1223, "step": 29312 }, { "epoch": 43.881736526946106, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1236, "step": 29313 }, { "epoch": 43.883233532934135, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1318, "step": 29314 }, { "epoch": 43.884730538922156, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1268, "step": 29315 }, { "epoch": 43.88622754491018, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1186, "step": 29316 }, { "epoch": 43.887724550898206, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 29317 }, { "epoch": 43.88922155688623, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1196, "step": 29318 }, { "epoch": 43.89071856287425, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1218, "step": 29319 }, { "epoch": 43.89221556886228, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1242, "step": 29320 }, { "epoch": 43.8937125748503, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1155, "step": 29321 }, { "epoch": 43.895209580838326, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1184, "step": 29322 }, { "epoch": 43.89670658682635, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1227, "step": 29323 }, { "epoch": 43.89820359281437, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1174, "step": 29324 }, { "epoch": 43.8997005988024, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1181, "step": 29325 }, { "epoch": 43.90119760479042, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1253, "step": 29326 }, { "epoch": 43.90269461077844, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1215, "step": 29327 }, { "epoch": 43.90419161676647, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1229, "step": 29328 }, { "epoch": 43.90568862275449, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1204, "step": 29329 }, { "epoch": 43.90718562874252, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1193, "step": 29330 }, { "epoch": 43.90868263473054, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1237, "step": 29331 }, { "epoch": 43.91017964071856, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1257, "step": 29332 }, { "epoch": 43.91167664670659, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1199, "step": 29333 }, { "epoch": 43.91317365269461, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1158, "step": 29334 }, { "epoch": 43.91467065868264, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1266, "step": 29335 }, { "epoch": 43.91616766467066, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1285, "step": 29336 }, { "epoch": 43.91766467065868, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.125, "step": 29337 }, { "epoch": 43.91916167664671, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 29338 }, { "epoch": 43.92065868263473, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1219, "step": 29339 }, { "epoch": 43.92215568862275, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1229, "step": 29340 }, { "epoch": 43.92365269461078, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1181, "step": 29341 }, { "epoch": 43.9251497005988, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1217, "step": 29342 }, { "epoch": 43.92664670658683, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1187, "step": 29343 }, { "epoch": 43.92814371257485, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.125, "step": 29344 }, { "epoch": 43.92964071856287, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1266, "step": 29345 }, { "epoch": 43.9311377245509, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1137, "step": 29346 }, { "epoch": 43.93263473053892, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1218, "step": 29347 }, { "epoch": 43.93413173652694, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1202, "step": 29348 }, { "epoch": 43.93562874251497, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1306, "step": 29349 }, { "epoch": 43.93712574850299, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1202, "step": 29350 }, { "epoch": 43.93862275449102, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1253, "step": 29351 }, { "epoch": 43.94011976047904, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1141, "step": 29352 }, { "epoch": 43.941616766467064, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1281, "step": 29353 }, { "epoch": 43.94311377245509, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1237, "step": 29354 }, { "epoch": 43.94461077844311, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1164, "step": 29355 }, { "epoch": 43.946107784431135, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1216, "step": 29356 }, { "epoch": 43.94760479041916, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1212, "step": 29357 }, { "epoch": 43.949101796407184, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.122, "step": 29358 }, { "epoch": 43.95059880239521, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1255, "step": 29359 }, { "epoch": 43.952095808383234, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1205, "step": 29360 }, { "epoch": 43.953592814371255, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.119, "step": 29361 }, { "epoch": 43.955089820359284, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1188, "step": 29362 }, { "epoch": 43.956586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1261, "step": 29363 }, { "epoch": 43.958083832335326, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1153, "step": 29364 }, { "epoch": 43.959580838323355, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1234, "step": 29365 }, { "epoch": 43.961077844311376, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1219, "step": 29366 }, { "epoch": 43.962574850299404, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1282, "step": 29367 }, { "epoch": 43.964071856287426, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1199, "step": 29368 }, { "epoch": 43.96556886227545, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1253, "step": 29369 }, { "epoch": 43.967065868263475, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1215, "step": 29370 }, { "epoch": 43.9685628742515, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1255, "step": 29371 }, { "epoch": 43.97005988023952, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.12, "step": 29372 }, { "epoch": 43.971556886227546, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 29373 }, { "epoch": 43.97305389221557, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1224, "step": 29374 }, { "epoch": 43.974550898203596, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1186, "step": 29375 }, { "epoch": 43.97604790419162, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1182, "step": 29376 }, { "epoch": 43.97754491017964, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1256, "step": 29377 }, { "epoch": 43.97904191616767, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1179, "step": 29378 }, { "epoch": 43.98053892215569, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1226, "step": 29379 }, { "epoch": 43.98203592814371, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1279, "step": 29380 }, { "epoch": 43.98353293413174, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1226, "step": 29381 }, { "epoch": 43.98502994011976, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1189, "step": 29382 }, { "epoch": 43.98652694610779, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1197, "step": 29383 }, { "epoch": 43.98802395209581, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1168, "step": 29384 }, { "epoch": 43.98952095808383, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1329, "step": 29385 }, { "epoch": 43.99101796407186, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1139, "step": 29386 }, { "epoch": 43.99251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1172, "step": 29387 }, { "epoch": 43.9940119760479, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1182, "step": 29388 }, { "epoch": 43.99550898203593, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1219, "step": 29389 }, { "epoch": 43.99700598802395, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1281, "step": 29390 }, { "epoch": 43.99850299401198, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1259, "step": 29391 }, { "epoch": 44.0, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1229, "step": 29392 }, { "epoch": 44.00149700598802, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1219, "step": 29393 }, { "epoch": 44.00299401197605, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1258, "step": 29394 }, { "epoch": 44.00449101796407, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1243, "step": 29395 }, { "epoch": 44.0059880239521, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1211, "step": 29396 }, { "epoch": 44.00748502994012, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1277, "step": 29397 }, { "epoch": 44.00898203592814, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1199, "step": 29398 }, { "epoch": 44.01047904191617, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1243, "step": 29399 }, { "epoch": 44.01197604790419, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1125, "step": 29400 }, { "epoch": 44.01347305389221, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1214, "step": 29401 }, { "epoch": 44.01497005988024, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1213, "step": 29402 }, { "epoch": 44.01646706586826, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1149, "step": 29403 }, { "epoch": 44.01796407185629, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1201, "step": 29404 }, { "epoch": 44.01946107784431, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1216, "step": 29405 }, { "epoch": 44.02095808383233, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1159, "step": 29406 }, { "epoch": 44.02245508982036, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1181, "step": 29407 }, { "epoch": 44.02395209580838, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1168, "step": 29408 }, { "epoch": 44.025449101796404, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1191, "step": 29409 }, { "epoch": 44.02694610778443, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1179, "step": 29410 }, { "epoch": 44.028443113772454, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1238, "step": 29411 }, { "epoch": 44.02994011976048, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1234, "step": 29412 }, { "epoch": 44.0314371257485, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1249, "step": 29413 }, { "epoch": 44.032934131736525, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1219, "step": 29414 }, { "epoch": 44.03443113772455, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1277, "step": 29415 }, { "epoch": 44.035928143712574, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.126, "step": 29416 }, { "epoch": 44.037425149700596, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.12, "step": 29417 }, { "epoch": 44.038922155688624, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.126, "step": 29418 }, { "epoch": 44.040419161676645, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1201, "step": 29419 }, { "epoch": 44.041916167664674, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1212, "step": 29420 }, { "epoch": 44.043413173652695, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1234, "step": 29421 }, { "epoch": 44.044910179640716, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1221, "step": 29422 }, { "epoch": 44.046407185628745, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1257, "step": 29423 }, { "epoch": 44.047904191616766, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1205, "step": 29424 }, { "epoch": 44.04940119760479, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1247, "step": 29425 }, { "epoch": 44.050898203592816, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1148, "step": 29426 }, { "epoch": 44.05239520958084, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1176, "step": 29427 }, { "epoch": 44.053892215568865, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1287, "step": 29428 }, { "epoch": 44.05538922155689, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1159, "step": 29429 }, { "epoch": 44.05688622754491, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1196, "step": 29430 }, { "epoch": 44.058383233532936, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1205, "step": 29431 }, { "epoch": 44.05988023952096, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1185, "step": 29432 }, { "epoch": 44.06137724550898, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1262, "step": 29433 }, { "epoch": 44.06287425149701, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1279, "step": 29434 }, { "epoch": 44.06437125748503, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1233, "step": 29435 }, { "epoch": 44.06586826347306, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 29436 }, { "epoch": 44.06736526946108, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1168, "step": 29437 }, { "epoch": 44.0688622754491, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1242, "step": 29438 }, { "epoch": 44.07035928143713, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1276, "step": 29439 }, { "epoch": 44.07185628742515, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1197, "step": 29440 }, { "epoch": 44.07335329341317, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1191, "step": 29441 }, { "epoch": 44.0748502994012, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1215, "step": 29442 }, { "epoch": 44.07634730538922, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.126, "step": 29443 }, { "epoch": 44.07784431137725, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1241, "step": 29444 }, { "epoch": 44.07934131736527, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1199, "step": 29445 }, { "epoch": 44.08083832335329, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1178, "step": 29446 }, { "epoch": 44.08233532934132, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1273, "step": 29447 }, { "epoch": 44.08383233532934, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1203, "step": 29448 }, { "epoch": 44.08532934131737, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1169, "step": 29449 }, { "epoch": 44.08682634730539, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1245, "step": 29450 }, { "epoch": 44.08832335329341, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1216, "step": 29451 }, { "epoch": 44.08982035928144, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.128, "step": 29452 }, { "epoch": 44.09131736526946, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1221, "step": 29453 }, { "epoch": 44.09281437125748, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1199, "step": 29454 }, { "epoch": 44.09431137724551, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1179, "step": 29455 }, { "epoch": 44.09580838323353, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.123, "step": 29456 }, { "epoch": 44.09730538922156, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1224, "step": 29457 }, { "epoch": 44.09880239520958, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1274, "step": 29458 }, { "epoch": 44.1002994011976, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1236, "step": 29459 }, { "epoch": 44.10179640718563, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1235, "step": 29460 }, { "epoch": 44.10329341317365, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1225, "step": 29461 }, { "epoch": 44.104790419161674, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1175, "step": 29462 }, { "epoch": 44.1062874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 29463 }, { "epoch": 44.10778443113772, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1226, "step": 29464 }, { "epoch": 44.10928143712575, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1309, "step": 29465 }, { "epoch": 44.11077844311377, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1188, "step": 29466 }, { "epoch": 44.112275449101794, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1212, "step": 29467 }, { "epoch": 44.11377245508982, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1253, "step": 29468 }, { "epoch": 44.115269461077844, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1254, "step": 29469 }, { "epoch": 44.116766467065865, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1285, "step": 29470 }, { "epoch": 44.118263473053894, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1239, "step": 29471 }, { "epoch": 44.119760479041915, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1229, "step": 29472 }, { "epoch": 44.12125748502994, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1228, "step": 29473 }, { "epoch": 44.122754491017965, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1241, "step": 29474 }, { "epoch": 44.124251497005986, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1238, "step": 29475 }, { "epoch": 44.125748502994014, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1215, "step": 29476 }, { "epoch": 44.127245508982035, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1202, "step": 29477 }, { "epoch": 44.12874251497006, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1178, "step": 29478 }, { "epoch": 44.130239520958085, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1232, "step": 29479 }, { "epoch": 44.131736526946106, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1254, "step": 29480 }, { "epoch": 44.133233532934135, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1202, "step": 29481 }, { "epoch": 44.134730538922156, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1243, "step": 29482 }, { "epoch": 44.13622754491018, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.122, "step": 29483 }, { "epoch": 44.137724550898206, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1185, "step": 29484 }, { "epoch": 44.13922155688623, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1227, "step": 29485 }, { "epoch": 44.14071856287425, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1217, "step": 29486 }, { "epoch": 44.14221556886228, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1232, "step": 29487 }, { "epoch": 44.1437125748503, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1216, "step": 29488 }, { "epoch": 44.145209580838326, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1257, "step": 29489 }, { "epoch": 44.14670658682635, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1225, "step": 29490 }, { "epoch": 44.14820359281437, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.124, "step": 29491 }, { "epoch": 44.1497005988024, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1236, "step": 29492 }, { "epoch": 44.15119760479042, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1277, "step": 29493 }, { "epoch": 44.15269461077844, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1226, "step": 29494 }, { "epoch": 44.15419161676647, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.124, "step": 29495 }, { "epoch": 44.15568862275449, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1199, "step": 29496 }, { "epoch": 44.15718562874252, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1244, "step": 29497 }, { "epoch": 44.15868263473054, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1243, "step": 29498 }, { "epoch": 44.16017964071856, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1245, "step": 29499 }, { "epoch": 44.16167664670659, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1223, "step": 29500 }, { "epoch": 44.16317365269461, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1255, "step": 29501 }, { "epoch": 44.16467065868263, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1226, "step": 29502 }, { "epoch": 44.16616766467066, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1176, "step": 29503 }, { "epoch": 44.16766467065868, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1138, "step": 29504 }, { "epoch": 44.16916167664671, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 29505 }, { "epoch": 44.17065868263473, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1171, "step": 29506 }, { "epoch": 44.17215568862275, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.126, "step": 29507 }, { "epoch": 44.17365269461078, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1258, "step": 29508 }, { "epoch": 44.1751497005988, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1173, "step": 29509 }, { "epoch": 44.17664670658683, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1236, "step": 29510 }, { "epoch": 44.17814371257485, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1253, "step": 29511 }, { "epoch": 44.17964071856287, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1205, "step": 29512 }, { "epoch": 44.1811377245509, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1251, "step": 29513 }, { "epoch": 44.18263473053892, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1194, "step": 29514 }, { "epoch": 44.18413173652694, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1212, "step": 29515 }, { "epoch": 44.18562874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1219, "step": 29516 }, { "epoch": 44.18712574850299, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1258, "step": 29517 }, { "epoch": 44.18862275449102, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1243, "step": 29518 }, { "epoch": 44.19011976047904, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1297, "step": 29519 }, { "epoch": 44.191616766467064, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1315, "step": 29520 }, { "epoch": 44.19311377245509, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1189, "step": 29521 }, { "epoch": 44.19461077844311, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1284, "step": 29522 }, { "epoch": 44.196107784431135, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1209, "step": 29523 }, { "epoch": 44.19760479041916, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1197, "step": 29524 }, { "epoch": 44.199101796407184, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1199, "step": 29525 }, { "epoch": 44.20059880239521, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1247, "step": 29526 }, { "epoch": 44.202095808383234, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1245, "step": 29527 }, { "epoch": 44.203592814371255, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1168, "step": 29528 }, { "epoch": 44.205089820359284, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1218, "step": 29529 }, { "epoch": 44.206586826347305, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.12, "step": 29530 }, { "epoch": 44.208083832335326, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1133, "step": 29531 }, { "epoch": 44.209580838323355, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1336, "step": 29532 }, { "epoch": 44.211077844311376, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1191, "step": 29533 }, { "epoch": 44.212574850299404, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.128, "step": 29534 }, { "epoch": 44.214071856287426, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1185, "step": 29535 }, { "epoch": 44.21556886227545, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1176, "step": 29536 }, { "epoch": 44.217065868263475, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1234, "step": 29537 }, { "epoch": 44.2185628742515, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1261, "step": 29538 }, { "epoch": 44.22005988023952, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1247, "step": 29539 }, { "epoch": 44.221556886227546, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.128, "step": 29540 }, { "epoch": 44.22305389221557, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1214, "step": 29541 }, { "epoch": 44.224550898203596, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.119, "step": 29542 }, { "epoch": 44.22604790419162, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1183, "step": 29543 }, { "epoch": 44.22754491017964, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.123, "step": 29544 }, { "epoch": 44.22904191616767, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1177, "step": 29545 }, { "epoch": 44.23053892215569, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1176, "step": 29546 }, { "epoch": 44.23203592814371, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1247, "step": 29547 }, { "epoch": 44.23353293413174, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1188, "step": 29548 }, { "epoch": 44.23502994011976, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1222, "step": 29549 }, { "epoch": 44.23652694610779, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.118, "step": 29550 }, { "epoch": 44.23802395209581, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1197, "step": 29551 }, { "epoch": 44.23952095808383, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1205, "step": 29552 }, { "epoch": 44.24101796407186, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1259, "step": 29553 }, { "epoch": 44.24251497005988, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1201, "step": 29554 }, { "epoch": 44.2440119760479, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1268, "step": 29555 }, { "epoch": 44.24550898203593, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1172, "step": 29556 }, { "epoch": 44.24700598802395, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1303, "step": 29557 }, { "epoch": 44.24850299401198, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1191, "step": 29558 }, { "epoch": 44.25, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1177, "step": 29559 }, { "epoch": 44.25149700598802, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1258, "step": 29560 }, { "epoch": 44.25299401197605, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 29561 }, { "epoch": 44.25449101796407, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1193, "step": 29562 }, { "epoch": 44.2559880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.125, "step": 29563 }, { "epoch": 44.25748502994012, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1238, "step": 29564 }, { "epoch": 44.25898203592814, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1122, "step": 29565 }, { "epoch": 44.26047904191617, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1202, "step": 29566 }, { "epoch": 44.26197604790419, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1235, "step": 29567 }, { "epoch": 44.26347305389221, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1207, "step": 29568 }, { "epoch": 44.26497005988024, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1243, "step": 29569 }, { "epoch": 44.26646706586826, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1218, "step": 29570 }, { "epoch": 44.26796407185629, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1183, "step": 29571 }, { "epoch": 44.26946107784431, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1232, "step": 29572 }, { "epoch": 44.27095808383233, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1209, "step": 29573 }, { "epoch": 44.27245508982036, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1182, "step": 29574 }, { "epoch": 44.27395209580838, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1159, "step": 29575 }, { "epoch": 44.275449101796404, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1275, "step": 29576 }, { "epoch": 44.27694610778443, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1248, "step": 29577 }, { "epoch": 44.278443113772454, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1226, "step": 29578 }, { "epoch": 44.27994011976048, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1193, "step": 29579 }, { "epoch": 44.2814371257485, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.129, "step": 29580 }, { "epoch": 44.282934131736525, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1232, "step": 29581 }, { "epoch": 44.28443113772455, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1144, "step": 29582 }, { "epoch": 44.285928143712574, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1234, "step": 29583 }, { "epoch": 44.287425149700596, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.116, "step": 29584 }, { "epoch": 44.288922155688624, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1235, "step": 29585 }, { "epoch": 44.290419161676645, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1248, "step": 29586 }, { "epoch": 44.291916167664674, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1233, "step": 29587 }, { "epoch": 44.293413173652695, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1181, "step": 29588 }, { "epoch": 44.294910179640716, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1244, "step": 29589 }, { "epoch": 44.296407185628745, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1196, "step": 29590 }, { "epoch": 44.297904191616766, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1254, "step": 29591 }, { "epoch": 44.29940119760479, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.126, "step": 29592 }, { "epoch": 44.300898203592816, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1229, "step": 29593 }, { "epoch": 44.30239520958084, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1203, "step": 29594 }, { "epoch": 44.303892215568865, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1194, "step": 29595 }, { "epoch": 44.30538922155689, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1212, "step": 29596 }, { "epoch": 44.30688622754491, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1198, "step": 29597 }, { "epoch": 44.308383233532936, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1228, "step": 29598 }, { "epoch": 44.30988023952096, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1252, "step": 29599 }, { "epoch": 44.31137724550898, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1242, "step": 29600 }, { "epoch": 44.31287425149701, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1211, "step": 29601 }, { "epoch": 44.31437125748503, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1245, "step": 29602 }, { "epoch": 44.31586826347306, "grad_norm": 0.04736328125, "learning_rate": 0.0008, "loss": 1.1191, "step": 29603 }, { "epoch": 44.31736526946108, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.119, "step": 29604 }, { "epoch": 44.3188622754491, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1238, "step": 29605 }, { "epoch": 44.32035928143713, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.123, "step": 29606 }, { "epoch": 44.32185628742515, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1275, "step": 29607 }, { "epoch": 44.32335329341317, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1216, "step": 29608 }, { "epoch": 44.3248502994012, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1216, "step": 29609 }, { "epoch": 44.32634730538922, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1249, "step": 29610 }, { "epoch": 44.32784431137725, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1191, "step": 29611 }, { "epoch": 44.32934131736527, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1224, "step": 29612 }, { "epoch": 44.33083832335329, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1194, "step": 29613 }, { "epoch": 44.33233532934132, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1242, "step": 29614 }, { "epoch": 44.33383233532934, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1219, "step": 29615 }, { "epoch": 44.33532934131736, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.125, "step": 29616 }, { "epoch": 44.33682634730539, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1207, "step": 29617 }, { "epoch": 44.33832335329341, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1291, "step": 29618 }, { "epoch": 44.33982035928144, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1172, "step": 29619 }, { "epoch": 44.34131736526946, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1221, "step": 29620 }, { "epoch": 44.34281437125748, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1172, "step": 29621 }, { "epoch": 44.34431137724551, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.121, "step": 29622 }, { "epoch": 44.34580838323353, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1218, "step": 29623 }, { "epoch": 44.34730538922156, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1245, "step": 29624 }, { "epoch": 44.34880239520958, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1242, "step": 29625 }, { "epoch": 44.3502994011976, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1202, "step": 29626 }, { "epoch": 44.35179640718563, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1222, "step": 29627 }, { "epoch": 44.35329341317365, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.121, "step": 29628 }, { "epoch": 44.354790419161674, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1235, "step": 29629 }, { "epoch": 44.3562874251497, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1173, "step": 29630 }, { "epoch": 44.35778443113772, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1183, "step": 29631 }, { "epoch": 44.35928143712575, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1236, "step": 29632 }, { "epoch": 44.36077844311377, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1165, "step": 29633 }, { "epoch": 44.362275449101794, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1291, "step": 29634 }, { "epoch": 44.36377245508982, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.127, "step": 29635 }, { "epoch": 44.365269461077844, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1164, "step": 29636 }, { "epoch": 44.366766467065865, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1253, "step": 29637 }, { "epoch": 44.368263473053894, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1283, "step": 29638 }, { "epoch": 44.369760479041915, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 29639 }, { "epoch": 44.37125748502994, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1208, "step": 29640 }, { "epoch": 44.372754491017965, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1294, "step": 29641 }, { "epoch": 44.374251497005986, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1245, "step": 29642 }, { "epoch": 44.375748502994014, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.119, "step": 29643 }, { "epoch": 44.377245508982035, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1254, "step": 29644 }, { "epoch": 44.37874251497006, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1256, "step": 29645 }, { "epoch": 44.380239520958085, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1269, "step": 29646 }, { "epoch": 44.381736526946106, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 29647 }, { "epoch": 44.383233532934135, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1278, "step": 29648 }, { "epoch": 44.384730538922156, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1297, "step": 29649 }, { "epoch": 44.38622754491018, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1261, "step": 29650 }, { "epoch": 44.387724550898206, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1226, "step": 29651 }, { "epoch": 44.38922155688623, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1267, "step": 29652 }, { "epoch": 44.39071856287425, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1196, "step": 29653 }, { "epoch": 44.39221556886228, "grad_norm": 0.047119140625, "learning_rate": 0.0008, "loss": 1.1207, "step": 29654 }, { "epoch": 44.3937125748503, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1219, "step": 29655 }, { "epoch": 44.395209580838326, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1268, "step": 29656 }, { "epoch": 44.39670658682635, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1157, "step": 29657 }, { "epoch": 44.39820359281437, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1248, "step": 29658 }, { "epoch": 44.3997005988024, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.127, "step": 29659 }, { "epoch": 44.40119760479042, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1205, "step": 29660 }, { "epoch": 44.40269461077844, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1207, "step": 29661 }, { "epoch": 44.40419161676647, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.126, "step": 29662 }, { "epoch": 44.40568862275449, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1246, "step": 29663 }, { "epoch": 44.40718562874252, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1189, "step": 29664 }, { "epoch": 44.40868263473054, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1261, "step": 29665 }, { "epoch": 44.41017964071856, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 29666 }, { "epoch": 44.41167664670659, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1264, "step": 29667 }, { "epoch": 44.41317365269461, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1216, "step": 29668 }, { "epoch": 44.41467065868264, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.125, "step": 29669 }, { "epoch": 44.41616766467066, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1255, "step": 29670 }, { "epoch": 44.41766467065868, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1253, "step": 29671 }, { "epoch": 44.41916167664671, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1216, "step": 29672 }, { "epoch": 44.42065868263473, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1213, "step": 29673 }, { "epoch": 44.42215568862275, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1195, "step": 29674 }, { "epoch": 44.42365269461078, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1222, "step": 29675 }, { "epoch": 44.4251497005988, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1238, "step": 29676 }, { "epoch": 44.42664670658683, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1202, "step": 29677 }, { "epoch": 44.42814371257485, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1245, "step": 29678 }, { "epoch": 44.42964071856287, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1239, "step": 29679 }, { "epoch": 44.4311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1186, "step": 29680 }, { "epoch": 44.43263473053892, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1212, "step": 29681 }, { "epoch": 44.43413173652694, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1205, "step": 29682 }, { "epoch": 44.43562874251497, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1251, "step": 29683 }, { "epoch": 44.43712574850299, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1204, "step": 29684 }, { "epoch": 44.43862275449102, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1163, "step": 29685 }, { "epoch": 44.44011976047904, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.125, "step": 29686 }, { "epoch": 44.441616766467064, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1174, "step": 29687 }, { "epoch": 44.44311377245509, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1257, "step": 29688 }, { "epoch": 44.44461077844311, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1305, "step": 29689 }, { "epoch": 44.446107784431135, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1181, "step": 29690 }, { "epoch": 44.44760479041916, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1169, "step": 29691 }, { "epoch": 44.449101796407184, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.124, "step": 29692 }, { "epoch": 44.45059880239521, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1263, "step": 29693 }, { "epoch": 44.452095808383234, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.122, "step": 29694 }, { "epoch": 44.453592814371255, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.126, "step": 29695 }, { "epoch": 44.455089820359284, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1217, "step": 29696 }, { "epoch": 44.456586826347305, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1199, "step": 29697 }, { "epoch": 44.458083832335326, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1257, "step": 29698 }, { "epoch": 44.459580838323355, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1248, "step": 29699 }, { "epoch": 44.461077844311376, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1192, "step": 29700 }, { "epoch": 44.462574850299404, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1277, "step": 29701 }, { "epoch": 44.464071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.12, "step": 29702 }, { "epoch": 44.46556886227545, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1355, "step": 29703 }, { "epoch": 44.467065868263475, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1215, "step": 29704 }, { "epoch": 44.4685628742515, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1266, "step": 29705 }, { "epoch": 44.47005988023952, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1243, "step": 29706 }, { "epoch": 44.471556886227546, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1252, "step": 29707 }, { "epoch": 44.47305389221557, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1165, "step": 29708 }, { "epoch": 44.474550898203596, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1256, "step": 29709 }, { "epoch": 44.47604790419162, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1218, "step": 29710 }, { "epoch": 44.47754491017964, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1182, "step": 29711 }, { "epoch": 44.47904191616767, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1207, "step": 29712 }, { "epoch": 44.48053892215569, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1179, "step": 29713 }, { "epoch": 44.48203592814371, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1235, "step": 29714 }, { "epoch": 44.48353293413174, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1253, "step": 29715 }, { "epoch": 44.48502994011976, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1224, "step": 29716 }, { "epoch": 44.48652694610779, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1284, "step": 29717 }, { "epoch": 44.48802395209581, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1244, "step": 29718 }, { "epoch": 44.48952095808383, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1261, "step": 29719 }, { "epoch": 44.49101796407186, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1259, "step": 29720 }, { "epoch": 44.49251497005988, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1194, "step": 29721 }, { "epoch": 44.4940119760479, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1263, "step": 29722 }, { "epoch": 44.49550898203593, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1205, "step": 29723 }, { "epoch": 44.49700598802395, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.121, "step": 29724 }, { "epoch": 44.49850299401198, "grad_norm": 0.05029296875, "learning_rate": 0.0008, "loss": 1.1208, "step": 29725 }, { "epoch": 44.5, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1237, "step": 29726 }, { "epoch": 44.50149700598802, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1187, "step": 29727 }, { "epoch": 44.50299401197605, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1204, "step": 29728 }, { "epoch": 44.50449101796407, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1241, "step": 29729 }, { "epoch": 44.5059880239521, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1223, "step": 29730 }, { "epoch": 44.50748502994012, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.12, "step": 29731 }, { "epoch": 44.50898203592814, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.123, "step": 29732 }, { "epoch": 44.51047904191617, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1194, "step": 29733 }, { "epoch": 44.51197604790419, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1199, "step": 29734 }, { "epoch": 44.51347305389221, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1148, "step": 29735 }, { "epoch": 44.51497005988024, "grad_norm": 0.049560546875, "learning_rate": 0.0008, "loss": 1.1184, "step": 29736 }, { "epoch": 44.51646706586826, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1143, "step": 29737 }, { "epoch": 44.51796407185629, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.124, "step": 29738 }, { "epoch": 44.51946107784431, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1219, "step": 29739 }, { "epoch": 44.52095808383233, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1241, "step": 29740 }, { "epoch": 44.52245508982036, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1285, "step": 29741 }, { "epoch": 44.52395209580838, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1266, "step": 29742 }, { "epoch": 44.525449101796404, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1262, "step": 29743 }, { "epoch": 44.52694610778443, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1196, "step": 29744 }, { "epoch": 44.528443113772454, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1167, "step": 29745 }, { "epoch": 44.52994011976048, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1249, "step": 29746 }, { "epoch": 44.5314371257485, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1214, "step": 29747 }, { "epoch": 44.532934131736525, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1238, "step": 29748 }, { "epoch": 44.53443113772455, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1249, "step": 29749 }, { "epoch": 44.535928143712574, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1233, "step": 29750 }, { "epoch": 44.537425149700596, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1241, "step": 29751 }, { "epoch": 44.538922155688624, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1245, "step": 29752 }, { "epoch": 44.540419161676645, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1324, "step": 29753 }, { "epoch": 44.541916167664674, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1249, "step": 29754 }, { "epoch": 44.543413173652695, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1283, "step": 29755 }, { "epoch": 44.544910179640716, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 29756 }, { "epoch": 44.546407185628745, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.123, "step": 29757 }, { "epoch": 44.547904191616766, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1259, "step": 29758 }, { "epoch": 44.54940119760479, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1217, "step": 29759 }, { "epoch": 44.550898203592816, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1245, "step": 29760 }, { "epoch": 44.55239520958084, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1297, "step": 29761 }, { "epoch": 44.553892215568865, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1211, "step": 29762 }, { "epoch": 44.55538922155689, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1196, "step": 29763 }, { "epoch": 44.55688622754491, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 29764 }, { "epoch": 44.558383233532936, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1275, "step": 29765 }, { "epoch": 44.55988023952096, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1216, "step": 29766 }, { "epoch": 44.56137724550898, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1283, "step": 29767 }, { "epoch": 44.56287425149701, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1222, "step": 29768 }, { "epoch": 44.56437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1175, "step": 29769 }, { "epoch": 44.56586826347306, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.122, "step": 29770 }, { "epoch": 44.56736526946108, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1277, "step": 29771 }, { "epoch": 44.5688622754491, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.119, "step": 29772 }, { "epoch": 44.57035928143713, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1228, "step": 29773 }, { "epoch": 44.57185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1168, "step": 29774 }, { "epoch": 44.57335329341317, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1266, "step": 29775 }, { "epoch": 44.5748502994012, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1217, "step": 29776 }, { "epoch": 44.57634730538922, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1231, "step": 29777 }, { "epoch": 44.57784431137725, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1257, "step": 29778 }, { "epoch": 44.57934131736527, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1279, "step": 29779 }, { "epoch": 44.58083832335329, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1188, "step": 29780 }, { "epoch": 44.58233532934132, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1165, "step": 29781 }, { "epoch": 44.58383233532934, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.123, "step": 29782 }, { "epoch": 44.58532934131736, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1244, "step": 29783 }, { "epoch": 44.58682634730539, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1201, "step": 29784 }, { "epoch": 44.58832335329341, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1246, "step": 29785 }, { "epoch": 44.58982035928144, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1194, "step": 29786 }, { "epoch": 44.59131736526946, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.128, "step": 29787 }, { "epoch": 44.59281437125748, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1202, "step": 29788 }, { "epoch": 44.59431137724551, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1243, "step": 29789 }, { "epoch": 44.59580838323353, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1212, "step": 29790 }, { "epoch": 44.59730538922156, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1191, "step": 29791 }, { "epoch": 44.59880239520958, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1259, "step": 29792 }, { "epoch": 44.6002994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.116, "step": 29793 }, { "epoch": 44.60179640718563, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1238, "step": 29794 }, { "epoch": 44.60329341317365, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1233, "step": 29795 }, { "epoch": 44.604790419161674, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1262, "step": 29796 }, { "epoch": 44.6062874251497, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1225, "step": 29797 }, { "epoch": 44.60778443113772, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1235, "step": 29798 }, { "epoch": 44.60928143712575, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1205, "step": 29799 }, { "epoch": 44.61077844311377, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1192, "step": 29800 }, { "epoch": 44.612275449101794, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1206, "step": 29801 }, { "epoch": 44.61377245508982, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1229, "step": 29802 }, { "epoch": 44.615269461077844, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1185, "step": 29803 }, { "epoch": 44.616766467065865, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1158, "step": 29804 }, { "epoch": 44.618263473053894, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1176, "step": 29805 }, { "epoch": 44.619760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1177, "step": 29806 }, { "epoch": 44.62125748502994, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1209, "step": 29807 }, { "epoch": 44.622754491017965, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1195, "step": 29808 }, { "epoch": 44.624251497005986, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1218, "step": 29809 }, { "epoch": 44.625748502994014, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1184, "step": 29810 }, { "epoch": 44.627245508982035, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1273, "step": 29811 }, { "epoch": 44.62874251497006, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1185, "step": 29812 }, { "epoch": 44.630239520958085, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1237, "step": 29813 }, { "epoch": 44.631736526946106, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1218, "step": 29814 }, { "epoch": 44.633233532934135, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1203, "step": 29815 }, { "epoch": 44.634730538922156, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.126, "step": 29816 }, { "epoch": 44.63622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1252, "step": 29817 }, { "epoch": 44.637724550898206, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1202, "step": 29818 }, { "epoch": 44.63922155688623, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.121, "step": 29819 }, { "epoch": 44.64071856287425, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1187, "step": 29820 }, { "epoch": 44.64221556886228, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1278, "step": 29821 }, { "epoch": 44.6437125748503, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1302, "step": 29822 }, { "epoch": 44.645209580838326, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1238, "step": 29823 }, { "epoch": 44.64670658682635, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.12, "step": 29824 }, { "epoch": 44.64820359281437, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1189, "step": 29825 }, { "epoch": 44.6497005988024, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 29826 }, { "epoch": 44.65119760479042, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1137, "step": 29827 }, { "epoch": 44.65269461077844, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1283, "step": 29828 }, { "epoch": 44.65419161676647, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.128, "step": 29829 }, { "epoch": 44.65568862275449, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.124, "step": 29830 }, { "epoch": 44.65718562874252, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1268, "step": 29831 }, { "epoch": 44.65868263473054, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1158, "step": 29832 }, { "epoch": 44.66017964071856, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1218, "step": 29833 }, { "epoch": 44.66167664670659, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1211, "step": 29834 }, { "epoch": 44.66317365269461, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1228, "step": 29835 }, { "epoch": 44.66467065868264, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1292, "step": 29836 }, { "epoch": 44.66616766467066, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1254, "step": 29837 }, { "epoch": 44.66766467065868, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.121, "step": 29838 }, { "epoch": 44.66916167664671, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1228, "step": 29839 }, { "epoch": 44.67065868263473, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1246, "step": 29840 }, { "epoch": 44.67215568862275, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1209, "step": 29841 }, { "epoch": 44.67365269461078, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1214, "step": 29842 }, { "epoch": 44.6751497005988, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1189, "step": 29843 }, { "epoch": 44.67664670658683, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1218, "step": 29844 }, { "epoch": 44.67814371257485, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.125, "step": 29845 }, { "epoch": 44.67964071856287, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1211, "step": 29846 }, { "epoch": 44.6811377245509, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1255, "step": 29847 }, { "epoch": 44.68263473053892, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1255, "step": 29848 }, { "epoch": 44.68413173652694, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1238, "step": 29849 }, { "epoch": 44.68562874251497, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.121, "step": 29850 }, { "epoch": 44.68712574850299, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.124, "step": 29851 }, { "epoch": 44.68862275449102, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1216, "step": 29852 }, { "epoch": 44.69011976047904, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1255, "step": 29853 }, { "epoch": 44.691616766467064, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1266, "step": 29854 }, { "epoch": 44.69311377245509, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.121, "step": 29855 }, { "epoch": 44.69461077844311, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.12, "step": 29856 }, { "epoch": 44.696107784431135, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1162, "step": 29857 }, { "epoch": 44.69760479041916, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1238, "step": 29858 }, { "epoch": 44.699101796407184, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1251, "step": 29859 }, { "epoch": 44.70059880239521, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1196, "step": 29860 }, { "epoch": 44.702095808383234, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1192, "step": 29861 }, { "epoch": 44.703592814371255, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1237, "step": 29862 }, { "epoch": 44.705089820359284, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1154, "step": 29863 }, { "epoch": 44.706586826347305, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1175, "step": 29864 }, { "epoch": 44.708083832335326, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1259, "step": 29865 }, { "epoch": 44.709580838323355, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1177, "step": 29866 }, { "epoch": 44.711077844311376, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1271, "step": 29867 }, { "epoch": 44.712574850299404, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1228, "step": 29868 }, { "epoch": 44.714071856287426, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 29869 }, { "epoch": 44.71556886227545, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1312, "step": 29870 }, { "epoch": 44.717065868263475, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1231, "step": 29871 }, { "epoch": 44.7185628742515, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1212, "step": 29872 }, { "epoch": 44.72005988023952, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1156, "step": 29873 }, { "epoch": 44.721556886227546, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1158, "step": 29874 }, { "epoch": 44.72305389221557, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1203, "step": 29875 }, { "epoch": 44.724550898203596, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1168, "step": 29876 }, { "epoch": 44.72604790419162, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1307, "step": 29877 }, { "epoch": 44.72754491017964, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.125, "step": 29878 }, { "epoch": 44.72904191616767, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1227, "step": 29879 }, { "epoch": 44.73053892215569, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1177, "step": 29880 }, { "epoch": 44.73203592814371, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1169, "step": 29881 }, { "epoch": 44.73353293413174, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1219, "step": 29882 }, { "epoch": 44.73502994011976, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1248, "step": 29883 }, { "epoch": 44.73652694610779, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1214, "step": 29884 }, { "epoch": 44.73802395209581, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1194, "step": 29885 }, { "epoch": 44.73952095808383, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1255, "step": 29886 }, { "epoch": 44.74101796407186, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1293, "step": 29887 }, { "epoch": 44.74251497005988, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1232, "step": 29888 }, { "epoch": 44.7440119760479, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1283, "step": 29889 }, { "epoch": 44.74550898203593, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1268, "step": 29890 }, { "epoch": 44.74700598802395, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1283, "step": 29891 }, { "epoch": 44.74850299401198, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1266, "step": 29892 }, { "epoch": 44.75, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1245, "step": 29893 }, { "epoch": 44.75149700598802, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1143, "step": 29894 }, { "epoch": 44.75299401197605, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1276, "step": 29895 }, { "epoch": 44.75449101796407, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.126, "step": 29896 }, { "epoch": 44.7559880239521, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1212, "step": 29897 }, { "epoch": 44.75748502994012, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1222, "step": 29898 }, { "epoch": 44.75898203592814, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.125, "step": 29899 }, { "epoch": 44.76047904191617, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1212, "step": 29900 }, { "epoch": 44.76197604790419, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1241, "step": 29901 }, { "epoch": 44.76347305389221, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1234, "step": 29902 }, { "epoch": 44.76497005988024, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1261, "step": 29903 }, { "epoch": 44.76646706586826, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1265, "step": 29904 }, { "epoch": 44.76796407185629, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1206, "step": 29905 }, { "epoch": 44.76946107784431, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1274, "step": 29906 }, { "epoch": 44.77095808383233, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1238, "step": 29907 }, { "epoch": 44.77245508982036, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1239, "step": 29908 }, { "epoch": 44.77395209580838, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.125, "step": 29909 }, { "epoch": 44.775449101796404, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.118, "step": 29910 }, { "epoch": 44.77694610778443, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1228, "step": 29911 }, { "epoch": 44.778443113772454, "grad_norm": 0.045654296875, "learning_rate": 0.0008, "loss": 1.1233, "step": 29912 }, { "epoch": 44.77994011976048, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1215, "step": 29913 }, { "epoch": 44.7814371257485, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1171, "step": 29914 }, { "epoch": 44.782934131736525, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1202, "step": 29915 }, { "epoch": 44.78443113772455, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1277, "step": 29916 }, { "epoch": 44.785928143712574, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1196, "step": 29917 }, { "epoch": 44.787425149700596, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1235, "step": 29918 }, { "epoch": 44.788922155688624, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.121, "step": 29919 }, { "epoch": 44.790419161676645, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1233, "step": 29920 }, { "epoch": 44.791916167664674, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1195, "step": 29921 }, { "epoch": 44.793413173652695, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1189, "step": 29922 }, { "epoch": 44.794910179640716, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1213, "step": 29923 }, { "epoch": 44.796407185628745, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1299, "step": 29924 }, { "epoch": 44.797904191616766, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1216, "step": 29925 }, { "epoch": 44.79940119760479, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1278, "step": 29926 }, { "epoch": 44.800898203592816, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1143, "step": 29927 }, { "epoch": 44.80239520958084, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1235, "step": 29928 }, { "epoch": 44.803892215568865, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1217, "step": 29929 }, { "epoch": 44.80538922155689, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1279, "step": 29930 }, { "epoch": 44.80688622754491, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1265, "step": 29931 }, { "epoch": 44.808383233532936, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1198, "step": 29932 }, { "epoch": 44.80988023952096, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1266, "step": 29933 }, { "epoch": 44.81137724550898, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1244, "step": 29934 }, { "epoch": 44.81287425149701, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1187, "step": 29935 }, { "epoch": 44.81437125748503, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1231, "step": 29936 }, { "epoch": 44.81586826347306, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1286, "step": 29937 }, { "epoch": 44.81736526946108, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1219, "step": 29938 }, { "epoch": 44.8188622754491, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1194, "step": 29939 }, { "epoch": 44.82035928143713, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.125, "step": 29940 }, { "epoch": 44.82185628742515, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1268, "step": 29941 }, { "epoch": 44.82335329341317, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1224, "step": 29942 }, { "epoch": 44.8248502994012, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1245, "step": 29943 }, { "epoch": 44.82634730538922, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1212, "step": 29944 }, { "epoch": 44.82784431137725, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1233, "step": 29945 }, { "epoch": 44.82934131736527, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1226, "step": 29946 }, { "epoch": 44.83083832335329, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1204, "step": 29947 }, { "epoch": 44.83233532934132, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.127, "step": 29948 }, { "epoch": 44.83383233532934, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1253, "step": 29949 }, { "epoch": 44.83532934131736, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1217, "step": 29950 }, { "epoch": 44.83682634730539, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1272, "step": 29951 }, { "epoch": 44.83832335329341, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1256, "step": 29952 }, { "epoch": 44.83982035928144, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1256, "step": 29953 }, { "epoch": 44.84131736526946, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.121, "step": 29954 }, { "epoch": 44.84281437125748, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1185, "step": 29955 }, { "epoch": 44.84431137724551, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1209, "step": 29956 }, { "epoch": 44.84580838323353, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1186, "step": 29957 }, { "epoch": 44.84730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1183, "step": 29958 }, { "epoch": 44.84880239520958, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1209, "step": 29959 }, { "epoch": 44.8502994011976, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.12, "step": 29960 }, { "epoch": 44.85179640718563, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1263, "step": 29961 }, { "epoch": 44.85329341317365, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1251, "step": 29962 }, { "epoch": 44.854790419161674, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1165, "step": 29963 }, { "epoch": 44.8562874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1219, "step": 29964 }, { "epoch": 44.85778443113772, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1221, "step": 29965 }, { "epoch": 44.85928143712575, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1248, "step": 29966 }, { "epoch": 44.86077844311377, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1247, "step": 29967 }, { "epoch": 44.862275449101794, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1256, "step": 29968 }, { "epoch": 44.86377245508982, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1274, "step": 29969 }, { "epoch": 44.865269461077844, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1215, "step": 29970 }, { "epoch": 44.866766467065865, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1183, "step": 29971 }, { "epoch": 44.868263473053894, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1165, "step": 29972 }, { "epoch": 44.869760479041915, "grad_norm": 0.049560546875, "learning_rate": 0.0008, "loss": 1.1242, "step": 29973 }, { "epoch": 44.87125748502994, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1173, "step": 29974 }, { "epoch": 44.872754491017965, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1242, "step": 29975 }, { "epoch": 44.874251497005986, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1241, "step": 29976 }, { "epoch": 44.875748502994014, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1181, "step": 29977 }, { "epoch": 44.877245508982035, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1206, "step": 29978 }, { "epoch": 44.87874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1199, "step": 29979 }, { "epoch": 44.880239520958085, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1229, "step": 29980 }, { "epoch": 44.881736526946106, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1245, "step": 29981 }, { "epoch": 44.883233532934135, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 29982 }, { "epoch": 44.884730538922156, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1215, "step": 29983 }, { "epoch": 44.88622754491018, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1242, "step": 29984 }, { "epoch": 44.887724550898206, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1232, "step": 29985 }, { "epoch": 44.88922155688623, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1174, "step": 29986 }, { "epoch": 44.89071856287425, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1262, "step": 29987 }, { "epoch": 44.89221556886228, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1257, "step": 29988 }, { "epoch": 44.8937125748503, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1259, "step": 29989 }, { "epoch": 44.895209580838326, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1161, "step": 29990 }, { "epoch": 44.89670658682635, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1246, "step": 29991 }, { "epoch": 44.89820359281437, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1284, "step": 29992 }, { "epoch": 44.8997005988024, "grad_norm": 0.049072265625, "learning_rate": 0.0008, "loss": 1.1241, "step": 29993 }, { "epoch": 44.90119760479042, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1238, "step": 29994 }, { "epoch": 44.90269461077844, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1242, "step": 29995 }, { "epoch": 44.90419161676647, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1308, "step": 29996 }, { "epoch": 44.90568862275449, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1252, "step": 29997 }, { "epoch": 44.90718562874252, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1183, "step": 29998 }, { "epoch": 44.90868263473054, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.12, "step": 29999 }, { "epoch": 44.91017964071856, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1209, "step": 30000 } ], "logging_steps": 1.0, "max_steps": 1002000, "num_input_tokens_seen": 0, "num_train_epochs": 1500, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.7895174231092625e+19, "train_batch_size": 120, "trial_name": null, "trial_params": null }