| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.9996064541519087, |
| "eval_steps": 500, |
| "global_step": 1270, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0007870916961826052, |
| "grad_norm": 4.882866791973475, |
| "learning_rate": 7.8125e-08, |
| "loss": 0.357, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.0015741833923652105, |
| "grad_norm": 4.89981767485179, |
| "learning_rate": 1.5625e-07, |
| "loss": 0.3398, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.0023612750885478157, |
| "grad_norm": 4.908915346462736, |
| "learning_rate": 2.3437500000000003e-07, |
| "loss": 0.3326, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.003148366784730421, |
| "grad_norm": 4.731649267914947, |
| "learning_rate": 3.125e-07, |
| "loss": 0.342, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.003935458480913027, |
| "grad_norm": 4.769633474207938, |
| "learning_rate": 3.90625e-07, |
| "loss": 0.3431, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.004722550177095631, |
| "grad_norm": 4.754569879633701, |
| "learning_rate": 4.6875000000000006e-07, |
| "loss": 0.3369, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.005509641873278237, |
| "grad_norm": 4.354074850343827, |
| "learning_rate": 5.468750000000001e-07, |
| "loss": 0.3416, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.006296733569460842, |
| "grad_norm": 4.51384309102365, |
| "learning_rate": 6.25e-07, |
| "loss": 0.3577, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.0070838252656434475, |
| "grad_norm": 4.127868399899779, |
| "learning_rate": 7.03125e-07, |
| "loss": 0.3447, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.007870916961826053, |
| "grad_norm": 4.268670776824985, |
| "learning_rate": 7.8125e-07, |
| "loss": 0.318, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.008658008658008658, |
| "grad_norm": 3.7083716156028674, |
| "learning_rate": 8.59375e-07, |
| "loss": 0.3106, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.009445100354191263, |
| "grad_norm": 3.1545864445099263, |
| "learning_rate": 9.375000000000001e-07, |
| "loss": 0.3131, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.01023219205037387, |
| "grad_norm": 3.020045714405798, |
| "learning_rate": 1.0156250000000001e-06, |
| "loss": 0.3215, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.011019283746556474, |
| "grad_norm": 2.748876831681126, |
| "learning_rate": 1.0937500000000001e-06, |
| "loss": 0.3072, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.011806375442739079, |
| "grad_norm": 2.2307366833759485, |
| "learning_rate": 1.1718750000000001e-06, |
| "loss": 0.2922, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.012593467138921684, |
| "grad_norm": 2.219422516987874, |
| "learning_rate": 1.25e-06, |
| "loss": 0.2842, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.01338055883510429, |
| "grad_norm": 2.7073031779339973, |
| "learning_rate": 1.328125e-06, |
| "loss": 0.2674, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.014167650531286895, |
| "grad_norm": 2.873035911017537, |
| "learning_rate": 1.40625e-06, |
| "loss": 0.294, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.0149547422274695, |
| "grad_norm": 2.119880363778339, |
| "learning_rate": 1.484375e-06, |
| "loss": 0.2693, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.015741833923652106, |
| "grad_norm": 1.7740660860958901, |
| "learning_rate": 1.5625e-06, |
| "loss": 0.2607, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.01652892561983471, |
| "grad_norm": 1.654838099179133, |
| "learning_rate": 1.640625e-06, |
| "loss": 0.2539, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.017316017316017316, |
| "grad_norm": 2.1067372096520884, |
| "learning_rate": 1.71875e-06, |
| "loss": 0.274, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.01810310901219992, |
| "grad_norm": 2.227492365997846, |
| "learning_rate": 1.796875e-06, |
| "loss": 0.2667, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.018890200708382526, |
| "grad_norm": 1.9818482942437547, |
| "learning_rate": 1.8750000000000003e-06, |
| "loss": 0.2507, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.01967729240456513, |
| "grad_norm": 1.9916019664977938, |
| "learning_rate": 1.953125e-06, |
| "loss": 0.2302, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.02046438410074774, |
| "grad_norm": 2.0987871479467533, |
| "learning_rate": 2.0312500000000002e-06, |
| "loss": 0.2563, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.021251475796930343, |
| "grad_norm": 1.7851505967742112, |
| "learning_rate": 2.109375e-06, |
| "loss": 0.2432, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.02203856749311295, |
| "grad_norm": 1.6067598902195293, |
| "learning_rate": 2.1875000000000002e-06, |
| "loss": 0.2466, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.022825659189295553, |
| "grad_norm": 1.436243142469347, |
| "learning_rate": 2.265625e-06, |
| "loss": 0.2486, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.023612750885478158, |
| "grad_norm": 1.631080710695958, |
| "learning_rate": 2.3437500000000002e-06, |
| "loss": 0.2692, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.024399842581660763, |
| "grad_norm": 1.42554302342302, |
| "learning_rate": 2.421875e-06, |
| "loss": 0.2374, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.025186934277843367, |
| "grad_norm": 1.479794666013743, |
| "learning_rate": 2.5e-06, |
| "loss": 0.238, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.025974025974025976, |
| "grad_norm": 1.3857185652178832, |
| "learning_rate": 2.5781250000000004e-06, |
| "loss": 0.2366, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.02676111767020858, |
| "grad_norm": 1.335993998237778, |
| "learning_rate": 2.65625e-06, |
| "loss": 0.2251, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.027548209366391185, |
| "grad_norm": 1.5950255189913525, |
| "learning_rate": 2.7343750000000004e-06, |
| "loss": 0.2506, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.02833530106257379, |
| "grad_norm": 1.3773024411686232, |
| "learning_rate": 2.8125e-06, |
| "loss": 0.2107, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.029122392758756395, |
| "grad_norm": 1.391558709917223, |
| "learning_rate": 2.8906250000000004e-06, |
| "loss": 0.236, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.029909484454939, |
| "grad_norm": 1.4317153691023394, |
| "learning_rate": 2.96875e-06, |
| "loss": 0.2369, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.030696576151121605, |
| "grad_norm": 1.5982850208202695, |
| "learning_rate": 3.0468750000000004e-06, |
| "loss": 0.2135, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.03148366784730421, |
| "grad_norm": 1.2463748947443163, |
| "learning_rate": 3.125e-06, |
| "loss": 0.2109, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.032270759543486814, |
| "grad_norm": 1.489195263138514, |
| "learning_rate": 3.2031250000000004e-06, |
| "loss": 0.2254, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.03305785123966942, |
| "grad_norm": 1.1952984228039816, |
| "learning_rate": 3.28125e-06, |
| "loss": 0.2124, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.033844942935852024, |
| "grad_norm": 1.3331698750786545, |
| "learning_rate": 3.3593750000000003e-06, |
| "loss": 0.2192, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.03463203463203463, |
| "grad_norm": 1.3944936006633961, |
| "learning_rate": 3.4375e-06, |
| "loss": 0.2024, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.03541912632821724, |
| "grad_norm": 1.3992213437238004, |
| "learning_rate": 3.5156250000000003e-06, |
| "loss": 0.2274, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.03620621802439984, |
| "grad_norm": 1.3664016160053327, |
| "learning_rate": 3.59375e-06, |
| "loss": 0.2152, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.03699330972058245, |
| "grad_norm": 1.4891884814728509, |
| "learning_rate": 3.6718750000000003e-06, |
| "loss": 0.2292, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.03778040141676505, |
| "grad_norm": 1.3270512221194979, |
| "learning_rate": 3.7500000000000005e-06, |
| "loss": 0.1997, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.03856749311294766, |
| "grad_norm": 1.4002885202427642, |
| "learning_rate": 3.828125000000001e-06, |
| "loss": 0.2049, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.03935458480913026, |
| "grad_norm": 1.3129491965923514, |
| "learning_rate": 3.90625e-06, |
| "loss": 0.2108, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.04014167650531287, |
| "grad_norm": 1.5308223134960726, |
| "learning_rate": 3.984375e-06, |
| "loss": 0.2185, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.04092876820149548, |
| "grad_norm": 1.4637864330743848, |
| "learning_rate": 4.0625000000000005e-06, |
| "loss": 0.1959, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.04171585989767808, |
| "grad_norm": 1.3727993193826824, |
| "learning_rate": 4.140625000000001e-06, |
| "loss": 0.2079, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.04250295159386069, |
| "grad_norm": 1.324719262777532, |
| "learning_rate": 4.21875e-06, |
| "loss": 0.1986, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.04329004329004329, |
| "grad_norm": 1.5169702547195179, |
| "learning_rate": 4.296875e-06, |
| "loss": 0.2233, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.0440771349862259, |
| "grad_norm": 1.2762861570952524, |
| "learning_rate": 4.3750000000000005e-06, |
| "loss": 0.1887, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.0448642266824085, |
| "grad_norm": 1.3276694806352698, |
| "learning_rate": 4.453125000000001e-06, |
| "loss": 0.2039, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.045651318378591106, |
| "grad_norm": 1.3146250055299598, |
| "learning_rate": 4.53125e-06, |
| "loss": 0.2142, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.046438410074773714, |
| "grad_norm": 1.333854988794211, |
| "learning_rate": 4.609375e-06, |
| "loss": 0.1999, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.047225501770956316, |
| "grad_norm": 1.3581177171903593, |
| "learning_rate": 4.6875000000000004e-06, |
| "loss": 0.1957, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.048012593467138924, |
| "grad_norm": 1.3130565355707189, |
| "learning_rate": 4.765625000000001e-06, |
| "loss": 0.2122, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.048799685163321525, |
| "grad_norm": 1.4142816132337854, |
| "learning_rate": 4.84375e-06, |
| "loss": 0.2147, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.049586776859504134, |
| "grad_norm": 1.357898232243354, |
| "learning_rate": 4.921875e-06, |
| "loss": 0.1877, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.050373868555686735, |
| "grad_norm": 1.4153342064426397, |
| "learning_rate": 5e-06, |
| "loss": 0.1945, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.05116096025186934, |
| "grad_norm": 1.4495342715013748, |
| "learning_rate": 4.999991517675219e-06, |
| "loss": 0.1939, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.05194805194805195, |
| "grad_norm": 1.1539274129121713, |
| "learning_rate": 4.999966070758437e-06, |
| "loss": 0.2003, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.05273514364423455, |
| "grad_norm": 1.3379283904444008, |
| "learning_rate": 4.999923659422332e-06, |
| "loss": 0.2007, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.05352223534041716, |
| "grad_norm": 1.3492954613335875, |
| "learning_rate": 4.999864283954702e-06, |
| "loss": 0.1989, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.05430932703659976, |
| "grad_norm": 1.1801853129144864, |
| "learning_rate": 4.99978794475846e-06, |
| "loss": 0.2114, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.05509641873278237, |
| "grad_norm": 1.2068999367428581, |
| "learning_rate": 4.999694642351633e-06, |
| "loss": 0.2033, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.05588351042896497, |
| "grad_norm": 1.2287271472480104, |
| "learning_rate": 4.999584377367359e-06, |
| "loss": 0.1895, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.05667060212514758, |
| "grad_norm": 1.3129837217534652, |
| "learning_rate": 4.99945715055388e-06, |
| "loss": 0.1905, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.05745769382133018, |
| "grad_norm": 1.1734967025843308, |
| "learning_rate": 4.99931296277454e-06, |
| "loss": 0.213, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.05824478551751279, |
| "grad_norm": 1.3738466570011791, |
| "learning_rate": 4.999151815007776e-06, |
| "loss": 0.2214, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.0590318772136954, |
| "grad_norm": 1.273179655688277, |
| "learning_rate": 4.9989737083471165e-06, |
| "loss": 0.1894, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.059818968909878, |
| "grad_norm": 1.0843431120214646, |
| "learning_rate": 4.998778644001165e-06, |
| "loss": 0.1967, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.06060606060606061, |
| "grad_norm": 1.4896402431576707, |
| "learning_rate": 4.998566623293603e-06, |
| "loss": 0.1752, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.06139315230224321, |
| "grad_norm": 1.3405458603738243, |
| "learning_rate": 4.9983376476631725e-06, |
| "loss": 0.1998, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.06218024399842582, |
| "grad_norm": 1.3641086369593634, |
| "learning_rate": 4.998091718663671e-06, |
| "loss": 0.2047, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.06296733569460843, |
| "grad_norm": 1.3391162585136267, |
| "learning_rate": 4.997828837963937e-06, |
| "loss": 0.181, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.06375442739079103, |
| "grad_norm": 1.1899411991269295, |
| "learning_rate": 4.997549007347843e-06, |
| "loss": 0.1946, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.06454151908697363, |
| "grad_norm": 1.3917818646896112, |
| "learning_rate": 4.997252228714279e-06, |
| "loss": 0.1919, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.06532861078315624, |
| "grad_norm": 1.2543099071691322, |
| "learning_rate": 4.996938504077145e-06, |
| "loss": 0.1948, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.06611570247933884, |
| "grad_norm": 1.3941008619735185, |
| "learning_rate": 4.99660783556533e-06, |
| "loss": 0.1861, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.06690279417552145, |
| "grad_norm": 1.1765528133487257, |
| "learning_rate": 4.9962602254227075e-06, |
| "loss": 0.1817, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.06768988587170405, |
| "grad_norm": 1.223066746932356, |
| "learning_rate": 4.995895676008109e-06, |
| "loss": 0.1934, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.06847697756788666, |
| "grad_norm": 1.3140944559909808, |
| "learning_rate": 4.995514189795316e-06, |
| "loss": 0.197, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.06926406926406926, |
| "grad_norm": 1.1819977914205286, |
| "learning_rate": 4.99511576937304e-06, |
| "loss": 0.1972, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.07005116096025187, |
| "grad_norm": 1.3152579578345207, |
| "learning_rate": 4.994700417444907e-06, |
| "loss": 0.207, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.07083825265643448, |
| "grad_norm": 1.2064669225701854, |
| "learning_rate": 4.994268136829438e-06, |
| "loss": 0.1953, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.07162534435261708, |
| "grad_norm": 1.1619755664518439, |
| "learning_rate": 4.993818930460026e-06, |
| "loss": 0.1982, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.07241243604879968, |
| "grad_norm": 1.1792837872493809, |
| "learning_rate": 4.993352801384924e-06, |
| "loss": 0.1886, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.07319952774498228, |
| "grad_norm": 1.097328306217708, |
| "learning_rate": 4.992869752767218e-06, |
| "loss": 0.1673, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.0739866194411649, |
| "grad_norm": 1.2788239338552108, |
| "learning_rate": 4.992369787884809e-06, |
| "loss": 0.1972, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.0747737111373475, |
| "grad_norm": 1.1905278770669998, |
| "learning_rate": 4.991852910130388e-06, |
| "loss": 0.1872, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.0755608028335301, |
| "grad_norm": 1.2133270115400816, |
| "learning_rate": 4.9913191230114154e-06, |
| "loss": 0.1748, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.07634789452971272, |
| "grad_norm": 1.2840440499091732, |
| "learning_rate": 4.990768430150096e-06, |
| "loss": 0.1942, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.07713498622589532, |
| "grad_norm": 1.5346248945491554, |
| "learning_rate": 4.990200835283353e-06, |
| "loss": 0.1861, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.07792207792207792, |
| "grad_norm": 1.1936205681426777, |
| "learning_rate": 4.989616342262807e-06, |
| "loss": 0.1975, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.07870916961826052, |
| "grad_norm": 1.2662437794316659, |
| "learning_rate": 4.989014955054746e-06, |
| "loss": 0.1853, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.07949626131444314, |
| "grad_norm": 1.116915072535967, |
| "learning_rate": 4.988396677740097e-06, |
| "loss": 0.1738, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.08028335301062574, |
| "grad_norm": 1.1577366023558335, |
| "learning_rate": 4.9877615145144055e-06, |
| "loss": 0.2045, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.08107044470680834, |
| "grad_norm": 1.1022178093801993, |
| "learning_rate": 4.9871094696878e-06, |
| "loss": 0.1814, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.08185753640299095, |
| "grad_norm": 1.218327314143879, |
| "learning_rate": 4.986440547684963e-06, |
| "loss": 0.1822, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.08264462809917356, |
| "grad_norm": 1.0747362510591434, |
| "learning_rate": 4.985754753045108e-06, |
| "loss": 0.1639, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.08343171979535616, |
| "grad_norm": 1.236686976609853, |
| "learning_rate": 4.9850520904219406e-06, |
| "loss": 0.1773, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.08421881149153876, |
| "grad_norm": 1.2843110878866029, |
| "learning_rate": 4.98433256458363e-06, |
| "loss": 0.1931, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.08500590318772137, |
| "grad_norm": 1.2556201190754803, |
| "learning_rate": 4.983596180412779e-06, |
| "loss": 0.1891, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.08579299488390398, |
| "grad_norm": 1.1736861180333642, |
| "learning_rate": 4.982842942906386e-06, |
| "loss": 0.1932, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.08658008658008658, |
| "grad_norm": 1.250703274500956, |
| "learning_rate": 4.982072857175816e-06, |
| "loss": 0.1979, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.08736717827626919, |
| "grad_norm": 1.111910462348759, |
| "learning_rate": 4.981285928446762e-06, |
| "loss": 0.1729, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.0881542699724518, |
| "grad_norm": 1.2267625409230847, |
| "learning_rate": 4.980482162059214e-06, |
| "loss": 0.1993, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.0889413616686344, |
| "grad_norm": 1.350342930816002, |
| "learning_rate": 4.979661563467415e-06, |
| "loss": 0.1914, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.089728453364817, |
| "grad_norm": 0.9837790085016399, |
| "learning_rate": 4.978824138239835e-06, |
| "loss": 0.1852, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.09051554506099961, |
| "grad_norm": 1.1412715811918805, |
| "learning_rate": 4.977969892059123e-06, |
| "loss": 0.1791, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.09130263675718221, |
| "grad_norm": 1.091735318231847, |
| "learning_rate": 4.977098830722074e-06, |
| "loss": 0.1879, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.09208972845336481, |
| "grad_norm": 1.1356995797773966, |
| "learning_rate": 4.976210960139587e-06, |
| "loss": 0.1942, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.09287682014954743, |
| "grad_norm": 1.197221158258512, |
| "learning_rate": 4.975306286336628e-06, |
| "loss": 0.1822, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.09366391184573003, |
| "grad_norm": 1.1622435205009634, |
| "learning_rate": 4.974384815452187e-06, |
| "loss": 0.1938, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.09445100354191263, |
| "grad_norm": 1.1980574826361372, |
| "learning_rate": 4.9734465537392365e-06, |
| "loss": 0.1703, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.09523809523809523, |
| "grad_norm": 1.090793092501407, |
| "learning_rate": 4.972491507564688e-06, |
| "loss": 0.1681, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.09602518693427785, |
| "grad_norm": 1.2120296842604672, |
| "learning_rate": 4.9715196834093525e-06, |
| "loss": 0.1562, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.09681227863046045, |
| "grad_norm": 1.1420618659168036, |
| "learning_rate": 4.97053108786789e-06, |
| "loss": 0.1812, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.09759937032664305, |
| "grad_norm": 1.145370838994205, |
| "learning_rate": 4.969525727648774e-06, |
| "loss": 0.1873, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.09838646202282567, |
| "grad_norm": 1.1676600414602372, |
| "learning_rate": 4.9685036095742365e-06, |
| "loss": 0.1972, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.09917355371900827, |
| "grad_norm": 1.204479600477317, |
| "learning_rate": 4.967464740580228e-06, |
| "loss": 0.1904, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.09996064541519087, |
| "grad_norm": 1.119994012971968, |
| "learning_rate": 4.9664091277163664e-06, |
| "loss": 0.1851, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.10074773711137347, |
| "grad_norm": 1.1043684264734095, |
| "learning_rate": 4.9653367781458946e-06, |
| "loss": 0.1926, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.10153482880755609, |
| "grad_norm": 1.127680976136701, |
| "learning_rate": 4.964247699145626e-06, |
| "loss": 0.1886, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.10232192050373869, |
| "grad_norm": 1.1813875803533898, |
| "learning_rate": 4.963141898105898e-06, |
| "loss": 0.1858, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.10310901219992129, |
| "grad_norm": 1.1497128287458092, |
| "learning_rate": 4.962019382530521e-06, |
| "loss": 0.1724, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.1038961038961039, |
| "grad_norm": 1.223498886081565, |
| "learning_rate": 4.960880160036728e-06, |
| "loss": 0.194, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.1046831955922865, |
| "grad_norm": 1.1861652899170938, |
| "learning_rate": 4.959724238355124e-06, |
| "loss": 0.1841, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.1054702872884691, |
| "grad_norm": 1.0805114288365025, |
| "learning_rate": 4.958551625329631e-06, |
| "loss": 0.1646, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.10625737898465171, |
| "grad_norm": 1.29808710670669, |
| "learning_rate": 4.957362328917437e-06, |
| "loss": 0.1833, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.10704447068083432, |
| "grad_norm": 1.2660501691777906, |
| "learning_rate": 4.95615635718894e-06, |
| "loss": 0.1753, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.10783156237701692, |
| "grad_norm": 1.1429230314494303, |
| "learning_rate": 4.954933718327697e-06, |
| "loss": 0.1734, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.10861865407319952, |
| "grad_norm": 1.114357361335831, |
| "learning_rate": 4.953694420630361e-06, |
| "loss": 0.1925, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.10940574576938213, |
| "grad_norm": 1.1238119767186239, |
| "learning_rate": 4.952438472506636e-06, |
| "loss": 0.1805, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.11019283746556474, |
| "grad_norm": 1.1524735878912507, |
| "learning_rate": 4.951165882479206e-06, |
| "loss": 0.1783, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.11097992916174734, |
| "grad_norm": 1.0546198047284017, |
| "learning_rate": 4.949876659183693e-06, |
| "loss": 0.1745, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.11176702085792994, |
| "grad_norm": 1.0925714956018635, |
| "learning_rate": 4.94857081136858e-06, |
| "loss": 0.1763, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.11255411255411256, |
| "grad_norm": 1.1039385653204372, |
| "learning_rate": 4.947248347895172e-06, |
| "loss": 0.1777, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.11334120425029516, |
| "grad_norm": 1.145622347104172, |
| "learning_rate": 4.945909277737519e-06, |
| "loss": 0.1804, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.11412829594647776, |
| "grad_norm": 1.0810330697861197, |
| "learning_rate": 4.944553609982363e-06, |
| "loss": 0.18, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.11491538764266036, |
| "grad_norm": 1.079722871077113, |
| "learning_rate": 4.943181353829077e-06, |
| "loss": 0.1805, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.11570247933884298, |
| "grad_norm": 1.2122723148500483, |
| "learning_rate": 4.941792518589596e-06, |
| "loss": 0.2113, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.11648957103502558, |
| "grad_norm": 1.1619622709214918, |
| "learning_rate": 4.940387113688364e-06, |
| "loss": 0.1714, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.11727666273120818, |
| "grad_norm": 1.0508760593348456, |
| "learning_rate": 4.93896514866226e-06, |
| "loss": 0.1625, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.1180637544273908, |
| "grad_norm": 1.0710088382142664, |
| "learning_rate": 4.93752663316054e-06, |
| "loss": 0.1778, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.1188508461235734, |
| "grad_norm": 1.0503531295721205, |
| "learning_rate": 4.936071576944769e-06, |
| "loss": 0.1726, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.119637937819756, |
| "grad_norm": 1.0686610020146463, |
| "learning_rate": 4.934599989888753e-06, |
| "loss": 0.1769, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.1204250295159386, |
| "grad_norm": 1.072378297090023, |
| "learning_rate": 4.933111881978478e-06, |
| "loss": 0.1866, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.12121212121212122, |
| "grad_norm": 1.2495883030259693, |
| "learning_rate": 4.931607263312033e-06, |
| "loss": 0.1998, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.12199921290830382, |
| "grad_norm": 1.109893027407933, |
| "learning_rate": 4.93008614409955e-06, |
| "loss": 0.1805, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.12278630460448642, |
| "grad_norm": 1.1570851370725408, |
| "learning_rate": 4.928548534663133e-06, |
| "loss": 0.1725, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.12357339630066903, |
| "grad_norm": 1.1758781032456742, |
| "learning_rate": 4.9269944454367815e-06, |
| "loss": 0.176, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.12436048799685163, |
| "grad_norm": 1.1408455648753233, |
| "learning_rate": 4.925423886966328e-06, |
| "loss": 0.1848, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.12514757969303425, |
| "grad_norm": 1.1318514267380126, |
| "learning_rate": 4.923836869909363e-06, |
| "loss": 0.1764, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.12593467138921685, |
| "grad_norm": 1.1451300788977063, |
| "learning_rate": 4.9222334050351595e-06, |
| "loss": 0.1756, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.12672176308539945, |
| "grad_norm": 1.1117305593028235, |
| "learning_rate": 4.920613503224608e-06, |
| "loss": 0.1797, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.12750885478158205, |
| "grad_norm": 1.1301581138966732, |
| "learning_rate": 4.9189771754701335e-06, |
| "loss": 0.1675, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.12829594647776466, |
| "grad_norm": 1.0326917294149387, |
| "learning_rate": 4.917324432875627e-06, |
| "loss": 0.1784, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.12908303817394726, |
| "grad_norm": 1.1983588521884831, |
| "learning_rate": 4.915655286656368e-06, |
| "loss": 0.1966, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.12987012987012986, |
| "grad_norm": 1.0140424703790007, |
| "learning_rate": 4.9139697481389505e-06, |
| "loss": 0.1744, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.1306572215663125, |
| "grad_norm": 1.223539092779737, |
| "learning_rate": 4.9122678287612e-06, |
| "loss": 0.1831, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.1314443132624951, |
| "grad_norm": 1.0918972348910556, |
| "learning_rate": 4.910549540072104e-06, |
| "loss": 0.1843, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.1322314049586777, |
| "grad_norm": 1.1292739304249166, |
| "learning_rate": 4.908814893731728e-06, |
| "loss": 0.1552, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.1330184966548603, |
| "grad_norm": 1.1923518362383727, |
| "learning_rate": 4.9070639015111406e-06, |
| "loss": 0.1895, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.1338055883510429, |
| "grad_norm": 1.083542335588892, |
| "learning_rate": 4.905296575292329e-06, |
| "loss": 0.1745, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.1345926800472255, |
| "grad_norm": 1.2673623015109376, |
| "learning_rate": 4.90351292706812e-06, |
| "loss": 0.1726, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.1353797717434081, |
| "grad_norm": 1.1129476624507257, |
| "learning_rate": 4.901712968942101e-06, |
| "loss": 0.1706, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.13616686343959072, |
| "grad_norm": 1.1735922432656085, |
| "learning_rate": 4.899896713128536e-06, |
| "loss": 0.1741, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.13695395513577333, |
| "grad_norm": 1.2331570034422519, |
| "learning_rate": 4.898064171952281e-06, |
| "loss": 0.1946, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.13774104683195593, |
| "grad_norm": 1.2376618802061816, |
| "learning_rate": 4.896215357848706e-06, |
| "loss": 0.1715, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.13852813852813853, |
| "grad_norm": 1.0860947256302276, |
| "learning_rate": 4.894350283363603e-06, |
| "loss": 0.1664, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.13931523022432113, |
| "grad_norm": 1.1284792933006988, |
| "learning_rate": 4.892468961153105e-06, |
| "loss": 0.1721, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.14010232192050373, |
| "grad_norm": 1.1811695933066144, |
| "learning_rate": 4.8905714039836026e-06, |
| "loss": 0.1768, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.14088941361668633, |
| "grad_norm": 1.1690172197627666, |
| "learning_rate": 4.888657624731652e-06, |
| "loss": 0.1784, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.14167650531286896, |
| "grad_norm": 1.2215187765329307, |
| "learning_rate": 4.88672763638389e-06, |
| "loss": 0.1762, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.14246359700905156, |
| "grad_norm": 1.1657625904368065, |
| "learning_rate": 4.884781452036948e-06, |
| "loss": 0.1754, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.14325068870523416, |
| "grad_norm": 1.0812019740421663, |
| "learning_rate": 4.88281908489736e-06, |
| "loss": 0.1745, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.14403778040141677, |
| "grad_norm": 1.1662972444477193, |
| "learning_rate": 4.880840548281475e-06, |
| "loss": 0.1844, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.14482487209759937, |
| "grad_norm": 1.1318261660435303, |
| "learning_rate": 4.878845855615364e-06, |
| "loss": 0.177, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.14561196379378197, |
| "grad_norm": 1.0454173174935852, |
| "learning_rate": 4.876835020434733e-06, |
| "loss": 0.1726, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.14639905548996457, |
| "grad_norm": 1.1649728572384528, |
| "learning_rate": 4.874808056384826e-06, |
| "loss": 0.1829, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.1471861471861472, |
| "grad_norm": 0.9809711097737751, |
| "learning_rate": 4.8727649772203375e-06, |
| "loss": 0.1626, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.1479732388823298, |
| "grad_norm": 1.0024677570018588, |
| "learning_rate": 4.8707057968053175e-06, |
| "loss": 0.1564, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.1487603305785124, |
| "grad_norm": 1.0801740218719516, |
| "learning_rate": 4.868630529113075e-06, |
| "loss": 0.1571, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.149547422274695, |
| "grad_norm": 1.0633734918657578, |
| "learning_rate": 4.866539188226086e-06, |
| "loss": 0.1558, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.1503345139708776, |
| "grad_norm": 1.1110942685300096, |
| "learning_rate": 4.864431788335895e-06, |
| "loss": 0.1739, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.1511216056670602, |
| "grad_norm": 1.088865739839623, |
| "learning_rate": 4.862308343743024e-06, |
| "loss": 0.1705, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.1519086973632428, |
| "grad_norm": 1.158763785538179, |
| "learning_rate": 4.86016886885687e-06, |
| "loss": 0.1754, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.15269578905942544, |
| "grad_norm": 1.0665033787081621, |
| "learning_rate": 4.858013378195609e-06, |
| "loss": 0.1814, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.15348288075560804, |
| "grad_norm": 1.0347506383595513, |
| "learning_rate": 4.855841886386099e-06, |
| "loss": 0.1659, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.15426997245179064, |
| "grad_norm": 1.3652087096932124, |
| "learning_rate": 4.8536544081637785e-06, |
| "loss": 0.1693, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.15505706414797324, |
| "grad_norm": 1.1571980267809596, |
| "learning_rate": 4.8514509583725685e-06, |
| "loss": 0.1735, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.15584415584415584, |
| "grad_norm": 1.1126611625924816, |
| "learning_rate": 4.849231551964771e-06, |
| "loss": 0.1878, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.15663124754033844, |
| "grad_norm": 1.0666827506948415, |
| "learning_rate": 4.846996204000967e-06, |
| "loss": 0.1686, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.15741833923652104, |
| "grad_norm": 1.1408187983192677, |
| "learning_rate": 4.844744929649912e-06, |
| "loss": 0.1785, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.15820543093270367, |
| "grad_norm": 1.1050850982745672, |
| "learning_rate": 4.842477744188441e-06, |
| "loss": 0.1663, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.15899252262888627, |
| "grad_norm": 1.0153624350350885, |
| "learning_rate": 4.840194663001354e-06, |
| "loss": 0.1755, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.15977961432506887, |
| "grad_norm": 1.0251264155888737, |
| "learning_rate": 4.837895701581322e-06, |
| "loss": 0.1537, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.16056670602125148, |
| "grad_norm": 1.0673153393456505, |
| "learning_rate": 4.835580875528776e-06, |
| "loss": 0.1633, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.16135379771743408, |
| "grad_norm": 1.0273828987011315, |
| "learning_rate": 4.833250200551798e-06, |
| "loss": 0.1746, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.16214088941361668, |
| "grad_norm": 1.0964068866663357, |
| "learning_rate": 4.830903692466023e-06, |
| "loss": 0.1674, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.16292798110979928, |
| "grad_norm": 1.1142080493277295, |
| "learning_rate": 4.828541367194527e-06, |
| "loss": 0.1828, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.1637150728059819, |
| "grad_norm": 1.0617790409690397, |
| "learning_rate": 4.826163240767717e-06, |
| "loss": 0.1676, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.1645021645021645, |
| "grad_norm": 1.2859855971245049, |
| "learning_rate": 4.8237693293232256e-06, |
| "loss": 0.1942, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.1652892561983471, |
| "grad_norm": 1.000840540957111, |
| "learning_rate": 4.821359649105801e-06, |
| "loss": 0.1686, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.1660763478945297, |
| "grad_norm": 1.049595380158752, |
| "learning_rate": 4.818934216467195e-06, |
| "loss": 0.1696, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.16686343959071231, |
| "grad_norm": 1.0218031530162965, |
| "learning_rate": 4.816493047866053e-06, |
| "loss": 0.1653, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.16765053128689492, |
| "grad_norm": 1.0715206508098112, |
| "learning_rate": 4.8140361598678034e-06, |
| "loss": 0.1735, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.16843762298307752, |
| "grad_norm": 1.093161202120212, |
| "learning_rate": 4.811563569144544e-06, |
| "loss": 0.1698, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.16922471467926015, |
| "grad_norm": 1.078958887147992, |
| "learning_rate": 4.809075292474929e-06, |
| "loss": 0.1671, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.17001180637544275, |
| "grad_norm": 1.1213364259804648, |
| "learning_rate": 4.806571346744053e-06, |
| "loss": 0.1798, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.17079889807162535, |
| "grad_norm": 1.102076724202232, |
| "learning_rate": 4.804051748943343e-06, |
| "loss": 0.1845, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.17158598976780795, |
| "grad_norm": 1.1103430873095865, |
| "learning_rate": 4.801516516170437e-06, |
| "loss": 0.177, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.17237308146399055, |
| "grad_norm": 1.228711789290585, |
| "learning_rate": 4.798965665629068e-06, |
| "loss": 0.1636, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.17316017316017315, |
| "grad_norm": 1.1219855198900837, |
| "learning_rate": 4.796399214628949e-06, |
| "loss": 0.1802, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.17394726485635575, |
| "grad_norm": 1.1846418832749555, |
| "learning_rate": 4.7938171805856596e-06, |
| "loss": 0.1717, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.17473435655253838, |
| "grad_norm": 1.0672386815907553, |
| "learning_rate": 4.791219581020518e-06, |
| "loss": 0.1663, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.17552144824872098, |
| "grad_norm": 1.0398388591323704, |
| "learning_rate": 4.788606433560473e-06, |
| "loss": 0.1593, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.1763085399449036, |
| "grad_norm": 1.1402534682960337, |
| "learning_rate": 4.785977755937977e-06, |
| "loss": 0.1876, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.1770956316410862, |
| "grad_norm": 1.1260603683997887, |
| "learning_rate": 4.783333565990865e-06, |
| "loss": 0.172, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.1778827233372688, |
| "grad_norm": 1.062290554096683, |
| "learning_rate": 4.780673881662242e-06, |
| "loss": 0.1709, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.1786698150334514, |
| "grad_norm": 1.0650729387286197, |
| "learning_rate": 4.777998721000353e-06, |
| "loss": 0.1614, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.179456906729634, |
| "grad_norm": 1.0365419204779498, |
| "learning_rate": 4.775308102158461e-06, |
| "loss": 0.1605, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.18024399842581662, |
| "grad_norm": 1.1444494636007958, |
| "learning_rate": 4.772602043394731e-06, |
| "loss": 0.1867, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.18103109012199922, |
| "grad_norm": 1.1053808430839196, |
| "learning_rate": 4.769880563072097e-06, |
| "loss": 0.1627, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.18181818181818182, |
| "grad_norm": 1.0763207393373317, |
| "learning_rate": 4.767143679658143e-06, |
| "loss": 0.1703, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.18260527351436442, |
| "grad_norm": 1.1302336936081483, |
| "learning_rate": 4.764391411724977e-06, |
| "loss": 0.1697, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.18339236521054703, |
| "grad_norm": 1.059980991296742, |
| "learning_rate": 4.7616237779491026e-06, |
| "loss": 0.1658, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.18417945690672963, |
| "grad_norm": 1.0952807461742509, |
| "learning_rate": 4.758840797111295e-06, |
| "loss": 0.1833, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.18496654860291223, |
| "grad_norm": 1.0263555674269131, |
| "learning_rate": 4.756042488096472e-06, |
| "loss": 0.1732, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.18575364029909486, |
| "grad_norm": 1.088261327233659, |
| "learning_rate": 4.753228869893566e-06, |
| "loss": 0.1646, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.18654073199527746, |
| "grad_norm": 1.0644325115229099, |
| "learning_rate": 4.750399961595395e-06, |
| "loss": 0.1576, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.18732782369146006, |
| "grad_norm": 0.9952967090049917, |
| "learning_rate": 4.747555782398537e-06, |
| "loss": 0.1598, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.18811491538764266, |
| "grad_norm": 1.0300249714418026, |
| "learning_rate": 4.7446963516031904e-06, |
| "loss": 0.1883, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.18890200708382526, |
| "grad_norm": 1.0275382678304879, |
| "learning_rate": 4.741821688613054e-06, |
| "loss": 0.1704, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.18968909878000786, |
| "grad_norm": 1.0616733952682182, |
| "learning_rate": 4.738931812935186e-06, |
| "loss": 0.1907, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.19047619047619047, |
| "grad_norm": 1.0103628221724312, |
| "learning_rate": 4.736026744179878e-06, |
| "loss": 0.1556, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.1912632821723731, |
| "grad_norm": 1.0535669337117792, |
| "learning_rate": 4.73310650206052e-06, |
| "loss": 0.1809, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.1920503738685557, |
| "grad_norm": 1.0554553563643476, |
| "learning_rate": 4.730171106393466e-06, |
| "loss": 0.1675, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.1928374655647383, |
| "grad_norm": 0.9417424551436594, |
| "learning_rate": 4.7272205770979e-06, |
| "loss": 0.1438, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.1936245572609209, |
| "grad_norm": 1.1154888244817747, |
| "learning_rate": 4.724254934195698e-06, |
| "loss": 0.1765, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.1944116489571035, |
| "grad_norm": 1.1742188581521773, |
| "learning_rate": 4.721274197811298e-06, |
| "loss": 0.1711, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.1951987406532861, |
| "grad_norm": 1.057640390538921, |
| "learning_rate": 4.71827838817156e-06, |
| "loss": 0.1678, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.1959858323494687, |
| "grad_norm": 1.022336905613029, |
| "learning_rate": 4.715267525605627e-06, |
| "loss": 0.1552, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.19677292404565133, |
| "grad_norm": 1.181830506383501, |
| "learning_rate": 4.712241630544792e-06, |
| "loss": 0.1765, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.19756001574183393, |
| "grad_norm": 1.1571296526874602, |
| "learning_rate": 4.709200723522353e-06, |
| "loss": 0.1758, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.19834710743801653, |
| "grad_norm": 1.082056647389628, |
| "learning_rate": 4.706144825173481e-06, |
| "loss": 0.1638, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.19913419913419914, |
| "grad_norm": 1.0648327864294944, |
| "learning_rate": 4.703073956235071e-06, |
| "loss": 0.1747, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.19992129083038174, |
| "grad_norm": 1.1273460773870558, |
| "learning_rate": 4.6999881375456116e-06, |
| "loss": 0.1767, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.20070838252656434, |
| "grad_norm": 1.0782376126285664, |
| "learning_rate": 4.696887390045035e-06, |
| "loss": 0.169, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.20149547422274694, |
| "grad_norm": 1.043398805036875, |
| "learning_rate": 4.693771734774578e-06, |
| "loss": 0.1774, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.20228256591892957, |
| "grad_norm": 1.067320862475683, |
| "learning_rate": 4.690641192876643e-06, |
| "loss": 0.1607, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.20306965761511217, |
| "grad_norm": 1.1843944163744937, |
| "learning_rate": 4.687495785594646e-06, |
| "loss": 0.1633, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.20385674931129477, |
| "grad_norm": 1.0931562611646284, |
| "learning_rate": 4.684335534272881e-06, |
| "loss": 0.1687, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.20464384100747737, |
| "grad_norm": 1.1204870400497637, |
| "learning_rate": 4.68116046035637e-06, |
| "loss": 0.1639, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.20543093270365997, |
| "grad_norm": 1.2082791443480092, |
| "learning_rate": 4.6779705853907205e-06, |
| "loss": 0.1683, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.20621802439984258, |
| "grad_norm": 1.0646518318192153, |
| "learning_rate": 4.674765931021976e-06, |
| "loss": 0.1611, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.20700511609602518, |
| "grad_norm": 1.1268791395123645, |
| "learning_rate": 4.671546518996473e-06, |
| "loss": 0.1553, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.2077922077922078, |
| "grad_norm": 1.0048534045343525, |
| "learning_rate": 4.668312371160688e-06, |
| "loss": 0.1571, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.2085792994883904, |
| "grad_norm": 1.0052893495164037, |
| "learning_rate": 4.665063509461098e-06, |
| "loss": 0.1679, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.209366391184573, |
| "grad_norm": 0.9679422598052939, |
| "learning_rate": 4.661799955944019e-06, |
| "loss": 0.1556, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.2101534828807556, |
| "grad_norm": 1.0487292157874373, |
| "learning_rate": 4.658521732755471e-06, |
| "loss": 0.183, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.2109405745769382, |
| "grad_norm": 1.0878511570789495, |
| "learning_rate": 4.655228862141017e-06, |
| "loss": 0.1762, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.2117276662731208, |
| "grad_norm": 0.9275216638767947, |
| "learning_rate": 4.651921366445613e-06, |
| "loss": 0.1483, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.21251475796930341, |
| "grad_norm": 1.0291173856009612, |
| "learning_rate": 4.648599268113464e-06, |
| "loss": 0.1657, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.21330184966548604, |
| "grad_norm": 0.9814951923963836, |
| "learning_rate": 4.645262589687861e-06, |
| "loss": 0.1737, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.21408894136166864, |
| "grad_norm": 0.9574503772544043, |
| "learning_rate": 4.641911353811038e-06, |
| "loss": 0.1638, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.21487603305785125, |
| "grad_norm": 0.9684496500051328, |
| "learning_rate": 4.638545583224011e-06, |
| "loss": 0.1649, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.21566312475403385, |
| "grad_norm": 1.0314787067828541, |
| "learning_rate": 4.635165300766428e-06, |
| "loss": 0.1699, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.21645021645021645, |
| "grad_norm": 1.0287264097080684, |
| "learning_rate": 4.63177052937641e-06, |
| "loss": 0.1602, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.21723730814639905, |
| "grad_norm": 1.1114659065296888, |
| "learning_rate": 4.628361292090403e-06, |
| "loss": 0.1783, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.21802439984258165, |
| "grad_norm": 1.0298788844790752, |
| "learning_rate": 4.6249376120430115e-06, |
| "loss": 0.1678, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.21881149153876425, |
| "grad_norm": 1.0099420287081406, |
| "learning_rate": 4.621499512466847e-06, |
| "loss": 0.1672, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.21959858323494688, |
| "grad_norm": 0.9892117727941296, |
| "learning_rate": 4.618047016692374e-06, |
| "loss": 0.1663, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.22038567493112948, |
| "grad_norm": 0.9289360238552057, |
| "learning_rate": 4.614580148147744e-06, |
| "loss": 0.1563, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.22117276662731208, |
| "grad_norm": 0.9603340451855991, |
| "learning_rate": 4.61109893035864e-06, |
| "loss": 0.1561, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.22195985832349469, |
| "grad_norm": 1.0449269347565262, |
| "learning_rate": 4.607603386948119e-06, |
| "loss": 0.165, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.2227469500196773, |
| "grad_norm": 0.990226128298578, |
| "learning_rate": 4.604093541636448e-06, |
| "loss": 0.1704, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.2235340417158599, |
| "grad_norm": 1.031797952555019, |
| "learning_rate": 4.600569418240946e-06, |
| "loss": 0.1677, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.2243211334120425, |
| "grad_norm": 1.0506428763431659, |
| "learning_rate": 4.597031040675819e-06, |
| "loss": 0.1802, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.22510822510822512, |
| "grad_norm": 0.980146123693525, |
| "learning_rate": 4.593478432952002e-06, |
| "loss": 0.1656, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.22589531680440772, |
| "grad_norm": 1.0058178922055618, |
| "learning_rate": 4.589911619176993e-06, |
| "loss": 0.1601, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.22668240850059032, |
| "grad_norm": 1.1532752501338874, |
| "learning_rate": 4.586330623554691e-06, |
| "loss": 0.1707, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.22746950019677292, |
| "grad_norm": 0.9925104519486038, |
| "learning_rate": 4.582735470385229e-06, |
| "loss": 0.1712, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.22825659189295552, |
| "grad_norm": 1.1312813134045174, |
| "learning_rate": 4.579126184064814e-06, |
| "loss": 0.1607, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.22904368358913813, |
| "grad_norm": 1.2454875330122912, |
| "learning_rate": 4.575502789085555e-06, |
| "loss": 0.1656, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.22983077528532073, |
| "grad_norm": 0.9825183210915687, |
| "learning_rate": 4.571865310035304e-06, |
| "loss": 0.1589, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.23061786698150336, |
| "grad_norm": 1.0887371255437703, |
| "learning_rate": 4.568213771597484e-06, |
| "loss": 0.1585, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.23140495867768596, |
| "grad_norm": 1.0975434488519114, |
| "learning_rate": 4.564548198550922e-06, |
| "loss": 0.1435, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.23219205037386856, |
| "grad_norm": 1.0593259383463134, |
| "learning_rate": 4.5608686157696844e-06, |
| "loss": 0.167, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.23297914207005116, |
| "grad_norm": 1.1536948102561841, |
| "learning_rate": 4.557175048222901e-06, |
| "loss": 0.1621, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.23376623376623376, |
| "grad_norm": 1.1369019291567328, |
| "learning_rate": 4.5534675209746076e-06, |
| "loss": 0.1654, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.23455332546241636, |
| "grad_norm": 0.9585590140764199, |
| "learning_rate": 4.5497460591835615e-06, |
| "loss": 0.148, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.23534041715859896, |
| "grad_norm": 1.2337420030262027, |
| "learning_rate": 4.546010688103082e-06, |
| "loss": 0.1599, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.2361275088547816, |
| "grad_norm": 1.1641848426244756, |
| "learning_rate": 4.542261433080874e-06, |
| "loss": 0.1641, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.2369146005509642, |
| "grad_norm": 0.9715264597638171, |
| "learning_rate": 4.538498319558854e-06, |
| "loss": 0.1604, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.2377016922471468, |
| "grad_norm": 1.2043568904283137, |
| "learning_rate": 4.534721373072986e-06, |
| "loss": 0.1561, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.2384887839433294, |
| "grad_norm": 1.087701432883666, |
| "learning_rate": 4.530930619253097e-06, |
| "loss": 0.1573, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.239275875639512, |
| "grad_norm": 1.0432095830081018, |
| "learning_rate": 4.527126083822713e-06, |
| "loss": 0.1576, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.2400629673356946, |
| "grad_norm": 1.1515388977241858, |
| "learning_rate": 4.523307792598877e-06, |
| "loss": 0.1836, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.2408500590318772, |
| "grad_norm": 1.1236907370811289, |
| "learning_rate": 4.519475771491978e-06, |
| "loss": 0.1654, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.24163715072805983, |
| "grad_norm": 1.0492490872684002, |
| "learning_rate": 4.515630046505575e-06, |
| "loss": 0.1604, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.24242424242424243, |
| "grad_norm": 1.0414505694174347, |
| "learning_rate": 4.511770643736217e-06, |
| "loss": 0.1587, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.24321133412042503, |
| "grad_norm": 0.9963463131455829, |
| "learning_rate": 4.507897589373272e-06, |
| "loss": 0.1536, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.24399842581660763, |
| "grad_norm": 0.9437267739253786, |
| "learning_rate": 4.504010909698744e-06, |
| "loss": 0.1573, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.24478551751279023, |
| "grad_norm": 0.9915304289222059, |
| "learning_rate": 4.500110631087095e-06, |
| "loss": 0.1519, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.24557260920897284, |
| "grad_norm": 0.9782358310573961, |
| "learning_rate": 4.496196780005069e-06, |
| "loss": 0.1629, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.24635970090515544, |
| "grad_norm": 1.0770165377269398, |
| "learning_rate": 4.492269383011512e-06, |
| "loss": 0.1623, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.24714679260133807, |
| "grad_norm": 1.052396599909024, |
| "learning_rate": 4.4883284667571894e-06, |
| "loss": 0.1533, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.24793388429752067, |
| "grad_norm": 1.0084809840218907, |
| "learning_rate": 4.4843740579846055e-06, |
| "loss": 0.1512, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.24872097599370327, |
| "grad_norm": 1.0756395659672484, |
| "learning_rate": 4.480406183527823e-06, |
| "loss": 0.1682, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.24950806768988587, |
| "grad_norm": 1.095604151904482, |
| "learning_rate": 4.476424870312286e-06, |
| "loss": 0.1588, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.2502951593860685, |
| "grad_norm": 1.073871876794014, |
| "learning_rate": 4.472430145354622e-06, |
| "loss": 0.1663, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.2510822510822511, |
| "grad_norm": 1.00181438336178, |
| "learning_rate": 4.46842203576248e-06, |
| "loss": 0.1668, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.2518693427784337, |
| "grad_norm": 1.0179064844212398, |
| "learning_rate": 4.464400568734327e-06, |
| "loss": 0.1618, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.2526564344746163, |
| "grad_norm": 1.1266566093245078, |
| "learning_rate": 4.460365771559275e-06, |
| "loss": 0.1726, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.2534435261707989, |
| "grad_norm": 1.0831980755033608, |
| "learning_rate": 4.456317671616892e-06, |
| "loss": 0.1674, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.2542306178669815, |
| "grad_norm": 0.9991360442603613, |
| "learning_rate": 4.452256296377017e-06, |
| "loss": 0.1534, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.2550177095631641, |
| "grad_norm": 0.9497710360440503, |
| "learning_rate": 4.448181673399573e-06, |
| "loss": 0.1562, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.25580480125934674, |
| "grad_norm": 1.1113260986403124, |
| "learning_rate": 4.444093830334381e-06, |
| "loss": 0.1639, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.2565918929555293, |
| "grad_norm": 1.1452949830587935, |
| "learning_rate": 4.4399927949209685e-06, |
| "loss": 0.1633, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.25737898465171194, |
| "grad_norm": 1.0842379105419755, |
| "learning_rate": 4.43587859498839e-06, |
| "loss": 0.1754, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.2581660763478945, |
| "grad_norm": 1.0361570331888057, |
| "learning_rate": 4.431751258455029e-06, |
| "loss": 0.1629, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.25895316804407714, |
| "grad_norm": 0.9514704452172565, |
| "learning_rate": 4.4276108133284115e-06, |
| "loss": 0.1615, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.2597402597402597, |
| "grad_norm": 1.0051943736689641, |
| "learning_rate": 4.4234572877050175e-06, |
| "loss": 0.1635, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.26052735143644234, |
| "grad_norm": 1.061826511574687, |
| "learning_rate": 4.419290709770091e-06, |
| "loss": 0.1572, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.261314443132625, |
| "grad_norm": 1.0098180333606226, |
| "learning_rate": 4.415111107797445e-06, |
| "loss": 0.1625, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.26210153482880755, |
| "grad_norm": 0.9258158779374888, |
| "learning_rate": 4.4109185101492735e-06, |
| "loss": 0.163, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.2628886265249902, |
| "grad_norm": 1.031959410480149, |
| "learning_rate": 4.406712945275955e-06, |
| "loss": 0.1601, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.26367571822117275, |
| "grad_norm": 1.098174422684468, |
| "learning_rate": 4.402494441715864e-06, |
| "loss": 0.1632, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.2644628099173554, |
| "grad_norm": 0.9325275936138202, |
| "learning_rate": 4.398263028095175e-06, |
| "loss": 0.1568, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.26524990161353795, |
| "grad_norm": 0.9452361980478395, |
| "learning_rate": 4.394018733127667e-06, |
| "loss": 0.1514, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.2660369933097206, |
| "grad_norm": 0.9440560796701104, |
| "learning_rate": 4.389761585614531e-06, |
| "loss": 0.1568, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.2668240850059032, |
| "grad_norm": 0.9825093172685871, |
| "learning_rate": 4.3854916144441714e-06, |
| "loss": 0.1513, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.2676111767020858, |
| "grad_norm": 0.9909422001877334, |
| "learning_rate": 4.381208848592017e-06, |
| "loss": 0.1607, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.2683982683982684, |
| "grad_norm": 1.026772957857381, |
| "learning_rate": 4.3769133171203146e-06, |
| "loss": 0.1579, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.269185360094451, |
| "grad_norm": 0.9727634660522837, |
| "learning_rate": 4.372605049177939e-06, |
| "loss": 0.1611, |
| "step": 342 |
| }, |
| { |
| "epoch": 0.2699724517906336, |
| "grad_norm": 0.9991705382361779, |
| "learning_rate": 4.368284074000193e-06, |
| "loss": 0.1423, |
| "step": 343 |
| }, |
| { |
| "epoch": 0.2707595434868162, |
| "grad_norm": 1.0413166825567135, |
| "learning_rate": 4.363950420908608e-06, |
| "loss": 0.1531, |
| "step": 344 |
| }, |
| { |
| "epoch": 0.2715466351829988, |
| "grad_norm": 1.051399331371367, |
| "learning_rate": 4.3596041193107475e-06, |
| "loss": 0.1537, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.27233372687918145, |
| "grad_norm": 1.1268002118202416, |
| "learning_rate": 4.355245198700003e-06, |
| "loss": 0.1687, |
| "step": 346 |
| }, |
| { |
| "epoch": 0.273120818575364, |
| "grad_norm": 1.0579162910588005, |
| "learning_rate": 4.3508736886554e-06, |
| "loss": 0.1545, |
| "step": 347 |
| }, |
| { |
| "epoch": 0.27390791027154665, |
| "grad_norm": 1.0780531804812832, |
| "learning_rate": 4.346489618841393e-06, |
| "loss": 0.1478, |
| "step": 348 |
| }, |
| { |
| "epoch": 0.2746950019677292, |
| "grad_norm": 1.1629336261073622, |
| "learning_rate": 4.342093019007664e-06, |
| "loss": 0.1507, |
| "step": 349 |
| }, |
| { |
| "epoch": 0.27548209366391185, |
| "grad_norm": 0.9806357134318359, |
| "learning_rate": 4.337683918988924e-06, |
| "loss": 0.1605, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.2762691853600944, |
| "grad_norm": 1.0271547256327147, |
| "learning_rate": 4.333262348704708e-06, |
| "loss": 0.1544, |
| "step": 351 |
| }, |
| { |
| "epoch": 0.27705627705627706, |
| "grad_norm": 1.040963108089893, |
| "learning_rate": 4.328828338159173e-06, |
| "loss": 0.1505, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.2778433687524597, |
| "grad_norm": 1.036202462349552, |
| "learning_rate": 4.324381917440891e-06, |
| "loss": 0.1558, |
| "step": 353 |
| }, |
| { |
| "epoch": 0.27863046044864226, |
| "grad_norm": 0.975994343559266, |
| "learning_rate": 4.319923116722651e-06, |
| "loss": 0.1641, |
| "step": 354 |
| }, |
| { |
| "epoch": 0.2794175521448249, |
| "grad_norm": 1.039409188253541, |
| "learning_rate": 4.315451966261248e-06, |
| "loss": 0.1549, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.28020464384100746, |
| "grad_norm": 1.047725080130562, |
| "learning_rate": 4.310968496397284e-06, |
| "loss": 0.165, |
| "step": 356 |
| }, |
| { |
| "epoch": 0.2809917355371901, |
| "grad_norm": 1.0011313336241248, |
| "learning_rate": 4.306472737554957e-06, |
| "loss": 0.1456, |
| "step": 357 |
| }, |
| { |
| "epoch": 0.28177882723337266, |
| "grad_norm": 0.9015679935576075, |
| "learning_rate": 4.301964720241857e-06, |
| "loss": 0.1369, |
| "step": 358 |
| }, |
| { |
| "epoch": 0.2825659189295553, |
| "grad_norm": 1.049381444767021, |
| "learning_rate": 4.297444475048755e-06, |
| "loss": 0.1563, |
| "step": 359 |
| }, |
| { |
| "epoch": 0.2833530106257379, |
| "grad_norm": 1.0194195709152667, |
| "learning_rate": 4.292912032649403e-06, |
| "loss": 0.1649, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.2841401023219205, |
| "grad_norm": 0.957368492301693, |
| "learning_rate": 4.2883674238003195e-06, |
| "loss": 0.1515, |
| "step": 361 |
| }, |
| { |
| "epoch": 0.2849271940181031, |
| "grad_norm": 1.1143901057936236, |
| "learning_rate": 4.2838106793405825e-06, |
| "loss": 0.1625, |
| "step": 362 |
| }, |
| { |
| "epoch": 0.2857142857142857, |
| "grad_norm": 1.0882799366794436, |
| "learning_rate": 4.2792418301916225e-06, |
| "loss": 0.153, |
| "step": 363 |
| }, |
| { |
| "epoch": 0.2865013774104683, |
| "grad_norm": 1.0192177035415801, |
| "learning_rate": 4.274660907357009e-06, |
| "loss": 0.1645, |
| "step": 364 |
| }, |
| { |
| "epoch": 0.2872884691066509, |
| "grad_norm": 1.0205240256871457, |
| "learning_rate": 4.2700679419222415e-06, |
| "loss": 0.1459, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.28807556080283353, |
| "grad_norm": 1.2141057030738465, |
| "learning_rate": 4.265462965054539e-06, |
| "loss": 0.1597, |
| "step": 366 |
| }, |
| { |
| "epoch": 0.28886265249901616, |
| "grad_norm": 1.017121088131926, |
| "learning_rate": 4.260846008002631e-06, |
| "loss": 0.1619, |
| "step": 367 |
| }, |
| { |
| "epoch": 0.28964974419519873, |
| "grad_norm": 1.0877328731947116, |
| "learning_rate": 4.25621710209654e-06, |
| "loss": 0.1716, |
| "step": 368 |
| }, |
| { |
| "epoch": 0.29043683589138136, |
| "grad_norm": 1.1099554764936985, |
| "learning_rate": 4.251576278747372e-06, |
| "loss": 0.1599, |
| "step": 369 |
| }, |
| { |
| "epoch": 0.29122392758756394, |
| "grad_norm": 0.962048395782395, |
| "learning_rate": 4.246923569447105e-06, |
| "loss": 0.1465, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.29201101928374656, |
| "grad_norm": 1.0782102946203345, |
| "learning_rate": 4.24225900576837e-06, |
| "loss": 0.1584, |
| "step": 371 |
| }, |
| { |
| "epoch": 0.29279811097992914, |
| "grad_norm": 1.0600722154446367, |
| "learning_rate": 4.237582619364244e-06, |
| "loss": 0.1518, |
| "step": 372 |
| }, |
| { |
| "epoch": 0.29358520267611177, |
| "grad_norm": 1.0154082912245785, |
| "learning_rate": 4.23289444196803e-06, |
| "loss": 0.1455, |
| "step": 373 |
| }, |
| { |
| "epoch": 0.2943722943722944, |
| "grad_norm": 1.1254176051245297, |
| "learning_rate": 4.228194505393041e-06, |
| "loss": 0.1544, |
| "step": 374 |
| }, |
| { |
| "epoch": 0.29515938606847697, |
| "grad_norm": 1.1003313998342341, |
| "learning_rate": 4.22348284153239e-06, |
| "loss": 0.1611, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.2959464777646596, |
| "grad_norm": 0.9110264218620379, |
| "learning_rate": 4.218759482358765e-06, |
| "loss": 0.1479, |
| "step": 376 |
| }, |
| { |
| "epoch": 0.2967335694608422, |
| "grad_norm": 1.0433752096490876, |
| "learning_rate": 4.214024459924221e-06, |
| "loss": 0.1561, |
| "step": 377 |
| }, |
| { |
| "epoch": 0.2975206611570248, |
| "grad_norm": 0.9985242964251728, |
| "learning_rate": 4.209277806359956e-06, |
| "loss": 0.1486, |
| "step": 378 |
| }, |
| { |
| "epoch": 0.2983077528532074, |
| "grad_norm": 0.9830203630270159, |
| "learning_rate": 4.204519553876095e-06, |
| "loss": 0.153, |
| "step": 379 |
| }, |
| { |
| "epoch": 0.29909484454939, |
| "grad_norm": 1.0623334389041004, |
| "learning_rate": 4.199749734761473e-06, |
| "loss": 0.1584, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.29988193624557263, |
| "grad_norm": 1.007050119697646, |
| "learning_rate": 4.194968381383414e-06, |
| "loss": 0.162, |
| "step": 381 |
| }, |
| { |
| "epoch": 0.3006690279417552, |
| "grad_norm": 0.9212276043601202, |
| "learning_rate": 4.1901755261875116e-06, |
| "loss": 0.1417, |
| "step": 382 |
| }, |
| { |
| "epoch": 0.30145611963793784, |
| "grad_norm": 1.0195210503773229, |
| "learning_rate": 4.18537120169741e-06, |
| "loss": 0.1631, |
| "step": 383 |
| }, |
| { |
| "epoch": 0.3022432113341204, |
| "grad_norm": 0.9791393783618954, |
| "learning_rate": 4.1805554405145805e-06, |
| "loss": 0.151, |
| "step": 384 |
| }, |
| { |
| "epoch": 0.30303030303030304, |
| "grad_norm": 0.9560471554995319, |
| "learning_rate": 4.175728275318105e-06, |
| "loss": 0.1537, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.3038173947264856, |
| "grad_norm": 0.9732207377472094, |
| "learning_rate": 4.170889738864448e-06, |
| "loss": 0.1541, |
| "step": 386 |
| }, |
| { |
| "epoch": 0.30460448642266824, |
| "grad_norm": 1.0273971232086052, |
| "learning_rate": 4.166039863987241e-06, |
| "loss": 0.1623, |
| "step": 387 |
| }, |
| { |
| "epoch": 0.30539157811885087, |
| "grad_norm": 1.0066781633766182, |
| "learning_rate": 4.161178683597055e-06, |
| "loss": 0.1623, |
| "step": 388 |
| }, |
| { |
| "epoch": 0.30617866981503344, |
| "grad_norm": 0.9519906303887405, |
| "learning_rate": 4.156306230681178e-06, |
| "loss": 0.1606, |
| "step": 389 |
| }, |
| { |
| "epoch": 0.3069657615112161, |
| "grad_norm": 1.0274010773396909, |
| "learning_rate": 4.151422538303393e-06, |
| "loss": 0.1588, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.30775285320739865, |
| "grad_norm": 1.0092975668609663, |
| "learning_rate": 4.1465276396037516e-06, |
| "loss": 0.1549, |
| "step": 391 |
| }, |
| { |
| "epoch": 0.3085399449035813, |
| "grad_norm": 0.9952464194936945, |
| "learning_rate": 4.141621567798351e-06, |
| "loss": 0.1468, |
| "step": 392 |
| }, |
| { |
| "epoch": 0.30932703659976385, |
| "grad_norm": 0.9809046515355889, |
| "learning_rate": 4.136704356179105e-06, |
| "loss": 0.1509, |
| "step": 393 |
| }, |
| { |
| "epoch": 0.3101141282959465, |
| "grad_norm": 1.0624718823483572, |
| "learning_rate": 4.131776038113524e-06, |
| "loss": 0.1629, |
| "step": 394 |
| }, |
| { |
| "epoch": 0.3109012199921291, |
| "grad_norm": 0.9361042861540975, |
| "learning_rate": 4.126836647044484e-06, |
| "loss": 0.1453, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.3116883116883117, |
| "grad_norm": 1.0675702598561039, |
| "learning_rate": 4.121886216489999e-06, |
| "loss": 0.1657, |
| "step": 396 |
| }, |
| { |
| "epoch": 0.3124754033844943, |
| "grad_norm": 1.0221190108601212, |
| "learning_rate": 4.116924780042997e-06, |
| "loss": 0.1609, |
| "step": 397 |
| }, |
| { |
| "epoch": 0.3132624950806769, |
| "grad_norm": 0.98812521742716, |
| "learning_rate": 4.111952371371091e-06, |
| "loss": 0.1488, |
| "step": 398 |
| }, |
| { |
| "epoch": 0.3140495867768595, |
| "grad_norm": 0.9689235987787954, |
| "learning_rate": 4.106969024216348e-06, |
| "loss": 0.1547, |
| "step": 399 |
| }, |
| { |
| "epoch": 0.3148366784730421, |
| "grad_norm": 1.0046393279094348, |
| "learning_rate": 4.101974772395066e-06, |
| "loss": 0.1467, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.3156237701692247, |
| "grad_norm": 0.968527185963086, |
| "learning_rate": 4.096969649797534e-06, |
| "loss": 0.1432, |
| "step": 401 |
| }, |
| { |
| "epoch": 0.31641086186540734, |
| "grad_norm": 1.0188815460176754, |
| "learning_rate": 4.091953690387815e-06, |
| "loss": 0.1521, |
| "step": 402 |
| }, |
| { |
| "epoch": 0.3171979535615899, |
| "grad_norm": 1.035965071382904, |
| "learning_rate": 4.086926928203506e-06, |
| "loss": 0.1575, |
| "step": 403 |
| }, |
| { |
| "epoch": 0.31798504525777255, |
| "grad_norm": 1.0400881738148544, |
| "learning_rate": 4.081889397355509e-06, |
| "loss": 0.1646, |
| "step": 404 |
| }, |
| { |
| "epoch": 0.3187721369539551, |
| "grad_norm": 1.0353365656909388, |
| "learning_rate": 4.076841132027805e-06, |
| "loss": 0.1578, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.31955922865013775, |
| "grad_norm": 0.9785090873779988, |
| "learning_rate": 4.071782166477213e-06, |
| "loss": 0.1485, |
| "step": 406 |
| }, |
| { |
| "epoch": 0.3203463203463203, |
| "grad_norm": 1.0365440161437718, |
| "learning_rate": 4.066712535033164e-06, |
| "loss": 0.1644, |
| "step": 407 |
| }, |
| { |
| "epoch": 0.32113341204250295, |
| "grad_norm": 0.9337858697268638, |
| "learning_rate": 4.061632272097467e-06, |
| "loss": 0.1396, |
| "step": 408 |
| }, |
| { |
| "epoch": 0.3219205037386856, |
| "grad_norm": 0.9930564105014524, |
| "learning_rate": 4.056541412144073e-06, |
| "loss": 0.1466, |
| "step": 409 |
| }, |
| { |
| "epoch": 0.32270759543486816, |
| "grad_norm": 1.0123860857315623, |
| "learning_rate": 4.051439989718845e-06, |
| "loss": 0.1718, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.3234946871310508, |
| "grad_norm": 0.9886983463565112, |
| "learning_rate": 4.0463280394393216e-06, |
| "loss": 0.1465, |
| "step": 411 |
| }, |
| { |
| "epoch": 0.32428177882723336, |
| "grad_norm": 0.9489896550219313, |
| "learning_rate": 4.041205595994478e-06, |
| "loss": 0.1553, |
| "step": 412 |
| }, |
| { |
| "epoch": 0.325068870523416, |
| "grad_norm": 0.935055903913981, |
| "learning_rate": 4.036072694144501e-06, |
| "loss": 0.1486, |
| "step": 413 |
| }, |
| { |
| "epoch": 0.32585596221959856, |
| "grad_norm": 1.0109287737515016, |
| "learning_rate": 4.030929368720539e-06, |
| "loss": 0.1563, |
| "step": 414 |
| }, |
| { |
| "epoch": 0.3266430539157812, |
| "grad_norm": 0.9682667210224672, |
| "learning_rate": 4.025775654624481e-06, |
| "loss": 0.154, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.3274301456119638, |
| "grad_norm": 0.9195115794003238, |
| "learning_rate": 4.020611586828705e-06, |
| "loss": 0.1433, |
| "step": 416 |
| }, |
| { |
| "epoch": 0.3282172373081464, |
| "grad_norm": 0.886911970381121, |
| "learning_rate": 4.015437200375855e-06, |
| "loss": 0.1374, |
| "step": 417 |
| }, |
| { |
| "epoch": 0.329004329004329, |
| "grad_norm": 1.021240159520919, |
| "learning_rate": 4.01025253037859e-06, |
| "loss": 0.1567, |
| "step": 418 |
| }, |
| { |
| "epoch": 0.3297914207005116, |
| "grad_norm": 0.9462875663398478, |
| "learning_rate": 4.005057612019353e-06, |
| "loss": 0.1516, |
| "step": 419 |
| }, |
| { |
| "epoch": 0.3305785123966942, |
| "grad_norm": 0.9850150964188347, |
| "learning_rate": 3.9998524805501335e-06, |
| "loss": 0.149, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.3313656040928768, |
| "grad_norm": 1.0466582919958791, |
| "learning_rate": 3.994637171292223e-06, |
| "loss": 0.1504, |
| "step": 421 |
| }, |
| { |
| "epoch": 0.3321526957890594, |
| "grad_norm": 0.9892923577104711, |
| "learning_rate": 3.989411719635979e-06, |
| "loss": 0.1465, |
| "step": 422 |
| }, |
| { |
| "epoch": 0.33293978748524206, |
| "grad_norm": 1.0797299646481497, |
| "learning_rate": 3.984176161040585e-06, |
| "loss": 0.1655, |
| "step": 423 |
| }, |
| { |
| "epoch": 0.33372687918142463, |
| "grad_norm": 1.0280855278386611, |
| "learning_rate": 3.978930531033807e-06, |
| "loss": 0.1614, |
| "step": 424 |
| }, |
| { |
| "epoch": 0.33451397087760726, |
| "grad_norm": 1.0013220452351206, |
| "learning_rate": 3.973674865211754e-06, |
| "loss": 0.1529, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.33530106257378983, |
| "grad_norm": 1.0185754306460224, |
| "learning_rate": 3.968409199238639e-06, |
| "loss": 0.1535, |
| "step": 426 |
| }, |
| { |
| "epoch": 0.33608815426997246, |
| "grad_norm": 0.9680713968649773, |
| "learning_rate": 3.963133568846533e-06, |
| "loss": 0.1532, |
| "step": 427 |
| }, |
| { |
| "epoch": 0.33687524596615503, |
| "grad_norm": 1.0465737054070945, |
| "learning_rate": 3.957848009835125e-06, |
| "loss": 0.1557, |
| "step": 428 |
| }, |
| { |
| "epoch": 0.33766233766233766, |
| "grad_norm": 1.0072097384887637, |
| "learning_rate": 3.952552558071475e-06, |
| "loss": 0.1686, |
| "step": 429 |
| }, |
| { |
| "epoch": 0.3384494293585203, |
| "grad_norm": 1.0495298691679416, |
| "learning_rate": 3.947247249489779e-06, |
| "loss": 0.1487, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.33923652105470287, |
| "grad_norm": 1.0214586461562896, |
| "learning_rate": 3.941932120091116e-06, |
| "loss": 0.1621, |
| "step": 431 |
| }, |
| { |
| "epoch": 0.3400236127508855, |
| "grad_norm": 1.0494096714602847, |
| "learning_rate": 3.93660720594321e-06, |
| "loss": 0.1598, |
| "step": 432 |
| }, |
| { |
| "epoch": 0.34081070444706807, |
| "grad_norm": 1.0334818385570048, |
| "learning_rate": 3.93127254318018e-06, |
| "loss": 0.1577, |
| "step": 433 |
| }, |
| { |
| "epoch": 0.3415977961432507, |
| "grad_norm": 0.9700994625756835, |
| "learning_rate": 3.925928168002302e-06, |
| "loss": 0.1526, |
| "step": 434 |
| }, |
| { |
| "epoch": 0.34238488783943327, |
| "grad_norm": 1.047736033995709, |
| "learning_rate": 3.920574116675756e-06, |
| "loss": 0.1581, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.3431719795356159, |
| "grad_norm": 1.0493869403649712, |
| "learning_rate": 3.915210425532383e-06, |
| "loss": 0.1495, |
| "step": 436 |
| }, |
| { |
| "epoch": 0.34395907123179853, |
| "grad_norm": 1.010254528268069, |
| "learning_rate": 3.90983713096944e-06, |
| "loss": 0.1539, |
| "step": 437 |
| }, |
| { |
| "epoch": 0.3447461629279811, |
| "grad_norm": 0.9846398029609658, |
| "learning_rate": 3.9044542694493515e-06, |
| "loss": 0.1463, |
| "step": 438 |
| }, |
| { |
| "epoch": 0.34553325462416373, |
| "grad_norm": 1.2083136674514858, |
| "learning_rate": 3.899061877499461e-06, |
| "loss": 0.1601, |
| "step": 439 |
| }, |
| { |
| "epoch": 0.3463203463203463, |
| "grad_norm": 0.97978554217786, |
| "learning_rate": 3.893659991711782e-06, |
| "loss": 0.139, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.34710743801652894, |
| "grad_norm": 1.1022405018344112, |
| "learning_rate": 3.888248648742756e-06, |
| "loss": 0.1617, |
| "step": 441 |
| }, |
| { |
| "epoch": 0.3478945297127115, |
| "grad_norm": 1.0077367730076683, |
| "learning_rate": 3.882827885312999e-06, |
| "loss": 0.1488, |
| "step": 442 |
| }, |
| { |
| "epoch": 0.34868162140889414, |
| "grad_norm": 1.0119669193080498, |
| "learning_rate": 3.877397738207051e-06, |
| "loss": 0.1433, |
| "step": 443 |
| }, |
| { |
| "epoch": 0.34946871310507677, |
| "grad_norm": 0.9336119872704435, |
| "learning_rate": 3.8719582442731276e-06, |
| "loss": 0.1393, |
| "step": 444 |
| }, |
| { |
| "epoch": 0.35025580480125934, |
| "grad_norm": 1.0144372790745282, |
| "learning_rate": 3.866509440422873e-06, |
| "loss": 0.1515, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.35104289649744197, |
| "grad_norm": 1.0618851735205919, |
| "learning_rate": 3.861051363631107e-06, |
| "loss": 0.1403, |
| "step": 446 |
| }, |
| { |
| "epoch": 0.35182998819362454, |
| "grad_norm": 1.0256940692518137, |
| "learning_rate": 3.855584050935574e-06, |
| "loss": 0.1533, |
| "step": 447 |
| }, |
| { |
| "epoch": 0.3526170798898072, |
| "grad_norm": 1.004262449427633, |
| "learning_rate": 3.85010753943669e-06, |
| "loss": 0.1437, |
| "step": 448 |
| }, |
| { |
| "epoch": 0.35340417158598975, |
| "grad_norm": 0.9608822952661715, |
| "learning_rate": 3.844621866297295e-06, |
| "loss": 0.1374, |
| "step": 449 |
| }, |
| { |
| "epoch": 0.3541912632821724, |
| "grad_norm": 1.032805552257636, |
| "learning_rate": 3.839127068742399e-06, |
| "loss": 0.1612, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.354978354978355, |
| "grad_norm": 1.089158815357864, |
| "learning_rate": 3.833623184058926e-06, |
| "loss": 0.1564, |
| "step": 451 |
| }, |
| { |
| "epoch": 0.3557654466745376, |
| "grad_norm": 1.0527347217082683, |
| "learning_rate": 3.8281102495954684e-06, |
| "loss": 0.1475, |
| "step": 452 |
| }, |
| { |
| "epoch": 0.3565525383707202, |
| "grad_norm": 1.012969201356965, |
| "learning_rate": 3.8225883027620245e-06, |
| "loss": 0.1443, |
| "step": 453 |
| }, |
| { |
| "epoch": 0.3573396300669028, |
| "grad_norm": 0.9952397622221426, |
| "learning_rate": 3.817057381029752e-06, |
| "loss": 0.1488, |
| "step": 454 |
| }, |
| { |
| "epoch": 0.3581267217630854, |
| "grad_norm": 0.9773911500192811, |
| "learning_rate": 3.811517521930711e-06, |
| "loss": 0.1419, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.358913813459268, |
| "grad_norm": 1.04344141144674, |
| "learning_rate": 3.805968763057609e-06, |
| "loss": 0.1335, |
| "step": 456 |
| }, |
| { |
| "epoch": 0.3597009051554506, |
| "grad_norm": 0.9127224357677829, |
| "learning_rate": 3.8004111420635453e-06, |
| "loss": 0.1421, |
| "step": 457 |
| }, |
| { |
| "epoch": 0.36048799685163324, |
| "grad_norm": 0.948335811441799, |
| "learning_rate": 3.7948446966617568e-06, |
| "loss": 0.1545, |
| "step": 458 |
| }, |
| { |
| "epoch": 0.3612750885478158, |
| "grad_norm": 1.054156015531643, |
| "learning_rate": 3.7892694646253624e-06, |
| "loss": 0.1462, |
| "step": 459 |
| }, |
| { |
| "epoch": 0.36206218024399844, |
| "grad_norm": 1.0883694334017704, |
| "learning_rate": 3.783685483787105e-06, |
| "loss": 0.1469, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.362849271940181, |
| "grad_norm": 1.0265972829923478, |
| "learning_rate": 3.7780927920390965e-06, |
| "loss": 0.1572, |
| "step": 461 |
| }, |
| { |
| "epoch": 0.36363636363636365, |
| "grad_norm": 0.9351090223515385, |
| "learning_rate": 3.772491427332557e-06, |
| "loss": 0.1317, |
| "step": 462 |
| }, |
| { |
| "epoch": 0.3644234553325462, |
| "grad_norm": 0.96672130032329, |
| "learning_rate": 3.766881427677563e-06, |
| "loss": 0.1474, |
| "step": 463 |
| }, |
| { |
| "epoch": 0.36521054702872885, |
| "grad_norm": 0.9284227954997755, |
| "learning_rate": 3.761262831142788e-06, |
| "loss": 0.144, |
| "step": 464 |
| }, |
| { |
| "epoch": 0.3659976387249115, |
| "grad_norm": 1.02329013434613, |
| "learning_rate": 3.755635675855238e-06, |
| "loss": 0.1459, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.36678473042109405, |
| "grad_norm": 0.9548918394606087, |
| "learning_rate": 3.7500000000000005e-06, |
| "loss": 0.1431, |
| "step": 466 |
| }, |
| { |
| "epoch": 0.3675718221172767, |
| "grad_norm": 1.0029018534160843, |
| "learning_rate": 3.744355841819983e-06, |
| "loss": 0.1551, |
| "step": 467 |
| }, |
| { |
| "epoch": 0.36835891381345925, |
| "grad_norm": 1.0170466682076178, |
| "learning_rate": 3.7387032396156497e-06, |
| "loss": 0.1574, |
| "step": 468 |
| }, |
| { |
| "epoch": 0.3691460055096419, |
| "grad_norm": 0.950504547373793, |
| "learning_rate": 3.7330422317447686e-06, |
| "loss": 0.1413, |
| "step": 469 |
| }, |
| { |
| "epoch": 0.36993309720582446, |
| "grad_norm": 1.00490564394254, |
| "learning_rate": 3.7273728566221447e-06, |
| "loss": 0.1539, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.3707201889020071, |
| "grad_norm": 1.0241073155182219, |
| "learning_rate": 3.721695152719364e-06, |
| "loss": 0.1505, |
| "step": 471 |
| }, |
| { |
| "epoch": 0.3715072805981897, |
| "grad_norm": 1.0650129974030413, |
| "learning_rate": 3.716009158564528e-06, |
| "loss": 0.1517, |
| "step": 472 |
| }, |
| { |
| "epoch": 0.3722943722943723, |
| "grad_norm": 1.0412500508709561, |
| "learning_rate": 3.710314912741997e-06, |
| "loss": 0.1447, |
| "step": 473 |
| }, |
| { |
| "epoch": 0.3730814639905549, |
| "grad_norm": 1.0273490151395026, |
| "learning_rate": 3.7046124538921237e-06, |
| "loss": 0.1429, |
| "step": 474 |
| }, |
| { |
| "epoch": 0.3738685556867375, |
| "grad_norm": 0.9952543111661871, |
| "learning_rate": 3.698901820710995e-06, |
| "loss": 0.1418, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.3746556473829201, |
| "grad_norm": 1.0824700054534682, |
| "learning_rate": 3.693183051950168e-06, |
| "loss": 0.1437, |
| "step": 476 |
| }, |
| { |
| "epoch": 0.3754427390791027, |
| "grad_norm": 1.0142752196453109, |
| "learning_rate": 3.6874561864164056e-06, |
| "loss": 0.1435, |
| "step": 477 |
| }, |
| { |
| "epoch": 0.3762298307752853, |
| "grad_norm": 0.9888106276082754, |
| "learning_rate": 3.6817212629714135e-06, |
| "loss": 0.1395, |
| "step": 478 |
| }, |
| { |
| "epoch": 0.37701692247146795, |
| "grad_norm": 0.9673698851206235, |
| "learning_rate": 3.675978320531579e-06, |
| "loss": 0.1425, |
| "step": 479 |
| }, |
| { |
| "epoch": 0.3778040141676505, |
| "grad_norm": 1.096283920214562, |
| "learning_rate": 3.670227398067705e-06, |
| "loss": 0.1515, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.37859110586383315, |
| "grad_norm": 1.0303413811027284, |
| "learning_rate": 3.664468534604745e-06, |
| "loss": 0.1462, |
| "step": 481 |
| }, |
| { |
| "epoch": 0.37937819756001573, |
| "grad_norm": 0.9550517801003708, |
| "learning_rate": 3.6587017692215387e-06, |
| "loss": 0.1483, |
| "step": 482 |
| }, |
| { |
| "epoch": 0.38016528925619836, |
| "grad_norm": 1.0499765951347195, |
| "learning_rate": 3.6529271410505483e-06, |
| "loss": 0.1516, |
| "step": 483 |
| }, |
| { |
| "epoch": 0.38095238095238093, |
| "grad_norm": 1.0612285971687154, |
| "learning_rate": 3.6471446892775896e-06, |
| "loss": 0.145, |
| "step": 484 |
| }, |
| { |
| "epoch": 0.38173947264856356, |
| "grad_norm": 0.9976574649139153, |
| "learning_rate": 3.6413544531415712e-06, |
| "loss": 0.1493, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.3825265643447462, |
| "grad_norm": 1.011974051278155, |
| "learning_rate": 3.635556471934224e-06, |
| "loss": 0.1557, |
| "step": 486 |
| }, |
| { |
| "epoch": 0.38331365604092876, |
| "grad_norm": 1.015959048224715, |
| "learning_rate": 3.629750784999835e-06, |
| "loss": 0.152, |
| "step": 487 |
| }, |
| { |
| "epoch": 0.3841007477371114, |
| "grad_norm": 0.9638439392236781, |
| "learning_rate": 3.623937431734982e-06, |
| "loss": 0.1464, |
| "step": 488 |
| }, |
| { |
| "epoch": 0.38488783943329397, |
| "grad_norm": 0.9820530085625633, |
| "learning_rate": 3.6181164515882663e-06, |
| "loss": 0.1468, |
| "step": 489 |
| }, |
| { |
| "epoch": 0.3856749311294766, |
| "grad_norm": 0.9281524539517508, |
| "learning_rate": 3.6122878840600417e-06, |
| "loss": 0.1451, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.38646202282565917, |
| "grad_norm": 1.039305922376239, |
| "learning_rate": 3.606451768702151e-06, |
| "loss": 0.1486, |
| "step": 491 |
| }, |
| { |
| "epoch": 0.3872491145218418, |
| "grad_norm": 1.026987888426606, |
| "learning_rate": 3.600608145117656e-06, |
| "loss": 0.1381, |
| "step": 492 |
| }, |
| { |
| "epoch": 0.3880362062180244, |
| "grad_norm": 1.058827889093346, |
| "learning_rate": 3.594757052960566e-06, |
| "loss": 0.1555, |
| "step": 493 |
| }, |
| { |
| "epoch": 0.388823297914207, |
| "grad_norm": 1.0027016575115129, |
| "learning_rate": 3.588898531935573e-06, |
| "loss": 0.1413, |
| "step": 494 |
| }, |
| { |
| "epoch": 0.38961038961038963, |
| "grad_norm": 1.0766471714794614, |
| "learning_rate": 3.583032621797778e-06, |
| "loss": 0.1418, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.3903974813065722, |
| "grad_norm": 1.0326313481110534, |
| "learning_rate": 3.5771593623524263e-06, |
| "loss": 0.1345, |
| "step": 496 |
| }, |
| { |
| "epoch": 0.39118457300275483, |
| "grad_norm": 0.9649958546178075, |
| "learning_rate": 3.5712787934546336e-06, |
| "loss": 0.1397, |
| "step": 497 |
| }, |
| { |
| "epoch": 0.3919716646989374, |
| "grad_norm": 1.0461258832079, |
| "learning_rate": 3.5653909550091138e-06, |
| "loss": 0.16, |
| "step": 498 |
| }, |
| { |
| "epoch": 0.39275875639512003, |
| "grad_norm": 0.9741702004779168, |
| "learning_rate": 3.559495886969916e-06, |
| "loss": 0.1366, |
| "step": 499 |
| }, |
| { |
| "epoch": 0.39354584809130266, |
| "grad_norm": 0.9875184829637668, |
| "learning_rate": 3.553593629340144e-06, |
| "loss": 0.1391, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.39354584809130266, |
| "eval_loss": 0.14773064851760864, |
| "eval_runtime": 18.0322, |
| "eval_samples_per_second": 45.585, |
| "eval_steps_per_second": 5.712, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.39433293978748524, |
| "grad_norm": 0.9563741831859393, |
| "learning_rate": 3.5476842221716915e-06, |
| "loss": 0.1453, |
| "step": 501 |
| }, |
| { |
| "epoch": 0.39512003148366787, |
| "grad_norm": 0.9839000041167648, |
| "learning_rate": 3.541767705564967e-06, |
| "loss": 0.1509, |
| "step": 502 |
| }, |
| { |
| "epoch": 0.39590712317985044, |
| "grad_norm": 0.9666175985112762, |
| "learning_rate": 3.535844119668622e-06, |
| "loss": 0.1436, |
| "step": 503 |
| }, |
| { |
| "epoch": 0.39669421487603307, |
| "grad_norm": 1.0513295542177603, |
| "learning_rate": 3.5299135046792816e-06, |
| "loss": 0.1371, |
| "step": 504 |
| }, |
| { |
| "epoch": 0.39748130657221564, |
| "grad_norm": 1.0136623338528887, |
| "learning_rate": 3.5239759008412666e-06, |
| "loss": 0.1498, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.39826839826839827, |
| "grad_norm": 0.9764920494655156, |
| "learning_rate": 3.518031348446324e-06, |
| "loss": 0.1371, |
| "step": 506 |
| }, |
| { |
| "epoch": 0.3990554899645809, |
| "grad_norm": 1.0113031849627157, |
| "learning_rate": 3.5120798878333544e-06, |
| "loss": 0.1453, |
| "step": 507 |
| }, |
| { |
| "epoch": 0.3998425816607635, |
| "grad_norm": 0.9947509560502654, |
| "learning_rate": 3.506121559388135e-06, |
| "loss": 0.1233, |
| "step": 508 |
| }, |
| { |
| "epoch": 0.4006296733569461, |
| "grad_norm": 1.1135464243984814, |
| "learning_rate": 3.500156403543046e-06, |
| "loss": 0.151, |
| "step": 509 |
| }, |
| { |
| "epoch": 0.4014167650531287, |
| "grad_norm": 1.0687025563863246, |
| "learning_rate": 3.4941844607768007e-06, |
| "loss": 0.1384, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.4022038567493113, |
| "grad_norm": 0.9654525860741724, |
| "learning_rate": 3.488205771614164e-06, |
| "loss": 0.1348, |
| "step": 511 |
| }, |
| { |
| "epoch": 0.4029909484454939, |
| "grad_norm": 1.07357744190682, |
| "learning_rate": 3.4822203766256834e-06, |
| "loss": 0.1412, |
| "step": 512 |
| }, |
| { |
| "epoch": 0.4037780401416765, |
| "grad_norm": 1.2491546536330014, |
| "learning_rate": 3.4762283164274104e-06, |
| "loss": 0.1523, |
| "step": 513 |
| }, |
| { |
| "epoch": 0.40456513183785914, |
| "grad_norm": 1.0398955239354635, |
| "learning_rate": 3.4702296316806243e-06, |
| "loss": 0.1507, |
| "step": 514 |
| }, |
| { |
| "epoch": 0.4053522235340417, |
| "grad_norm": 0.947562520308943, |
| "learning_rate": 3.4642243630915606e-06, |
| "loss": 0.1486, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.40613931523022434, |
| "grad_norm": 0.9405204018759319, |
| "learning_rate": 3.45821255141113e-06, |
| "loss": 0.1287, |
| "step": 516 |
| }, |
| { |
| "epoch": 0.4069264069264069, |
| "grad_norm": 1.01025400774114, |
| "learning_rate": 3.452194237434642e-06, |
| "loss": 0.1349, |
| "step": 517 |
| }, |
| { |
| "epoch": 0.40771349862258954, |
| "grad_norm": 1.0404932578099988, |
| "learning_rate": 3.446169462001534e-06, |
| "loss": 0.1508, |
| "step": 518 |
| }, |
| { |
| "epoch": 0.4085005903187721, |
| "grad_norm": 1.029425420995215, |
| "learning_rate": 3.4401382659950868e-06, |
| "loss": 0.1362, |
| "step": 519 |
| }, |
| { |
| "epoch": 0.40928768201495475, |
| "grad_norm": 1.025768159905711, |
| "learning_rate": 3.4341006903421493e-06, |
| "loss": 0.1437, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.4100747737111374, |
| "grad_norm": 0.9507044448226175, |
| "learning_rate": 3.4280567760128658e-06, |
| "loss": 0.1393, |
| "step": 521 |
| }, |
| { |
| "epoch": 0.41086186540731995, |
| "grad_norm": 1.0374082813027519, |
| "learning_rate": 3.4220065640203916e-06, |
| "loss": 0.16, |
| "step": 522 |
| }, |
| { |
| "epoch": 0.4116489571035026, |
| "grad_norm": 0.9378353888939086, |
| "learning_rate": 3.415950095420616e-06, |
| "loss": 0.1355, |
| "step": 523 |
| }, |
| { |
| "epoch": 0.41243604879968515, |
| "grad_norm": 0.924561930587711, |
| "learning_rate": 3.4098874113118863e-06, |
| "loss": 0.1452, |
| "step": 524 |
| }, |
| { |
| "epoch": 0.4132231404958678, |
| "grad_norm": 0.9505049489489825, |
| "learning_rate": 3.403818552834727e-06, |
| "loss": 0.1448, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.41401023219205035, |
| "grad_norm": 0.9701870488491394, |
| "learning_rate": 3.397743561171562e-06, |
| "loss": 0.1341, |
| "step": 526 |
| }, |
| { |
| "epoch": 0.414797323888233, |
| "grad_norm": 0.9122288876708122, |
| "learning_rate": 3.3916624775464318e-06, |
| "loss": 0.1291, |
| "step": 527 |
| }, |
| { |
| "epoch": 0.4155844155844156, |
| "grad_norm": 0.985150804267496, |
| "learning_rate": 3.385575343224718e-06, |
| "loss": 0.141, |
| "step": 528 |
| }, |
| { |
| "epoch": 0.4163715072805982, |
| "grad_norm": 0.9910844190276262, |
| "learning_rate": 3.3794821995128606e-06, |
| "loss": 0.1473, |
| "step": 529 |
| }, |
| { |
| "epoch": 0.4171585989767808, |
| "grad_norm": 0.9925292173111532, |
| "learning_rate": 3.3733830877580796e-06, |
| "loss": 0.1492, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.4179456906729634, |
| "grad_norm": 0.9483537804421872, |
| "learning_rate": 3.3672780493480927e-06, |
| "loss": 0.1476, |
| "step": 531 |
| }, |
| { |
| "epoch": 0.418732782369146, |
| "grad_norm": 0.9716970355806354, |
| "learning_rate": 3.3611671257108323e-06, |
| "loss": 0.1288, |
| "step": 532 |
| }, |
| { |
| "epoch": 0.4195198740653286, |
| "grad_norm": 1.0525983321400059, |
| "learning_rate": 3.3550503583141726e-06, |
| "loss": 0.1541, |
| "step": 533 |
| }, |
| { |
| "epoch": 0.4203069657615112, |
| "grad_norm": 0.9476841095185634, |
| "learning_rate": 3.3489277886656373e-06, |
| "loss": 0.1395, |
| "step": 534 |
| }, |
| { |
| "epoch": 0.42109405745769385, |
| "grad_norm": 0.8883884320293254, |
| "learning_rate": 3.342799458312127e-06, |
| "loss": 0.1374, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.4218811491538764, |
| "grad_norm": 1.026818858865084, |
| "learning_rate": 3.336665408839633e-06, |
| "loss": 0.1413, |
| "step": 536 |
| }, |
| { |
| "epoch": 0.42266824085005905, |
| "grad_norm": 0.9146805645048051, |
| "learning_rate": 3.330525681872954e-06, |
| "loss": 0.1352, |
| "step": 537 |
| }, |
| { |
| "epoch": 0.4234553325462416, |
| "grad_norm": 1.0439955820386841, |
| "learning_rate": 3.3243803190754166e-06, |
| "loss": 0.1482, |
| "step": 538 |
| }, |
| { |
| "epoch": 0.42424242424242425, |
| "grad_norm": 0.9964413472110166, |
| "learning_rate": 3.3182293621485923e-06, |
| "loss": 0.1524, |
| "step": 539 |
| }, |
| { |
| "epoch": 0.42502951593860683, |
| "grad_norm": 0.9626977177442709, |
| "learning_rate": 3.312072852832012e-06, |
| "loss": 0.1427, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.42581660763478946, |
| "grad_norm": 0.8939878261316884, |
| "learning_rate": 3.3059108329028845e-06, |
| "loss": 0.1283, |
| "step": 541 |
| }, |
| { |
| "epoch": 0.4266036993309721, |
| "grad_norm": 1.0155176108909485, |
| "learning_rate": 3.299743344175814e-06, |
| "loss": 0.1434, |
| "step": 542 |
| }, |
| { |
| "epoch": 0.42739079102715466, |
| "grad_norm": 0.983969699589635, |
| "learning_rate": 3.293570428502515e-06, |
| "loss": 0.1479, |
| "step": 543 |
| }, |
| { |
| "epoch": 0.4281778827233373, |
| "grad_norm": 1.006021089515589, |
| "learning_rate": 3.287392127771526e-06, |
| "loss": 0.1386, |
| "step": 544 |
| }, |
| { |
| "epoch": 0.42896497441951986, |
| "grad_norm": 0.9594215523929834, |
| "learning_rate": 3.2812084839079316e-06, |
| "loss": 0.1326, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.4297520661157025, |
| "grad_norm": 0.958170693889915, |
| "learning_rate": 3.275019538873071e-06, |
| "loss": 0.1418, |
| "step": 546 |
| }, |
| { |
| "epoch": 0.43053915781188506, |
| "grad_norm": 1.0256772435691563, |
| "learning_rate": 3.268825334664259e-06, |
| "loss": 0.1526, |
| "step": 547 |
| }, |
| { |
| "epoch": 0.4313262495080677, |
| "grad_norm": 1.0552921930006323, |
| "learning_rate": 3.2626259133144955e-06, |
| "loss": 0.1441, |
| "step": 548 |
| }, |
| { |
| "epoch": 0.43211334120425027, |
| "grad_norm": 1.0605556822333475, |
| "learning_rate": 3.2564213168921867e-06, |
| "loss": 0.1431, |
| "step": 549 |
| }, |
| { |
| "epoch": 0.4329004329004329, |
| "grad_norm": 0.982864648139079, |
| "learning_rate": 3.2502115875008523e-06, |
| "loss": 0.149, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.4336875245966155, |
| "grad_norm": 1.0049025924744737, |
| "learning_rate": 3.2439967672788462e-06, |
| "loss": 0.1334, |
| "step": 551 |
| }, |
| { |
| "epoch": 0.4344746162927981, |
| "grad_norm": 0.9263761254409442, |
| "learning_rate": 3.2377768983990677e-06, |
| "loss": 0.1401, |
| "step": 552 |
| }, |
| { |
| "epoch": 0.43526170798898073, |
| "grad_norm": 1.0390883988019344, |
| "learning_rate": 3.2315520230686747e-06, |
| "loss": 0.1493, |
| "step": 553 |
| }, |
| { |
| "epoch": 0.4360487996851633, |
| "grad_norm": 0.9913895430005143, |
| "learning_rate": 3.2253221835287984e-06, |
| "loss": 0.1406, |
| "step": 554 |
| }, |
| { |
| "epoch": 0.43683589138134593, |
| "grad_norm": 0.9801664753977715, |
| "learning_rate": 3.2190874220542577e-06, |
| "loss": 0.1341, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.4376229830775285, |
| "grad_norm": 0.9519041413125566, |
| "learning_rate": 3.2128477809532687e-06, |
| "loss": 0.1469, |
| "step": 556 |
| }, |
| { |
| "epoch": 0.43841007477371113, |
| "grad_norm": 1.0289764585305627, |
| "learning_rate": 3.2066033025671612e-06, |
| "loss": 0.1473, |
| "step": 557 |
| }, |
| { |
| "epoch": 0.43919716646989376, |
| "grad_norm": 1.0005689404595521, |
| "learning_rate": 3.200354029270091e-06, |
| "loss": 0.1477, |
| "step": 558 |
| }, |
| { |
| "epoch": 0.43998425816607634, |
| "grad_norm": 1.056463362355424, |
| "learning_rate": 3.1941000034687516e-06, |
| "loss": 0.1488, |
| "step": 559 |
| }, |
| { |
| "epoch": 0.44077134986225897, |
| "grad_norm": 0.9574144607007496, |
| "learning_rate": 3.187841267602084e-06, |
| "loss": 0.1445, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.44155844155844154, |
| "grad_norm": 0.9562953543913302, |
| "learning_rate": 3.1815778641409924e-06, |
| "loss": 0.1414, |
| "step": 561 |
| }, |
| { |
| "epoch": 0.44234553325462417, |
| "grad_norm": 0.9444651486667015, |
| "learning_rate": 3.1753098355880557e-06, |
| "loss": 0.138, |
| "step": 562 |
| }, |
| { |
| "epoch": 0.44313262495080674, |
| "grad_norm": 0.9465351053953429, |
| "learning_rate": 3.169037224477236e-06, |
| "loss": 0.1437, |
| "step": 563 |
| }, |
| { |
| "epoch": 0.44391971664698937, |
| "grad_norm": 1.0206836940486426, |
| "learning_rate": 3.162760073373594e-06, |
| "loss": 0.1411, |
| "step": 564 |
| }, |
| { |
| "epoch": 0.444706808343172, |
| "grad_norm": 1.0878905236564318, |
| "learning_rate": 3.1564784248729965e-06, |
| "loss": 0.1408, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.4454939000393546, |
| "grad_norm": 1.0130102955883906, |
| "learning_rate": 3.15019232160183e-06, |
| "loss": 0.1428, |
| "step": 566 |
| }, |
| { |
| "epoch": 0.4462809917355372, |
| "grad_norm": 0.980086016231054, |
| "learning_rate": 3.1439018062167092e-06, |
| "loss": 0.143, |
| "step": 567 |
| }, |
| { |
| "epoch": 0.4470680834317198, |
| "grad_norm": 1.0249915137559014, |
| "learning_rate": 3.1376069214041917e-06, |
| "loss": 0.1471, |
| "step": 568 |
| }, |
| { |
| "epoch": 0.4478551751279024, |
| "grad_norm": 1.1016327132095007, |
| "learning_rate": 3.1313077098804817e-06, |
| "loss": 0.1606, |
| "step": 569 |
| }, |
| { |
| "epoch": 0.448642266824085, |
| "grad_norm": 1.0411771801722989, |
| "learning_rate": 3.1250042143911462e-06, |
| "loss": 0.1499, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.4494293585202676, |
| "grad_norm": 1.0122030093902548, |
| "learning_rate": 3.118696477710822e-06, |
| "loss": 0.141, |
| "step": 571 |
| }, |
| { |
| "epoch": 0.45021645021645024, |
| "grad_norm": 1.0708872672849516, |
| "learning_rate": 3.1123845426429265e-06, |
| "loss": 0.128, |
| "step": 572 |
| }, |
| { |
| "epoch": 0.4510035419126328, |
| "grad_norm": 1.029737403462412, |
| "learning_rate": 3.106068452019365e-06, |
| "loss": 0.1383, |
| "step": 573 |
| }, |
| { |
| "epoch": 0.45179063360881544, |
| "grad_norm": 0.9988296671107193, |
| "learning_rate": 3.099748248700245e-06, |
| "loss": 0.1376, |
| "step": 574 |
| }, |
| { |
| "epoch": 0.452577725304998, |
| "grad_norm": 1.0475513726672416, |
| "learning_rate": 3.0934239755735782e-06, |
| "loss": 0.1355, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.45336481700118064, |
| "grad_norm": 1.0654745191838768, |
| "learning_rate": 3.0870956755549973e-06, |
| "loss": 0.143, |
| "step": 576 |
| }, |
| { |
| "epoch": 0.4541519086973632, |
| "grad_norm": 0.9397526290083124, |
| "learning_rate": 3.0807633915874585e-06, |
| "loss": 0.1406, |
| "step": 577 |
| }, |
| { |
| "epoch": 0.45493900039354584, |
| "grad_norm": 1.052837564760308, |
| "learning_rate": 3.0744271666409526e-06, |
| "loss": 0.1454, |
| "step": 578 |
| }, |
| { |
| "epoch": 0.4557260920897285, |
| "grad_norm": 1.1289865006459998, |
| "learning_rate": 3.0680870437122145e-06, |
| "loss": 0.1554, |
| "step": 579 |
| }, |
| { |
| "epoch": 0.45651318378591105, |
| "grad_norm": 0.9614320131595296, |
| "learning_rate": 3.0617430658244295e-06, |
| "loss": 0.1368, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.4573002754820937, |
| "grad_norm": 0.9849943444472453, |
| "learning_rate": 3.0553952760269427e-06, |
| "loss": 0.1372, |
| "step": 581 |
| }, |
| { |
| "epoch": 0.45808736717827625, |
| "grad_norm": 0.9938446057985301, |
| "learning_rate": 3.0490437173949656e-06, |
| "loss": 0.1397, |
| "step": 582 |
| }, |
| { |
| "epoch": 0.4588744588744589, |
| "grad_norm": 0.9430590894578916, |
| "learning_rate": 3.0426884330292844e-06, |
| "loss": 0.1404, |
| "step": 583 |
| }, |
| { |
| "epoch": 0.45966155057064145, |
| "grad_norm": 0.8998337127762756, |
| "learning_rate": 3.0363294660559685e-06, |
| "loss": 0.133, |
| "step": 584 |
| }, |
| { |
| "epoch": 0.4604486422668241, |
| "grad_norm": 0.9469777276964015, |
| "learning_rate": 3.0299668596260755e-06, |
| "loss": 0.1429, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.4612357339630067, |
| "grad_norm": 0.9961208676961326, |
| "learning_rate": 3.023600656915362e-06, |
| "loss": 0.1381, |
| "step": 586 |
| }, |
| { |
| "epoch": 0.4620228256591893, |
| "grad_norm": 0.9726679702119774, |
| "learning_rate": 3.017230901123985e-06, |
| "loss": 0.1391, |
| "step": 587 |
| }, |
| { |
| "epoch": 0.4628099173553719, |
| "grad_norm": 1.016233754336966, |
| "learning_rate": 3.0108576354762176e-06, |
| "loss": 0.1464, |
| "step": 588 |
| }, |
| { |
| "epoch": 0.4635970090515545, |
| "grad_norm": 0.891890572894692, |
| "learning_rate": 3.0044809032201448e-06, |
| "loss": 0.1312, |
| "step": 589 |
| }, |
| { |
| "epoch": 0.4643841007477371, |
| "grad_norm": 0.9300922465018149, |
| "learning_rate": 2.9981007476273787e-06, |
| "loss": 0.1272, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.4651711924439197, |
| "grad_norm": 1.0381540629264334, |
| "learning_rate": 2.9917172119927607e-06, |
| "loss": 0.1479, |
| "step": 591 |
| }, |
| { |
| "epoch": 0.4659582841401023, |
| "grad_norm": 1.0642195977009175, |
| "learning_rate": 2.9853303396340695e-06, |
| "loss": 0.1364, |
| "step": 592 |
| }, |
| { |
| "epoch": 0.46674537583628495, |
| "grad_norm": 0.9295272897205104, |
| "learning_rate": 2.9789401738917244e-06, |
| "loss": 0.1249, |
| "step": 593 |
| }, |
| { |
| "epoch": 0.4675324675324675, |
| "grad_norm": 1.0180029223750298, |
| "learning_rate": 2.9725467581284944e-06, |
| "loss": 0.1407, |
| "step": 594 |
| }, |
| { |
| "epoch": 0.46831955922865015, |
| "grad_norm": 1.1385262618991847, |
| "learning_rate": 2.966150135729203e-06, |
| "loss": 0.1502, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.4691066509248327, |
| "grad_norm": 1.0067715931565462, |
| "learning_rate": 2.9597503501004345e-06, |
| "loss": 0.1286, |
| "step": 596 |
| }, |
| { |
| "epoch": 0.46989374262101535, |
| "grad_norm": 0.9465710841629198, |
| "learning_rate": 2.9533474446702346e-06, |
| "loss": 0.1358, |
| "step": 597 |
| }, |
| { |
| "epoch": 0.4706808343171979, |
| "grad_norm": 1.04804051578767, |
| "learning_rate": 2.946941462887824e-06, |
| "loss": 0.1333, |
| "step": 598 |
| }, |
| { |
| "epoch": 0.47146792601338056, |
| "grad_norm": 1.0917713383450702, |
| "learning_rate": 2.940532448223296e-06, |
| "loss": 0.1462, |
| "step": 599 |
| }, |
| { |
| "epoch": 0.4722550177095632, |
| "grad_norm": 0.9580513732250364, |
| "learning_rate": 2.9341204441673267e-06, |
| "loss": 0.1321, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.47304210940574576, |
| "grad_norm": 0.9439921102070582, |
| "learning_rate": 2.927705494230875e-06, |
| "loss": 0.1441, |
| "step": 601 |
| }, |
| { |
| "epoch": 0.4738292011019284, |
| "grad_norm": 1.0178216949448748, |
| "learning_rate": 2.9212876419448943e-06, |
| "loss": 0.1405, |
| "step": 602 |
| }, |
| { |
| "epoch": 0.47461629279811096, |
| "grad_norm": 1.0297426762245179, |
| "learning_rate": 2.9148669308600298e-06, |
| "loss": 0.1392, |
| "step": 603 |
| }, |
| { |
| "epoch": 0.4754033844942936, |
| "grad_norm": 0.9415986568330708, |
| "learning_rate": 2.9084434045463255e-06, |
| "loss": 0.1282, |
| "step": 604 |
| }, |
| { |
| "epoch": 0.47619047619047616, |
| "grad_norm": 1.0337230890115443, |
| "learning_rate": 2.9020171065929327e-06, |
| "loss": 0.1394, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.4769775678866588, |
| "grad_norm": 1.0540052550471415, |
| "learning_rate": 2.895588080607807e-06, |
| "loss": 0.1472, |
| "step": 606 |
| }, |
| { |
| "epoch": 0.4777646595828414, |
| "grad_norm": 1.0081872244466563, |
| "learning_rate": 2.8891563702174174e-06, |
| "loss": 0.1372, |
| "step": 607 |
| }, |
| { |
| "epoch": 0.478551751279024, |
| "grad_norm": 1.0145019904402564, |
| "learning_rate": 2.8827220190664505e-06, |
| "loss": 0.1399, |
| "step": 608 |
| }, |
| { |
| "epoch": 0.4793388429752066, |
| "grad_norm": 1.0258604105718838, |
| "learning_rate": 2.8762850708175098e-06, |
| "loss": 0.1499, |
| "step": 609 |
| }, |
| { |
| "epoch": 0.4801259346713892, |
| "grad_norm": 1.0836484331180423, |
| "learning_rate": 2.869845569150825e-06, |
| "loss": 0.1388, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.4809130263675718, |
| "grad_norm": 0.9946389106293178, |
| "learning_rate": 2.863403557763951e-06, |
| "loss": 0.1323, |
| "step": 611 |
| }, |
| { |
| "epoch": 0.4817001180637544, |
| "grad_norm": 0.9968164583365795, |
| "learning_rate": 2.856959080371474e-06, |
| "loss": 0.1402, |
| "step": 612 |
| }, |
| { |
| "epoch": 0.48248720975993703, |
| "grad_norm": 1.0526146596249044, |
| "learning_rate": 2.8505121807047155e-06, |
| "loss": 0.1342, |
| "step": 613 |
| }, |
| { |
| "epoch": 0.48327430145611966, |
| "grad_norm": 0.9881771003275511, |
| "learning_rate": 2.8440629025114308e-06, |
| "loss": 0.1414, |
| "step": 614 |
| }, |
| { |
| "epoch": 0.48406139315230223, |
| "grad_norm": 1.0170639400089367, |
| "learning_rate": 2.8376112895555184e-06, |
| "loss": 0.1415, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.48484848484848486, |
| "grad_norm": 0.9618458339894986, |
| "learning_rate": 2.83115738561672e-06, |
| "loss": 0.125, |
| "step": 616 |
| }, |
| { |
| "epoch": 0.48563557654466744, |
| "grad_norm": 1.166675709546666, |
| "learning_rate": 2.8247012344903235e-06, |
| "loss": 0.1537, |
| "step": 617 |
| }, |
| { |
| "epoch": 0.48642266824085006, |
| "grad_norm": 1.0308351089525765, |
| "learning_rate": 2.8182428799868643e-06, |
| "loss": 0.1435, |
| "step": 618 |
| }, |
| { |
| "epoch": 0.48720975993703264, |
| "grad_norm": 0.9008466844444718, |
| "learning_rate": 2.811782365931832e-06, |
| "loss": 0.1255, |
| "step": 619 |
| }, |
| { |
| "epoch": 0.48799685163321527, |
| "grad_norm": 1.0328591551300574, |
| "learning_rate": 2.8053197361653684e-06, |
| "loss": 0.1431, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.4887839433293979, |
| "grad_norm": 1.0223227370370647, |
| "learning_rate": 2.7988550345419733e-06, |
| "loss": 0.1302, |
| "step": 621 |
| }, |
| { |
| "epoch": 0.48957103502558047, |
| "grad_norm": 1.0130656273790444, |
| "learning_rate": 2.792388304930207e-06, |
| "loss": 0.1413, |
| "step": 622 |
| }, |
| { |
| "epoch": 0.4903581267217631, |
| "grad_norm": 0.9678629630338841, |
| "learning_rate": 2.7859195912123875e-06, |
| "loss": 0.1411, |
| "step": 623 |
| }, |
| { |
| "epoch": 0.4911452184179457, |
| "grad_norm": 1.0630235458290422, |
| "learning_rate": 2.779448937284302e-06, |
| "loss": 0.144, |
| "step": 624 |
| }, |
| { |
| "epoch": 0.4919323101141283, |
| "grad_norm": 1.1368466359085148, |
| "learning_rate": 2.772976387054899e-06, |
| "loss": 0.1603, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.4927194018103109, |
| "grad_norm": 1.0638972206646764, |
| "learning_rate": 2.766501984445999e-06, |
| "loss": 0.1469, |
| "step": 626 |
| }, |
| { |
| "epoch": 0.4935064935064935, |
| "grad_norm": 0.9878723437777639, |
| "learning_rate": 2.7600257733919887e-06, |
| "loss": 0.1347, |
| "step": 627 |
| }, |
| { |
| "epoch": 0.49429358520267613, |
| "grad_norm": 0.9482438523704221, |
| "learning_rate": 2.7535477978395295e-06, |
| "loss": 0.1301, |
| "step": 628 |
| }, |
| { |
| "epoch": 0.4950806768988587, |
| "grad_norm": 1.0213978192147322, |
| "learning_rate": 2.7470681017472556e-06, |
| "loss": 0.1442, |
| "step": 629 |
| }, |
| { |
| "epoch": 0.49586776859504134, |
| "grad_norm": 1.0113916573838844, |
| "learning_rate": 2.740586729085476e-06, |
| "loss": 0.1477, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.4966548602912239, |
| "grad_norm": 1.0353820062718653, |
| "learning_rate": 2.7341037238358774e-06, |
| "loss": 0.1483, |
| "step": 631 |
| }, |
| { |
| "epoch": 0.49744195198740654, |
| "grad_norm": 1.0552352024187672, |
| "learning_rate": 2.727619129991224e-06, |
| "loss": 0.1328, |
| "step": 632 |
| }, |
| { |
| "epoch": 0.4982290436835891, |
| "grad_norm": 0.9937705442973395, |
| "learning_rate": 2.7211329915550615e-06, |
| "loss": 0.1409, |
| "step": 633 |
| }, |
| { |
| "epoch": 0.49901613537977174, |
| "grad_norm": 1.0486309341654392, |
| "learning_rate": 2.714645352541415e-06, |
| "loss": 0.15, |
| "step": 634 |
| }, |
| { |
| "epoch": 0.49980322707595437, |
| "grad_norm": 1.015369060592149, |
| "learning_rate": 2.7081562569744948e-06, |
| "loss": 0.1298, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.500590318772137, |
| "grad_norm": 1.014091287328762, |
| "learning_rate": 2.701665748888393e-06, |
| "loss": 0.139, |
| "step": 636 |
| }, |
| { |
| "epoch": 0.5013774104683195, |
| "grad_norm": 1.010797057516188, |
| "learning_rate": 2.695173872326788e-06, |
| "loss": 0.1306, |
| "step": 637 |
| }, |
| { |
| "epoch": 0.5021645021645021, |
| "grad_norm": 0.9886264059190445, |
| "learning_rate": 2.6886806713426435e-06, |
| "loss": 0.1493, |
| "step": 638 |
| }, |
| { |
| "epoch": 0.5029515938606848, |
| "grad_norm": 0.9006497838538798, |
| "learning_rate": 2.6821861899979116e-06, |
| "loss": 0.127, |
| "step": 639 |
| }, |
| { |
| "epoch": 0.5037386855568674, |
| "grad_norm": 1.0409028373992908, |
| "learning_rate": 2.6756904723632325e-06, |
| "loss": 0.1453, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.50452577725305, |
| "grad_norm": 0.9741943151013064, |
| "learning_rate": 2.6691935625176357e-06, |
| "loss": 0.1353, |
| "step": 641 |
| }, |
| { |
| "epoch": 0.5053128689492326, |
| "grad_norm": 0.949636504358609, |
| "learning_rate": 2.6626955045482405e-06, |
| "loss": 0.1335, |
| "step": 642 |
| }, |
| { |
| "epoch": 0.5060999606454152, |
| "grad_norm": 0.9249297082390363, |
| "learning_rate": 2.6561963425499575e-06, |
| "loss": 0.1338, |
| "step": 643 |
| }, |
| { |
| "epoch": 0.5068870523415978, |
| "grad_norm": 1.0151555535359889, |
| "learning_rate": 2.649696120625188e-06, |
| "loss": 0.1515, |
| "step": 644 |
| }, |
| { |
| "epoch": 0.5076741440377804, |
| "grad_norm": 1.039472398997662, |
| "learning_rate": 2.643194882883528e-06, |
| "loss": 0.1474, |
| "step": 645 |
| }, |
| { |
| "epoch": 0.508461235733963, |
| "grad_norm": 0.9434610266773801, |
| "learning_rate": 2.6366926734414648e-06, |
| "loss": 0.1304, |
| "step": 646 |
| }, |
| { |
| "epoch": 0.5092483274301456, |
| "grad_norm": 0.8865198426440791, |
| "learning_rate": 2.6301895364220816e-06, |
| "loss": 0.1202, |
| "step": 647 |
| }, |
| { |
| "epoch": 0.5100354191263282, |
| "grad_norm": 0.9546278944005607, |
| "learning_rate": 2.6236855159547527e-06, |
| "loss": 0.1291, |
| "step": 648 |
| }, |
| { |
| "epoch": 0.5108225108225108, |
| "grad_norm": 0.9693013564144493, |
| "learning_rate": 2.6171806561748503e-06, |
| "loss": 0.1339, |
| "step": 649 |
| }, |
| { |
| "epoch": 0.5116096025186935, |
| "grad_norm": 1.0027100891356027, |
| "learning_rate": 2.610675001223441e-06, |
| "loss": 0.1407, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.512396694214876, |
| "grad_norm": 0.8560729540932264, |
| "learning_rate": 2.6041685952469877e-06, |
| "loss": 0.116, |
| "step": 651 |
| }, |
| { |
| "epoch": 0.5131837859110586, |
| "grad_norm": 0.9257606438562741, |
| "learning_rate": 2.597661482397049e-06, |
| "loss": 0.1262, |
| "step": 652 |
| }, |
| { |
| "epoch": 0.5139708776072412, |
| "grad_norm": 1.0514657045725575, |
| "learning_rate": 2.5911537068299803e-06, |
| "loss": 0.1469, |
| "step": 653 |
| }, |
| { |
| "epoch": 0.5147579693034239, |
| "grad_norm": 0.9545058570137028, |
| "learning_rate": 2.584645312706634e-06, |
| "loss": 0.1302, |
| "step": 654 |
| }, |
| { |
| "epoch": 0.5155450609996065, |
| "grad_norm": 0.9392962167917809, |
| "learning_rate": 2.5781363441920614e-06, |
| "loss": 0.1335, |
| "step": 655 |
| }, |
| { |
| "epoch": 0.516332152695789, |
| "grad_norm": 0.9496925045032614, |
| "learning_rate": 2.5716268454552094e-06, |
| "loss": 0.135, |
| "step": 656 |
| }, |
| { |
| "epoch": 0.5171192443919717, |
| "grad_norm": 1.002838327785164, |
| "learning_rate": 2.565116860668625e-06, |
| "loss": 0.1316, |
| "step": 657 |
| }, |
| { |
| "epoch": 0.5179063360881543, |
| "grad_norm": 1.053225279922735, |
| "learning_rate": 2.5586064340081516e-06, |
| "loss": 0.1512, |
| "step": 658 |
| }, |
| { |
| "epoch": 0.5186934277843369, |
| "grad_norm": 0.9441130740117648, |
| "learning_rate": 2.5520956096526323e-06, |
| "loss": 0.131, |
| "step": 659 |
| }, |
| { |
| "epoch": 0.5194805194805194, |
| "grad_norm": 0.9889334534500898, |
| "learning_rate": 2.5455844317836077e-06, |
| "loss": 0.1331, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.5202676111767021, |
| "grad_norm": 0.9646380257679634, |
| "learning_rate": 2.53907294458502e-06, |
| "loss": 0.1291, |
| "step": 661 |
| }, |
| { |
| "epoch": 0.5210547028728847, |
| "grad_norm": 1.0529655580058879, |
| "learning_rate": 2.5325611922429074e-06, |
| "loss": 0.1491, |
| "step": 662 |
| }, |
| { |
| "epoch": 0.5218417945690673, |
| "grad_norm": 0.9926818678117324, |
| "learning_rate": 2.5260492189451076e-06, |
| "loss": 0.1443, |
| "step": 663 |
| }, |
| { |
| "epoch": 0.52262888626525, |
| "grad_norm": 0.9144135454846201, |
| "learning_rate": 2.51953706888096e-06, |
| "loss": 0.1217, |
| "step": 664 |
| }, |
| { |
| "epoch": 0.5234159779614325, |
| "grad_norm": 0.921592430215234, |
| "learning_rate": 2.513024786241001e-06, |
| "loss": 0.1248, |
| "step": 665 |
| }, |
| { |
| "epoch": 0.5242030696576151, |
| "grad_norm": 0.9514782593826102, |
| "learning_rate": 2.5065124152166692e-06, |
| "loss": 0.1297, |
| "step": 666 |
| }, |
| { |
| "epoch": 0.5249901613537977, |
| "grad_norm": 1.0117305817250293, |
| "learning_rate": 2.5e-06, |
| "loss": 0.1497, |
| "step": 667 |
| }, |
| { |
| "epoch": 0.5257772530499804, |
| "grad_norm": 1.013668842181626, |
| "learning_rate": 2.4934875847833308e-06, |
| "loss": 0.1224, |
| "step": 668 |
| }, |
| { |
| "epoch": 0.526564344746163, |
| "grad_norm": 1.000211936689413, |
| "learning_rate": 2.4869752137589994e-06, |
| "loss": 0.1419, |
| "step": 669 |
| }, |
| { |
| "epoch": 0.5273514364423455, |
| "grad_norm": 0.9733370358487723, |
| "learning_rate": 2.48046293111904e-06, |
| "loss": 0.1245, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.5281385281385281, |
| "grad_norm": 1.0646062724041805, |
| "learning_rate": 2.473950781054893e-06, |
| "loss": 0.1383, |
| "step": 671 |
| }, |
| { |
| "epoch": 0.5289256198347108, |
| "grad_norm": 0.930785726380819, |
| "learning_rate": 2.467438807757094e-06, |
| "loss": 0.1295, |
| "step": 672 |
| }, |
| { |
| "epoch": 0.5297127115308934, |
| "grad_norm": 0.9786127857256359, |
| "learning_rate": 2.460927055414981e-06, |
| "loss": 0.146, |
| "step": 673 |
| }, |
| { |
| "epoch": 0.5304998032270759, |
| "grad_norm": 0.9744140929407867, |
| "learning_rate": 2.4544155682163922e-06, |
| "loss": 0.1298, |
| "step": 674 |
| }, |
| { |
| "epoch": 0.5312868949232585, |
| "grad_norm": 0.9484782784554407, |
| "learning_rate": 2.447904390347369e-06, |
| "loss": 0.1278, |
| "step": 675 |
| }, |
| { |
| "epoch": 0.5320739866194412, |
| "grad_norm": 1.1172109264151044, |
| "learning_rate": 2.441393565991849e-06, |
| "loss": 0.146, |
| "step": 676 |
| }, |
| { |
| "epoch": 0.5328610783156238, |
| "grad_norm": 0.9832334733375834, |
| "learning_rate": 2.4348831393313763e-06, |
| "loss": 0.1341, |
| "step": 677 |
| }, |
| { |
| "epoch": 0.5336481700118064, |
| "grad_norm": 0.9993612072993626, |
| "learning_rate": 2.428373154544791e-06, |
| "loss": 0.1348, |
| "step": 678 |
| }, |
| { |
| "epoch": 0.5344352617079889, |
| "grad_norm": 0.8743067419696096, |
| "learning_rate": 2.42186365580794e-06, |
| "loss": 0.1127, |
| "step": 679 |
| }, |
| { |
| "epoch": 0.5352223534041716, |
| "grad_norm": 0.9878917132746777, |
| "learning_rate": 2.4153546872933667e-06, |
| "loss": 0.1289, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.5360094451003542, |
| "grad_norm": 0.9872701224310093, |
| "learning_rate": 2.4088462931700214e-06, |
| "loss": 0.1382, |
| "step": 681 |
| }, |
| { |
| "epoch": 0.5367965367965368, |
| "grad_norm": 1.0291331541759994, |
| "learning_rate": 2.4023385176029516e-06, |
| "loss": 0.1398, |
| "step": 682 |
| }, |
| { |
| "epoch": 0.5375836284927195, |
| "grad_norm": 1.0314844938730774, |
| "learning_rate": 2.3958314047530127e-06, |
| "loss": 0.1407, |
| "step": 683 |
| }, |
| { |
| "epoch": 0.538370720188902, |
| "grad_norm": 0.9922009235690711, |
| "learning_rate": 2.3893249987765598e-06, |
| "loss": 0.1375, |
| "step": 684 |
| }, |
| { |
| "epoch": 0.5391578118850846, |
| "grad_norm": 1.0407160996339295, |
| "learning_rate": 2.3828193438251497e-06, |
| "loss": 0.1356, |
| "step": 685 |
| }, |
| { |
| "epoch": 0.5399449035812672, |
| "grad_norm": 0.9860703004700557, |
| "learning_rate": 2.376314484045248e-06, |
| "loss": 0.132, |
| "step": 686 |
| }, |
| { |
| "epoch": 0.5407319952774499, |
| "grad_norm": 1.0540933767364977, |
| "learning_rate": 2.369810463577919e-06, |
| "loss": 0.1467, |
| "step": 687 |
| }, |
| { |
| "epoch": 0.5415190869736324, |
| "grad_norm": 1.0135356185084303, |
| "learning_rate": 2.3633073265585356e-06, |
| "loss": 0.1381, |
| "step": 688 |
| }, |
| { |
| "epoch": 0.542306178669815, |
| "grad_norm": 0.9743937278639236, |
| "learning_rate": 2.3568051171164724e-06, |
| "loss": 0.1324, |
| "step": 689 |
| }, |
| { |
| "epoch": 0.5430932703659976, |
| "grad_norm": 1.0422560526589146, |
| "learning_rate": 2.350303879374813e-06, |
| "loss": 0.136, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.5438803620621803, |
| "grad_norm": 1.0503391352080245, |
| "learning_rate": 2.3438036574500434e-06, |
| "loss": 0.147, |
| "step": 691 |
| }, |
| { |
| "epoch": 0.5446674537583629, |
| "grad_norm": 0.9557517793781123, |
| "learning_rate": 2.3373044954517603e-06, |
| "loss": 0.1216, |
| "step": 692 |
| }, |
| { |
| "epoch": 0.5454545454545454, |
| "grad_norm": 0.9898057468780994, |
| "learning_rate": 2.330806437482365e-06, |
| "loss": 0.1342, |
| "step": 693 |
| }, |
| { |
| "epoch": 0.546241637150728, |
| "grad_norm": 0.9685373418602369, |
| "learning_rate": 2.3243095276367687e-06, |
| "loss": 0.1294, |
| "step": 694 |
| }, |
| { |
| "epoch": 0.5470287288469107, |
| "grad_norm": 1.0187901801029866, |
| "learning_rate": 2.317813810002089e-06, |
| "loss": 0.1366, |
| "step": 695 |
| }, |
| { |
| "epoch": 0.5478158205430933, |
| "grad_norm": 1.036393473441657, |
| "learning_rate": 2.3113193286573577e-06, |
| "loss": 0.1384, |
| "step": 696 |
| }, |
| { |
| "epoch": 0.5486029122392759, |
| "grad_norm": 0.9735402694275894, |
| "learning_rate": 2.3048261276732133e-06, |
| "loss": 0.1325, |
| "step": 697 |
| }, |
| { |
| "epoch": 0.5493900039354584, |
| "grad_norm": 0.9435211562075637, |
| "learning_rate": 2.298334251111607e-06, |
| "loss": 0.1272, |
| "step": 698 |
| }, |
| { |
| "epoch": 0.5501770956316411, |
| "grad_norm": 0.9238771765346788, |
| "learning_rate": 2.2918437430255056e-06, |
| "loss": 0.1329, |
| "step": 699 |
| }, |
| { |
| "epoch": 0.5509641873278237, |
| "grad_norm": 0.9732329075427437, |
| "learning_rate": 2.285354647458585e-06, |
| "loss": 0.1316, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.5517512790240063, |
| "grad_norm": 1.032698839528823, |
| "learning_rate": 2.2788670084449393e-06, |
| "loss": 0.1438, |
| "step": 701 |
| }, |
| { |
| "epoch": 0.5525383707201889, |
| "grad_norm": 0.9453000245373157, |
| "learning_rate": 2.2723808700087764e-06, |
| "loss": 0.1349, |
| "step": 702 |
| }, |
| { |
| "epoch": 0.5533254624163715, |
| "grad_norm": 1.017447417352295, |
| "learning_rate": 2.2658962761641235e-06, |
| "loss": 0.1346, |
| "step": 703 |
| }, |
| { |
| "epoch": 0.5541125541125541, |
| "grad_norm": 1.0593240948345142, |
| "learning_rate": 2.2594132709145245e-06, |
| "loss": 0.1391, |
| "step": 704 |
| }, |
| { |
| "epoch": 0.5548996458087367, |
| "grad_norm": 1.0524825811903469, |
| "learning_rate": 2.2529318982527453e-06, |
| "loss": 0.1397, |
| "step": 705 |
| }, |
| { |
| "epoch": 0.5556867375049194, |
| "grad_norm": 0.9994684504324172, |
| "learning_rate": 2.246452202160471e-06, |
| "loss": 0.1416, |
| "step": 706 |
| }, |
| { |
| "epoch": 0.5564738292011019, |
| "grad_norm": 1.033406524556106, |
| "learning_rate": 2.2399742266080126e-06, |
| "loss": 0.1269, |
| "step": 707 |
| }, |
| { |
| "epoch": 0.5572609208972845, |
| "grad_norm": 0.9246218098662428, |
| "learning_rate": 2.233498015554002e-06, |
| "loss": 0.1242, |
| "step": 708 |
| }, |
| { |
| "epoch": 0.5580480125934671, |
| "grad_norm": 0.904021008692359, |
| "learning_rate": 2.227023612945102e-06, |
| "loss": 0.1217, |
| "step": 709 |
| }, |
| { |
| "epoch": 0.5588351042896498, |
| "grad_norm": 0.9921447266760961, |
| "learning_rate": 2.220551062715699e-06, |
| "loss": 0.1241, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.5596221959858324, |
| "grad_norm": 1.0273952935358304, |
| "learning_rate": 2.2140804087876134e-06, |
| "loss": 0.1244, |
| "step": 711 |
| }, |
| { |
| "epoch": 0.5604092876820149, |
| "grad_norm": 0.9958421204937957, |
| "learning_rate": 2.207611695069794e-06, |
| "loss": 0.1277, |
| "step": 712 |
| }, |
| { |
| "epoch": 0.5611963793781976, |
| "grad_norm": 1.0226304738126037, |
| "learning_rate": 2.2011449654580266e-06, |
| "loss": 0.1319, |
| "step": 713 |
| }, |
| { |
| "epoch": 0.5619834710743802, |
| "grad_norm": 0.9427241568832295, |
| "learning_rate": 2.1946802638346324e-06, |
| "loss": 0.1208, |
| "step": 714 |
| }, |
| { |
| "epoch": 0.5627705627705628, |
| "grad_norm": 0.9526667511261941, |
| "learning_rate": 2.1882176340681682e-06, |
| "loss": 0.1234, |
| "step": 715 |
| }, |
| { |
| "epoch": 0.5635576544667453, |
| "grad_norm": 0.9726636294262463, |
| "learning_rate": 2.181757120013136e-06, |
| "loss": 0.1241, |
| "step": 716 |
| }, |
| { |
| "epoch": 0.564344746162928, |
| "grad_norm": 0.9577642489452165, |
| "learning_rate": 2.1752987655096765e-06, |
| "loss": 0.1286, |
| "step": 717 |
| }, |
| { |
| "epoch": 0.5651318378591106, |
| "grad_norm": 0.9119267395234483, |
| "learning_rate": 2.1688426143832804e-06, |
| "loss": 0.132, |
| "step": 718 |
| }, |
| { |
| "epoch": 0.5659189295552932, |
| "grad_norm": 0.944139995902989, |
| "learning_rate": 2.162388710444482e-06, |
| "loss": 0.1234, |
| "step": 719 |
| }, |
| { |
| "epoch": 0.5667060212514758, |
| "grad_norm": 0.9645692490749199, |
| "learning_rate": 2.155937097488571e-06, |
| "loss": 0.1251, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.5674931129476584, |
| "grad_norm": 0.9720255939912888, |
| "learning_rate": 2.1494878192952857e-06, |
| "loss": 0.1319, |
| "step": 721 |
| }, |
| { |
| "epoch": 0.568280204643841, |
| "grad_norm": 0.9511775624645177, |
| "learning_rate": 2.1430409196285268e-06, |
| "loss": 0.1327, |
| "step": 722 |
| }, |
| { |
| "epoch": 0.5690672963400236, |
| "grad_norm": 0.9008868958605895, |
| "learning_rate": 2.1365964422360495e-06, |
| "loss": 0.1257, |
| "step": 723 |
| }, |
| { |
| "epoch": 0.5698543880362062, |
| "grad_norm": 1.0087528087899673, |
| "learning_rate": 2.1301544308491755e-06, |
| "loss": 0.1404, |
| "step": 724 |
| }, |
| { |
| "epoch": 0.5706414797323889, |
| "grad_norm": 0.9236247008656706, |
| "learning_rate": 2.1237149291824906e-06, |
| "loss": 0.1122, |
| "step": 725 |
| }, |
| { |
| "epoch": 0.5714285714285714, |
| "grad_norm": 0.9656011748968637, |
| "learning_rate": 2.11727798093355e-06, |
| "loss": 0.1238, |
| "step": 726 |
| }, |
| { |
| "epoch": 0.572215663124754, |
| "grad_norm": 0.9574761410065884, |
| "learning_rate": 2.110843629782583e-06, |
| "loss": 0.1205, |
| "step": 727 |
| }, |
| { |
| "epoch": 0.5730027548209367, |
| "grad_norm": 1.0073901156504852, |
| "learning_rate": 2.1044119193921935e-06, |
| "loss": 0.141, |
| "step": 728 |
| }, |
| { |
| "epoch": 0.5737898465171193, |
| "grad_norm": 1.0135184994615516, |
| "learning_rate": 2.097982893407068e-06, |
| "loss": 0.1391, |
| "step": 729 |
| }, |
| { |
| "epoch": 0.5745769382133018, |
| "grad_norm": 0.9943855979768463, |
| "learning_rate": 2.0915565954536745e-06, |
| "loss": 0.1261, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.5753640299094844, |
| "grad_norm": 0.9263403288426786, |
| "learning_rate": 2.085133069139971e-06, |
| "loss": 0.1199, |
| "step": 731 |
| }, |
| { |
| "epoch": 0.5761511216056671, |
| "grad_norm": 0.9250174978891127, |
| "learning_rate": 2.078712358055106e-06, |
| "loss": 0.1292, |
| "step": 732 |
| }, |
| { |
| "epoch": 0.5769382133018497, |
| "grad_norm": 0.9594966083023022, |
| "learning_rate": 2.0722945057691253e-06, |
| "loss": 0.13, |
| "step": 733 |
| }, |
| { |
| "epoch": 0.5777253049980323, |
| "grad_norm": 0.9996158846425939, |
| "learning_rate": 2.0658795558326745e-06, |
| "loss": 0.1346, |
| "step": 734 |
| }, |
| { |
| "epoch": 0.5785123966942148, |
| "grad_norm": 0.9512575411801811, |
| "learning_rate": 2.059467551776705e-06, |
| "loss": 0.1277, |
| "step": 735 |
| }, |
| { |
| "epoch": 0.5792994883903975, |
| "grad_norm": 0.9956048932258434, |
| "learning_rate": 2.053058537112177e-06, |
| "loss": 0.1364, |
| "step": 736 |
| }, |
| { |
| "epoch": 0.5800865800865801, |
| "grad_norm": 0.9822563309180476, |
| "learning_rate": 2.0466525553297666e-06, |
| "loss": 0.124, |
| "step": 737 |
| }, |
| { |
| "epoch": 0.5808736717827627, |
| "grad_norm": 0.9983781840041562, |
| "learning_rate": 2.0402496498995667e-06, |
| "loss": 0.1347, |
| "step": 738 |
| }, |
| { |
| "epoch": 0.5816607634789452, |
| "grad_norm": 0.9827604253780587, |
| "learning_rate": 2.0338498642707977e-06, |
| "loss": 0.1369, |
| "step": 739 |
| }, |
| { |
| "epoch": 0.5824478551751279, |
| "grad_norm": 1.0756801683767687, |
| "learning_rate": 2.027453241871506e-06, |
| "loss": 0.1323, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.5832349468713105, |
| "grad_norm": 1.0081971409472221, |
| "learning_rate": 2.0210598261082764e-06, |
| "loss": 0.1356, |
| "step": 741 |
| }, |
| { |
| "epoch": 0.5840220385674931, |
| "grad_norm": 0.9898906943423369, |
| "learning_rate": 2.014669660365931e-06, |
| "loss": 0.1368, |
| "step": 742 |
| }, |
| { |
| "epoch": 0.5848091302636758, |
| "grad_norm": 0.9690524566063999, |
| "learning_rate": 2.0082827880072393e-06, |
| "loss": 0.135, |
| "step": 743 |
| }, |
| { |
| "epoch": 0.5855962219598583, |
| "grad_norm": 0.9708412001010785, |
| "learning_rate": 2.0018992523726217e-06, |
| "loss": 0.1252, |
| "step": 744 |
| }, |
| { |
| "epoch": 0.5863833136560409, |
| "grad_norm": 0.9322317029959182, |
| "learning_rate": 1.995519096779855e-06, |
| "loss": 0.1205, |
| "step": 745 |
| }, |
| { |
| "epoch": 0.5871704053522235, |
| "grad_norm": 0.934365579766912, |
| "learning_rate": 1.9891423645237832e-06, |
| "loss": 0.1194, |
| "step": 746 |
| }, |
| { |
| "epoch": 0.5879574970484062, |
| "grad_norm": 0.9421279165900748, |
| "learning_rate": 1.982769098876015e-06, |
| "loss": 0.1319, |
| "step": 747 |
| }, |
| { |
| "epoch": 0.5887445887445888, |
| "grad_norm": 0.9954570155310445, |
| "learning_rate": 1.9763993430846394e-06, |
| "loss": 0.1369, |
| "step": 748 |
| }, |
| { |
| "epoch": 0.5895316804407713, |
| "grad_norm": 0.931791564112829, |
| "learning_rate": 1.970033140373925e-06, |
| "loss": 0.1315, |
| "step": 749 |
| }, |
| { |
| "epoch": 0.5903187721369539, |
| "grad_norm": 0.9679723780616554, |
| "learning_rate": 1.9636705339440327e-06, |
| "loss": 0.1377, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.5911058638331366, |
| "grad_norm": 0.9863750681505877, |
| "learning_rate": 1.957311566970716e-06, |
| "loss": 0.1293, |
| "step": 751 |
| }, |
| { |
| "epoch": 0.5918929555293192, |
| "grad_norm": 0.9598202963903522, |
| "learning_rate": 1.9509562826050353e-06, |
| "loss": 0.1273, |
| "step": 752 |
| }, |
| { |
| "epoch": 0.5926800472255017, |
| "grad_norm": 0.9800756257622318, |
| "learning_rate": 1.944604723973058e-06, |
| "loss": 0.1284, |
| "step": 753 |
| }, |
| { |
| "epoch": 0.5934671389216843, |
| "grad_norm": 1.0001757874575956, |
| "learning_rate": 1.938256934175571e-06, |
| "loss": 0.1303, |
| "step": 754 |
| }, |
| { |
| "epoch": 0.594254230617867, |
| "grad_norm": 0.9299561635899479, |
| "learning_rate": 1.9319129562877863e-06, |
| "loss": 0.1239, |
| "step": 755 |
| }, |
| { |
| "epoch": 0.5950413223140496, |
| "grad_norm": 1.0345095738407815, |
| "learning_rate": 1.925572833359048e-06, |
| "loss": 0.1305, |
| "step": 756 |
| }, |
| { |
| "epoch": 0.5958284140102322, |
| "grad_norm": 1.0520535233317054, |
| "learning_rate": 1.9192366084125423e-06, |
| "loss": 0.1373, |
| "step": 757 |
| }, |
| { |
| "epoch": 0.5966155057064148, |
| "grad_norm": 1.029018918955376, |
| "learning_rate": 1.9129043244450027e-06, |
| "loss": 0.1382, |
| "step": 758 |
| }, |
| { |
| "epoch": 0.5974025974025974, |
| "grad_norm": 1.0294584813791954, |
| "learning_rate": 1.906576024426422e-06, |
| "loss": 0.1368, |
| "step": 759 |
| }, |
| { |
| "epoch": 0.59818968909878, |
| "grad_norm": 0.9330122675132353, |
| "learning_rate": 1.9002517512997555e-06, |
| "loss": 0.1145, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.5989767807949626, |
| "grad_norm": 0.9623676868988281, |
| "learning_rate": 1.8939315479806352e-06, |
| "loss": 0.1335, |
| "step": 761 |
| }, |
| { |
| "epoch": 0.5997638724911453, |
| "grad_norm": 0.9245436138689049, |
| "learning_rate": 1.8876154573570744e-06, |
| "loss": 0.1307, |
| "step": 762 |
| }, |
| { |
| "epoch": 0.6005509641873278, |
| "grad_norm": 0.8942520983106202, |
| "learning_rate": 1.8813035222891785e-06, |
| "loss": 0.1272, |
| "step": 763 |
| }, |
| { |
| "epoch": 0.6013380558835104, |
| "grad_norm": 0.9343347217079488, |
| "learning_rate": 1.8749957856088546e-06, |
| "loss": 0.1317, |
| "step": 764 |
| }, |
| { |
| "epoch": 0.602125147579693, |
| "grad_norm": 0.9393324326245188, |
| "learning_rate": 1.8686922901195197e-06, |
| "loss": 0.1313, |
| "step": 765 |
| }, |
| { |
| "epoch": 0.6029122392758757, |
| "grad_norm": 0.9333264421793994, |
| "learning_rate": 1.8623930785958092e-06, |
| "loss": 0.1226, |
| "step": 766 |
| }, |
| { |
| "epoch": 0.6036993309720582, |
| "grad_norm": 0.9718728327996774, |
| "learning_rate": 1.8560981937832916e-06, |
| "loss": 0.1314, |
| "step": 767 |
| }, |
| { |
| "epoch": 0.6044864226682408, |
| "grad_norm": 0.9437466891844623, |
| "learning_rate": 1.849807678398171e-06, |
| "loss": 0.1271, |
| "step": 768 |
| }, |
| { |
| "epoch": 0.6052735143644234, |
| "grad_norm": 0.9433172532376, |
| "learning_rate": 1.8435215751270048e-06, |
| "loss": 0.1083, |
| "step": 769 |
| }, |
| { |
| "epoch": 0.6060606060606061, |
| "grad_norm": 0.9486485907428178, |
| "learning_rate": 1.8372399266264069e-06, |
| "loss": 0.1245, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.6068476977567887, |
| "grad_norm": 0.9345585045873044, |
| "learning_rate": 1.8309627755227643e-06, |
| "loss": 0.1205, |
| "step": 771 |
| }, |
| { |
| "epoch": 0.6076347894529712, |
| "grad_norm": 1.0082946745736912, |
| "learning_rate": 1.8246901644119447e-06, |
| "loss": 0.1337, |
| "step": 772 |
| }, |
| { |
| "epoch": 0.6084218811491539, |
| "grad_norm": 0.9640602588467792, |
| "learning_rate": 1.8184221358590078e-06, |
| "loss": 0.123, |
| "step": 773 |
| }, |
| { |
| "epoch": 0.6092089728453365, |
| "grad_norm": 1.0256774883323883, |
| "learning_rate": 1.812158732397917e-06, |
| "loss": 0.1331, |
| "step": 774 |
| }, |
| { |
| "epoch": 0.6099960645415191, |
| "grad_norm": 0.9485492161002549, |
| "learning_rate": 1.8058999965312484e-06, |
| "loss": 0.1328, |
| "step": 775 |
| }, |
| { |
| "epoch": 0.6107831562377017, |
| "grad_norm": 0.9763406590147844, |
| "learning_rate": 1.799645970729909e-06, |
| "loss": 0.1309, |
| "step": 776 |
| }, |
| { |
| "epoch": 0.6115702479338843, |
| "grad_norm": 0.9917227985654803, |
| "learning_rate": 1.793396697432839e-06, |
| "loss": 0.1349, |
| "step": 777 |
| }, |
| { |
| "epoch": 0.6123573396300669, |
| "grad_norm": 0.9926597353156553, |
| "learning_rate": 1.7871522190467327e-06, |
| "loss": 0.1303, |
| "step": 778 |
| }, |
| { |
| "epoch": 0.6131444313262495, |
| "grad_norm": 0.9259479446299848, |
| "learning_rate": 1.7809125779457432e-06, |
| "loss": 0.1145, |
| "step": 779 |
| }, |
| { |
| "epoch": 0.6139315230224321, |
| "grad_norm": 0.9291022839595524, |
| "learning_rate": 1.7746778164712024e-06, |
| "loss": 0.119, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.6147186147186147, |
| "grad_norm": 0.951539094911597, |
| "learning_rate": 1.768447976931326e-06, |
| "loss": 0.1261, |
| "step": 781 |
| }, |
| { |
| "epoch": 0.6155057064147973, |
| "grad_norm": 0.993556323836548, |
| "learning_rate": 1.7622231016009333e-06, |
| "loss": 0.1297, |
| "step": 782 |
| }, |
| { |
| "epoch": 0.6162927981109799, |
| "grad_norm": 0.9391833540663885, |
| "learning_rate": 1.7560032327211546e-06, |
| "loss": 0.124, |
| "step": 783 |
| }, |
| { |
| "epoch": 0.6170798898071626, |
| "grad_norm": 1.031878772377542, |
| "learning_rate": 1.7497884124991487e-06, |
| "loss": 0.1308, |
| "step": 784 |
| }, |
| { |
| "epoch": 0.6178669815033452, |
| "grad_norm": 0.9834671756142636, |
| "learning_rate": 1.7435786831078144e-06, |
| "loss": 0.1303, |
| "step": 785 |
| }, |
| { |
| "epoch": 0.6186540731995277, |
| "grad_norm": 0.9859388240495401, |
| "learning_rate": 1.7373740866855043e-06, |
| "loss": 0.1326, |
| "step": 786 |
| }, |
| { |
| "epoch": 0.6194411648957103, |
| "grad_norm": 1.0156315373671152, |
| "learning_rate": 1.731174665335742e-06, |
| "loss": 0.1333, |
| "step": 787 |
| }, |
| { |
| "epoch": 0.620228256591893, |
| "grad_norm": 0.8457875340285443, |
| "learning_rate": 1.724980461126929e-06, |
| "loss": 0.1149, |
| "step": 788 |
| }, |
| { |
| "epoch": 0.6210153482880756, |
| "grad_norm": 0.9812167735229308, |
| "learning_rate": 1.7187915160920692e-06, |
| "loss": 0.1341, |
| "step": 789 |
| }, |
| { |
| "epoch": 0.6218024399842582, |
| "grad_norm": 0.9479256338770862, |
| "learning_rate": 1.7126078722284739e-06, |
| "loss": 0.1171, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.6225895316804407, |
| "grad_norm": 0.9626147311559159, |
| "learning_rate": 1.706429571497486e-06, |
| "loss": 0.1195, |
| "step": 791 |
| }, |
| { |
| "epoch": 0.6233766233766234, |
| "grad_norm": 0.996537388600602, |
| "learning_rate": 1.7002566558241862e-06, |
| "loss": 0.1347, |
| "step": 792 |
| }, |
| { |
| "epoch": 0.624163715072806, |
| "grad_norm": 1.0979789094333103, |
| "learning_rate": 1.694089167097116e-06, |
| "loss": 0.1442, |
| "step": 793 |
| }, |
| { |
| "epoch": 0.6249508067689886, |
| "grad_norm": 0.9903394042224888, |
| "learning_rate": 1.6879271471679887e-06, |
| "loss": 0.1275, |
| "step": 794 |
| }, |
| { |
| "epoch": 0.6257378984651711, |
| "grad_norm": 0.9904940377814807, |
| "learning_rate": 1.681770637851409e-06, |
| "loss": 0.139, |
| "step": 795 |
| }, |
| { |
| "epoch": 0.6265249901613538, |
| "grad_norm": 0.9969630818236452, |
| "learning_rate": 1.675619680924584e-06, |
| "loss": 0.1325, |
| "step": 796 |
| }, |
| { |
| "epoch": 0.6273120818575364, |
| "grad_norm": 1.0558109930918702, |
| "learning_rate": 1.6694743181270474e-06, |
| "loss": 0.1448, |
| "step": 797 |
| }, |
| { |
| "epoch": 0.628099173553719, |
| "grad_norm": 0.956496508451797, |
| "learning_rate": 1.663334591160368e-06, |
| "loss": 0.1217, |
| "step": 798 |
| }, |
| { |
| "epoch": 0.6288862652499017, |
| "grad_norm": 0.9677341961932617, |
| "learning_rate": 1.657200541687874e-06, |
| "loss": 0.136, |
| "step": 799 |
| }, |
| { |
| "epoch": 0.6296733569460842, |
| "grad_norm": 1.0046971327809577, |
| "learning_rate": 1.6510722113343633e-06, |
| "loss": 0.1322, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.6304604486422668, |
| "grad_norm": 0.9254874616921521, |
| "learning_rate": 1.6449496416858285e-06, |
| "loss": 0.1227, |
| "step": 801 |
| }, |
| { |
| "epoch": 0.6312475403384494, |
| "grad_norm": 1.128513971443689, |
| "learning_rate": 1.6388328742891679e-06, |
| "loss": 0.1357, |
| "step": 802 |
| }, |
| { |
| "epoch": 0.6320346320346321, |
| "grad_norm": 1.0186779597478501, |
| "learning_rate": 1.6327219506519082e-06, |
| "loss": 0.1369, |
| "step": 803 |
| }, |
| { |
| "epoch": 0.6328217237308147, |
| "grad_norm": 0.9605839825909683, |
| "learning_rate": 1.6266169122419208e-06, |
| "loss": 0.1222, |
| "step": 804 |
| }, |
| { |
| "epoch": 0.6336088154269972, |
| "grad_norm": 1.0294396317293524, |
| "learning_rate": 1.6205178004871392e-06, |
| "loss": 0.1265, |
| "step": 805 |
| }, |
| { |
| "epoch": 0.6343959071231798, |
| "grad_norm": 0.9797067002404048, |
| "learning_rate": 1.6144246567752831e-06, |
| "loss": 0.1298, |
| "step": 806 |
| }, |
| { |
| "epoch": 0.6351829988193625, |
| "grad_norm": 0.9531255383226177, |
| "learning_rate": 1.6083375224535689e-06, |
| "loss": 0.1204, |
| "step": 807 |
| }, |
| { |
| "epoch": 0.6359700905155451, |
| "grad_norm": 0.9228494475342526, |
| "learning_rate": 1.6022564388284391e-06, |
| "loss": 0.1122, |
| "step": 808 |
| }, |
| { |
| "epoch": 0.6367571822117276, |
| "grad_norm": 0.9854787445128979, |
| "learning_rate": 1.596181447165273e-06, |
| "loss": 0.1287, |
| "step": 809 |
| }, |
| { |
| "epoch": 0.6375442739079102, |
| "grad_norm": 0.9205768495534565, |
| "learning_rate": 1.5901125886881147e-06, |
| "loss": 0.1206, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.6383313656040929, |
| "grad_norm": 0.9798975631304712, |
| "learning_rate": 1.5840499045793845e-06, |
| "loss": 0.1231, |
| "step": 811 |
| }, |
| { |
| "epoch": 0.6391184573002755, |
| "grad_norm": 0.9296415144186752, |
| "learning_rate": 1.5779934359796095e-06, |
| "loss": 0.1202, |
| "step": 812 |
| }, |
| { |
| "epoch": 0.6399055489964581, |
| "grad_norm": 0.9814672778856722, |
| "learning_rate": 1.5719432239871347e-06, |
| "loss": 0.1211, |
| "step": 813 |
| }, |
| { |
| "epoch": 0.6406926406926406, |
| "grad_norm": 0.9650667010737961, |
| "learning_rate": 1.5658993096578512e-06, |
| "loss": 0.123, |
| "step": 814 |
| }, |
| { |
| "epoch": 0.6414797323888233, |
| "grad_norm": 0.9606428490346777, |
| "learning_rate": 1.5598617340049145e-06, |
| "loss": 0.1196, |
| "step": 815 |
| }, |
| { |
| "epoch": 0.6422668240850059, |
| "grad_norm": 0.9865842262641049, |
| "learning_rate": 1.5538305379984661e-06, |
| "loss": 0.1414, |
| "step": 816 |
| }, |
| { |
| "epoch": 0.6430539157811885, |
| "grad_norm": 0.9734870716677574, |
| "learning_rate": 1.547805762565358e-06, |
| "loss": 0.1286, |
| "step": 817 |
| }, |
| { |
| "epoch": 0.6438410074773712, |
| "grad_norm": 0.9439567875437019, |
| "learning_rate": 1.5417874485888706e-06, |
| "loss": 0.1109, |
| "step": 818 |
| }, |
| { |
| "epoch": 0.6446280991735537, |
| "grad_norm": 1.0235998100882107, |
| "learning_rate": 1.5357756369084398e-06, |
| "loss": 0.123, |
| "step": 819 |
| }, |
| { |
| "epoch": 0.6454151908697363, |
| "grad_norm": 1.0269524388931728, |
| "learning_rate": 1.5297703683193755e-06, |
| "loss": 0.1324, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.6462022825659189, |
| "grad_norm": 0.9493642789135233, |
| "learning_rate": 1.5237716835725907e-06, |
| "loss": 0.1125, |
| "step": 821 |
| }, |
| { |
| "epoch": 0.6469893742621016, |
| "grad_norm": 1.162601877497598, |
| "learning_rate": 1.5177796233743174e-06, |
| "loss": 0.1249, |
| "step": 822 |
| }, |
| { |
| "epoch": 0.6477764659582841, |
| "grad_norm": 1.034038388079516, |
| "learning_rate": 1.511794228385837e-06, |
| "loss": 0.1217, |
| "step": 823 |
| }, |
| { |
| "epoch": 0.6485635576544667, |
| "grad_norm": 0.9672639615152381, |
| "learning_rate": 1.5058155392232004e-06, |
| "loss": 0.1208, |
| "step": 824 |
| }, |
| { |
| "epoch": 0.6493506493506493, |
| "grad_norm": 1.080256521732267, |
| "learning_rate": 1.4998435964569552e-06, |
| "loss": 0.1279, |
| "step": 825 |
| }, |
| { |
| "epoch": 0.650137741046832, |
| "grad_norm": 0.9417495839242918, |
| "learning_rate": 1.4938784406118663e-06, |
| "loss": 0.1249, |
| "step": 826 |
| }, |
| { |
| "epoch": 0.6509248327430146, |
| "grad_norm": 1.006350286001005, |
| "learning_rate": 1.4879201121666466e-06, |
| "loss": 0.1251, |
| "step": 827 |
| }, |
| { |
| "epoch": 0.6517119244391971, |
| "grad_norm": 0.968507626389286, |
| "learning_rate": 1.4819686515536763e-06, |
| "loss": 0.1203, |
| "step": 828 |
| }, |
| { |
| "epoch": 0.6524990161353798, |
| "grad_norm": 0.979256644659201, |
| "learning_rate": 1.4760240991587338e-06, |
| "loss": 0.1309, |
| "step": 829 |
| }, |
| { |
| "epoch": 0.6532861078315624, |
| "grad_norm": 1.041802414674734, |
| "learning_rate": 1.4700864953207192e-06, |
| "loss": 0.124, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.654073199527745, |
| "grad_norm": 0.9623673717149763, |
| "learning_rate": 1.4641558803313783e-06, |
| "loss": 0.1153, |
| "step": 831 |
| }, |
| { |
| "epoch": 0.6548602912239276, |
| "grad_norm": 1.0049463554640272, |
| "learning_rate": 1.4582322944350335e-06, |
| "loss": 0.123, |
| "step": 832 |
| }, |
| { |
| "epoch": 0.6556473829201102, |
| "grad_norm": 0.9822560730942449, |
| "learning_rate": 1.4523157778283082e-06, |
| "loss": 0.1253, |
| "step": 833 |
| }, |
| { |
| "epoch": 0.6564344746162928, |
| "grad_norm": 1.0300014906979744, |
| "learning_rate": 1.4464063706598563e-06, |
| "loss": 0.121, |
| "step": 834 |
| }, |
| { |
| "epoch": 0.6572215663124754, |
| "grad_norm": 0.9605069437184749, |
| "learning_rate": 1.440504113030084e-06, |
| "loss": 0.1303, |
| "step": 835 |
| }, |
| { |
| "epoch": 0.658008658008658, |
| "grad_norm": 1.0062748427154549, |
| "learning_rate": 1.4346090449908862e-06, |
| "loss": 0.1254, |
| "step": 836 |
| }, |
| { |
| "epoch": 0.6587957497048406, |
| "grad_norm": 1.003505120930448, |
| "learning_rate": 1.4287212065453681e-06, |
| "loss": 0.1293, |
| "step": 837 |
| }, |
| { |
| "epoch": 0.6595828414010232, |
| "grad_norm": 0.9215109848797975, |
| "learning_rate": 1.4228406376475741e-06, |
| "loss": 0.1156, |
| "step": 838 |
| }, |
| { |
| "epoch": 0.6603699330972058, |
| "grad_norm": 1.0375359512611602, |
| "learning_rate": 1.4169673782022232e-06, |
| "loss": 0.1251, |
| "step": 839 |
| }, |
| { |
| "epoch": 0.6611570247933884, |
| "grad_norm": 1.0075633482471045, |
| "learning_rate": 1.411101468064429e-06, |
| "loss": 0.1273, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.6619441164895711, |
| "grad_norm": 1.0079245494150497, |
| "learning_rate": 1.4052429470394353e-06, |
| "loss": 0.1302, |
| "step": 841 |
| }, |
| { |
| "epoch": 0.6627312081857536, |
| "grad_norm": 0.9589739631373009, |
| "learning_rate": 1.3993918548823453e-06, |
| "loss": 0.1219, |
| "step": 842 |
| }, |
| { |
| "epoch": 0.6635182998819362, |
| "grad_norm": 0.9854619269672102, |
| "learning_rate": 1.3935482312978494e-06, |
| "loss": 0.1264, |
| "step": 843 |
| }, |
| { |
| "epoch": 0.6643053915781189, |
| "grad_norm": 1.0139593156707545, |
| "learning_rate": 1.3877121159399587e-06, |
| "loss": 0.1352, |
| "step": 844 |
| }, |
| { |
| "epoch": 0.6650924832743015, |
| "grad_norm": 0.9879913850797528, |
| "learning_rate": 1.381883548411735e-06, |
| "loss": 0.1252, |
| "step": 845 |
| }, |
| { |
| "epoch": 0.6658795749704841, |
| "grad_norm": 0.9828821822604814, |
| "learning_rate": 1.376062568265018e-06, |
| "loss": 0.1262, |
| "step": 846 |
| }, |
| { |
| "epoch": 0.6666666666666666, |
| "grad_norm": 0.9902383754663022, |
| "learning_rate": 1.370249215000166e-06, |
| "loss": 0.1339, |
| "step": 847 |
| }, |
| { |
| "epoch": 0.6674537583628493, |
| "grad_norm": 1.0169925787410046, |
| "learning_rate": 1.3644435280657765e-06, |
| "loss": 0.1325, |
| "step": 848 |
| }, |
| { |
| "epoch": 0.6682408500590319, |
| "grad_norm": 0.9802382914836032, |
| "learning_rate": 1.3586455468584292e-06, |
| "loss": 0.1294, |
| "step": 849 |
| }, |
| { |
| "epoch": 0.6690279417552145, |
| "grad_norm": 0.9103087080426163, |
| "learning_rate": 1.3528553107224108e-06, |
| "loss": 0.1132, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.669815033451397, |
| "grad_norm": 1.0322697690605673, |
| "learning_rate": 1.347072858949453e-06, |
| "loss": 0.1326, |
| "step": 851 |
| }, |
| { |
| "epoch": 0.6706021251475797, |
| "grad_norm": 0.940497609406273, |
| "learning_rate": 1.3412982307784617e-06, |
| "loss": 0.1142, |
| "step": 852 |
| }, |
| { |
| "epoch": 0.6713892168437623, |
| "grad_norm": 0.9651333506256994, |
| "learning_rate": 1.3355314653952555e-06, |
| "loss": 0.12, |
| "step": 853 |
| }, |
| { |
| "epoch": 0.6721763085399449, |
| "grad_norm": 0.8974492403550183, |
| "learning_rate": 1.3297726019322948e-06, |
| "loss": 0.1252, |
| "step": 854 |
| }, |
| { |
| "epoch": 0.6729634002361276, |
| "grad_norm": 0.9779192150286001, |
| "learning_rate": 1.3240216794684212e-06, |
| "loss": 0.1265, |
| "step": 855 |
| }, |
| { |
| "epoch": 0.6737504919323101, |
| "grad_norm": 1.0060169889058102, |
| "learning_rate": 1.3182787370285865e-06, |
| "loss": 0.1305, |
| "step": 856 |
| }, |
| { |
| "epoch": 0.6745375836284927, |
| "grad_norm": 0.9623311050243877, |
| "learning_rate": 1.3125438135835955e-06, |
| "loss": 0.114, |
| "step": 857 |
| }, |
| { |
| "epoch": 0.6753246753246753, |
| "grad_norm": 1.005880860747008, |
| "learning_rate": 1.3068169480498333e-06, |
| "loss": 0.1237, |
| "step": 858 |
| }, |
| { |
| "epoch": 0.676111767020858, |
| "grad_norm": 1.0295442665880505, |
| "learning_rate": 1.3010981792890053e-06, |
| "loss": 0.141, |
| "step": 859 |
| }, |
| { |
| "epoch": 0.6768988587170406, |
| "grad_norm": 0.9746775819035803, |
| "learning_rate": 1.2953875461078777e-06, |
| "loss": 0.1174, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.6776859504132231, |
| "grad_norm": 0.9651023742880912, |
| "learning_rate": 1.289685087258004e-06, |
| "loss": 0.1179, |
| "step": 861 |
| }, |
| { |
| "epoch": 0.6784730421094057, |
| "grad_norm": 0.9778504990448126, |
| "learning_rate": 1.283990841435473e-06, |
| "loss": 0.1232, |
| "step": 862 |
| }, |
| { |
| "epoch": 0.6792601338055884, |
| "grad_norm": 0.9823411560425596, |
| "learning_rate": 1.2783048472806364e-06, |
| "loss": 0.1214, |
| "step": 863 |
| }, |
| { |
| "epoch": 0.680047225501771, |
| "grad_norm": 0.9509119170509043, |
| "learning_rate": 1.2726271433778559e-06, |
| "loss": 0.1331, |
| "step": 864 |
| }, |
| { |
| "epoch": 0.6808343171979535, |
| "grad_norm": 0.9637465369074552, |
| "learning_rate": 1.266957768255232e-06, |
| "loss": 0.1221, |
| "step": 865 |
| }, |
| { |
| "epoch": 0.6816214088941361, |
| "grad_norm": 1.0309739334485784, |
| "learning_rate": 1.2612967603843512e-06, |
| "loss": 0.1337, |
| "step": 866 |
| }, |
| { |
| "epoch": 0.6824085005903188, |
| "grad_norm": 0.9227141127754309, |
| "learning_rate": 1.2556441581800182e-06, |
| "loss": 0.1118, |
| "step": 867 |
| }, |
| { |
| "epoch": 0.6831955922865014, |
| "grad_norm": 0.983027599423059, |
| "learning_rate": 1.2500000000000007e-06, |
| "loss": 0.1201, |
| "step": 868 |
| }, |
| { |
| "epoch": 0.683982683982684, |
| "grad_norm": 0.969869074022873, |
| "learning_rate": 1.2443643241447629e-06, |
| "loss": 0.1205, |
| "step": 869 |
| }, |
| { |
| "epoch": 0.6847697756788665, |
| "grad_norm": 0.9626068462653994, |
| "learning_rate": 1.2387371688572133e-06, |
| "loss": 0.1294, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.6855568673750492, |
| "grad_norm": 0.9924688128052054, |
| "learning_rate": 1.233118572322437e-06, |
| "loss": 0.1193, |
| "step": 871 |
| }, |
| { |
| "epoch": 0.6863439590712318, |
| "grad_norm": 0.9409212105627156, |
| "learning_rate": 1.2275085726674442e-06, |
| "loss": 0.1186, |
| "step": 872 |
| }, |
| { |
| "epoch": 0.6871310507674144, |
| "grad_norm": 0.9321864217317675, |
| "learning_rate": 1.2219072079609046e-06, |
| "loss": 0.118, |
| "step": 873 |
| }, |
| { |
| "epoch": 0.6879181424635971, |
| "grad_norm": 0.8802354237634122, |
| "learning_rate": 1.2163145162128948e-06, |
| "loss": 0.1092, |
| "step": 874 |
| }, |
| { |
| "epoch": 0.6887052341597796, |
| "grad_norm": 0.9820858832906886, |
| "learning_rate": 1.2107305353746376e-06, |
| "loss": 0.1261, |
| "step": 875 |
| }, |
| { |
| "epoch": 0.6894923258559622, |
| "grad_norm": 1.0214787998802317, |
| "learning_rate": 1.2051553033382426e-06, |
| "loss": 0.121, |
| "step": 876 |
| }, |
| { |
| "epoch": 0.6902794175521448, |
| "grad_norm": 0.9157258726824631, |
| "learning_rate": 1.1995888579364551e-06, |
| "loss": 0.1189, |
| "step": 877 |
| }, |
| { |
| "epoch": 0.6910665092483275, |
| "grad_norm": 0.9531462191249618, |
| "learning_rate": 1.1940312369423919e-06, |
| "loss": 0.1184, |
| "step": 878 |
| }, |
| { |
| "epoch": 0.69185360094451, |
| "grad_norm": 0.9748879770068989, |
| "learning_rate": 1.18848247806929e-06, |
| "loss": 0.1201, |
| "step": 879 |
| }, |
| { |
| "epoch": 0.6926406926406926, |
| "grad_norm": 0.9952760658770881, |
| "learning_rate": 1.1829426189702487e-06, |
| "loss": 0.1211, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.6934277843368752, |
| "grad_norm": 0.9561514586133496, |
| "learning_rate": 1.177411697237977e-06, |
| "loss": 0.1208, |
| "step": 881 |
| }, |
| { |
| "epoch": 0.6942148760330579, |
| "grad_norm": 1.0289787958991654, |
| "learning_rate": 1.1718897504045328e-06, |
| "loss": 0.1329, |
| "step": 882 |
| }, |
| { |
| "epoch": 0.6950019677292405, |
| "grad_norm": 1.012367533381528, |
| "learning_rate": 1.1663768159410748e-06, |
| "loss": 0.1286, |
| "step": 883 |
| }, |
| { |
| "epoch": 0.695789059425423, |
| "grad_norm": 0.9932326189371155, |
| "learning_rate": 1.160872931257602e-06, |
| "loss": 0.1207, |
| "step": 884 |
| }, |
| { |
| "epoch": 0.6965761511216056, |
| "grad_norm": 0.9375475650331836, |
| "learning_rate": 1.1553781337027061e-06, |
| "loss": 0.1162, |
| "step": 885 |
| }, |
| { |
| "epoch": 0.6973632428177883, |
| "grad_norm": 1.0035582921316957, |
| "learning_rate": 1.149892460563311e-06, |
| "loss": 0.1272, |
| "step": 886 |
| }, |
| { |
| "epoch": 0.6981503345139709, |
| "grad_norm": 0.969216495536807, |
| "learning_rate": 1.1444159490644278e-06, |
| "loss": 0.1322, |
| "step": 887 |
| }, |
| { |
| "epoch": 0.6989374262101535, |
| "grad_norm": 0.9727140149487835, |
| "learning_rate": 1.1389486363688935e-06, |
| "loss": 0.1109, |
| "step": 888 |
| }, |
| { |
| "epoch": 0.699724517906336, |
| "grad_norm": 1.035921852021017, |
| "learning_rate": 1.1334905595771274e-06, |
| "loss": 0.125, |
| "step": 889 |
| }, |
| { |
| "epoch": 0.7005116096025187, |
| "grad_norm": 0.9817389469807767, |
| "learning_rate": 1.1280417557268735e-06, |
| "loss": 0.1263, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.7012987012987013, |
| "grad_norm": 0.941993125359632, |
| "learning_rate": 1.12260226179295e-06, |
| "loss": 0.1204, |
| "step": 891 |
| }, |
| { |
| "epoch": 0.7020857929948839, |
| "grad_norm": 0.9776393790876531, |
| "learning_rate": 1.1171721146870015e-06, |
| "loss": 0.1351, |
| "step": 892 |
| }, |
| { |
| "epoch": 0.7028728846910665, |
| "grad_norm": 1.0272253940679958, |
| "learning_rate": 1.1117513512572436e-06, |
| "loss": 0.1297, |
| "step": 893 |
| }, |
| { |
| "epoch": 0.7036599763872491, |
| "grad_norm": 0.9825257474446853, |
| "learning_rate": 1.1063400082882188e-06, |
| "loss": 0.1089, |
| "step": 894 |
| }, |
| { |
| "epoch": 0.7044470680834317, |
| "grad_norm": 0.9941185616779367, |
| "learning_rate": 1.10093812250054e-06, |
| "loss": 0.1182, |
| "step": 895 |
| }, |
| { |
| "epoch": 0.7052341597796143, |
| "grad_norm": 0.9809962090348159, |
| "learning_rate": 1.095545730550649e-06, |
| "loss": 0.1221, |
| "step": 896 |
| }, |
| { |
| "epoch": 0.706021251475797, |
| "grad_norm": 0.9359419759382669, |
| "learning_rate": 1.0901628690305593e-06, |
| "loss": 0.1175, |
| "step": 897 |
| }, |
| { |
| "epoch": 0.7068083431719795, |
| "grad_norm": 1.0254153011332428, |
| "learning_rate": 1.0847895744676173e-06, |
| "loss": 0.1364, |
| "step": 898 |
| }, |
| { |
| "epoch": 0.7075954348681621, |
| "grad_norm": 1.0451822058149052, |
| "learning_rate": 1.0794258833242452e-06, |
| "loss": 0.1341, |
| "step": 899 |
| }, |
| { |
| "epoch": 0.7083825265643447, |
| "grad_norm": 1.0267091614696302, |
| "learning_rate": 1.0740718319976992e-06, |
| "loss": 0.1284, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.7091696182605274, |
| "grad_norm": 0.8928053655240218, |
| "learning_rate": 1.0687274568198208e-06, |
| "loss": 0.1009, |
| "step": 901 |
| }, |
| { |
| "epoch": 0.70995670995671, |
| "grad_norm": 1.057234091878292, |
| "learning_rate": 1.063392794056792e-06, |
| "loss": 0.1346, |
| "step": 902 |
| }, |
| { |
| "epoch": 0.7107438016528925, |
| "grad_norm": 0.9612239375437197, |
| "learning_rate": 1.0580678799088847e-06, |
| "loss": 0.1158, |
| "step": 903 |
| }, |
| { |
| "epoch": 0.7115308933490752, |
| "grad_norm": 0.9876590104136502, |
| "learning_rate": 1.0527527505102213e-06, |
| "loss": 0.1193, |
| "step": 904 |
| }, |
| { |
| "epoch": 0.7123179850452578, |
| "grad_norm": 1.0155629806285287, |
| "learning_rate": 1.0474474419285255e-06, |
| "loss": 0.1206, |
| "step": 905 |
| }, |
| { |
| "epoch": 0.7131050767414404, |
| "grad_norm": 1.108914897353474, |
| "learning_rate": 1.0421519901648759e-06, |
| "loss": 0.1244, |
| "step": 906 |
| }, |
| { |
| "epoch": 0.7138921684376229, |
| "grad_norm": 0.9624208122062576, |
| "learning_rate": 1.0368664311534674e-06, |
| "loss": 0.122, |
| "step": 907 |
| }, |
| { |
| "epoch": 0.7146792601338056, |
| "grad_norm": 0.9051835119610858, |
| "learning_rate": 1.031590800761361e-06, |
| "loss": 0.1115, |
| "step": 908 |
| }, |
| { |
| "epoch": 0.7154663518299882, |
| "grad_norm": 1.001641822545354, |
| "learning_rate": 1.0263251347882467e-06, |
| "loss": 0.1205, |
| "step": 909 |
| }, |
| { |
| "epoch": 0.7162534435261708, |
| "grad_norm": 0.9587134043689033, |
| "learning_rate": 1.021069468966194e-06, |
| "loss": 0.114, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.7170405352223534, |
| "grad_norm": 0.9734138315261187, |
| "learning_rate": 1.0158238389594164e-06, |
| "loss": 0.1237, |
| "step": 911 |
| }, |
| { |
| "epoch": 0.717827626918536, |
| "grad_norm": 0.9654730718585164, |
| "learning_rate": 1.0105882803640215e-06, |
| "loss": 0.1241, |
| "step": 912 |
| }, |
| { |
| "epoch": 0.7186147186147186, |
| "grad_norm": 1.0069324283880368, |
| "learning_rate": 1.0053628287077782e-06, |
| "loss": 0.129, |
| "step": 913 |
| }, |
| { |
| "epoch": 0.7194018103109012, |
| "grad_norm": 0.9724112904663149, |
| "learning_rate": 1.000147519449867e-06, |
| "loss": 0.1217, |
| "step": 914 |
| }, |
| { |
| "epoch": 0.7201889020070839, |
| "grad_norm": 0.9009157337976961, |
| "learning_rate": 9.94942387980648e-07, |
| "loss": 0.1215, |
| "step": 915 |
| }, |
| { |
| "epoch": 0.7209759937032665, |
| "grad_norm": 0.9903710831464596, |
| "learning_rate": 9.89747469621411e-07, |
| "loss": 0.1247, |
| "step": 916 |
| }, |
| { |
| "epoch": 0.721763085399449, |
| "grad_norm": 0.9627081908787005, |
| "learning_rate": 9.845627996241459e-07, |
| "loss": 0.1235, |
| "step": 917 |
| }, |
| { |
| "epoch": 0.7225501770956316, |
| "grad_norm": 0.9666668030573422, |
| "learning_rate": 9.793884131712943e-07, |
| "loss": 0.123, |
| "step": 918 |
| }, |
| { |
| "epoch": 0.7233372687918143, |
| "grad_norm": 0.9644911943474369, |
| "learning_rate": 9.742243453755202e-07, |
| "loss": 0.1142, |
| "step": 919 |
| }, |
| { |
| "epoch": 0.7241243604879969, |
| "grad_norm": 0.9984844200263358, |
| "learning_rate": 9.690706312794618e-07, |
| "loss": 0.1251, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.7249114521841794, |
| "grad_norm": 0.988262708447867, |
| "learning_rate": 9.639273058555004e-07, |
| "loss": 0.1233, |
| "step": 921 |
| }, |
| { |
| "epoch": 0.725698543880362, |
| "grad_norm": 0.9062607929130434, |
| "learning_rate": 9.587944040055225e-07, |
| "loss": 0.1116, |
| "step": 922 |
| }, |
| { |
| "epoch": 0.7264856355765447, |
| "grad_norm": 1.0132516720132552, |
| "learning_rate": 9.536719605606795e-07, |
| "loss": 0.1314, |
| "step": 923 |
| }, |
| { |
| "epoch": 0.7272727272727273, |
| "grad_norm": 0.9210291352044477, |
| "learning_rate": 9.485600102811556e-07, |
| "loss": 0.108, |
| "step": 924 |
| }, |
| { |
| "epoch": 0.7280598189689099, |
| "grad_norm": 1.0099118734494892, |
| "learning_rate": 9.434585878559277e-07, |
| "loss": 0.1172, |
| "step": 925 |
| }, |
| { |
| "epoch": 0.7288469106650924, |
| "grad_norm": 1.0237482529235973, |
| "learning_rate": 9.383677279025347e-07, |
| "loss": 0.1186, |
| "step": 926 |
| }, |
| { |
| "epoch": 0.7296340023612751, |
| "grad_norm": 0.9855331385764105, |
| "learning_rate": 9.332874649668369e-07, |
| "loss": 0.1185, |
| "step": 927 |
| }, |
| { |
| "epoch": 0.7304210940574577, |
| "grad_norm": 0.9369233888911801, |
| "learning_rate": 9.282178335227885e-07, |
| "loss": 0.1067, |
| "step": 928 |
| }, |
| { |
| "epoch": 0.7312081857536403, |
| "grad_norm": 1.025834900254658, |
| "learning_rate": 9.231588679721956e-07, |
| "loss": 0.1256, |
| "step": 929 |
| }, |
| { |
| "epoch": 0.731995277449823, |
| "grad_norm": 1.0004815551544541, |
| "learning_rate": 9.181106026444913e-07, |
| "loss": 0.1171, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.7327823691460055, |
| "grad_norm": 0.9247417584553485, |
| "learning_rate": 9.130730717964948e-07, |
| "loss": 0.1132, |
| "step": 931 |
| }, |
| { |
| "epoch": 0.7335694608421881, |
| "grad_norm": 0.9769073592720867, |
| "learning_rate": 9.08046309612185e-07, |
| "loss": 0.1242, |
| "step": 932 |
| }, |
| { |
| "epoch": 0.7343565525383707, |
| "grad_norm": 0.96681906386633, |
| "learning_rate": 9.030303502024662e-07, |
| "loss": 0.1179, |
| "step": 933 |
| }, |
| { |
| "epoch": 0.7351436442345534, |
| "grad_norm": 1.021595769957744, |
| "learning_rate": 8.980252276049345e-07, |
| "loss": 0.1161, |
| "step": 934 |
| }, |
| { |
| "epoch": 0.7359307359307359, |
| "grad_norm": 0.9231065432942811, |
| "learning_rate": 8.930309757836517e-07, |
| "loss": 0.1149, |
| "step": 935 |
| }, |
| { |
| "epoch": 0.7367178276269185, |
| "grad_norm": 1.026367432921577, |
| "learning_rate": 8.880476286289091e-07, |
| "loss": 0.1284, |
| "step": 936 |
| }, |
| { |
| "epoch": 0.7375049193231011, |
| "grad_norm": 0.981215601065822, |
| "learning_rate": 8.830752199570033e-07, |
| "loss": 0.1133, |
| "step": 937 |
| }, |
| { |
| "epoch": 0.7382920110192838, |
| "grad_norm": 0.9212608185738064, |
| "learning_rate": 8.781137835100021e-07, |
| "loss": 0.1077, |
| "step": 938 |
| }, |
| { |
| "epoch": 0.7390791027154664, |
| "grad_norm": 0.9833427367903659, |
| "learning_rate": 8.731633529555167e-07, |
| "loss": 0.1164, |
| "step": 939 |
| }, |
| { |
| "epoch": 0.7398661944116489, |
| "grad_norm": 0.9854894539977124, |
| "learning_rate": 8.682239618864763e-07, |
| "loss": 0.1155, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.7406532861078315, |
| "grad_norm": 0.9551803394241506, |
| "learning_rate": 8.632956438208962e-07, |
| "loss": 0.1162, |
| "step": 941 |
| }, |
| { |
| "epoch": 0.7414403778040142, |
| "grad_norm": 0.9042419017178762, |
| "learning_rate": 8.583784322016503e-07, |
| "loss": 0.109, |
| "step": 942 |
| }, |
| { |
| "epoch": 0.7422274695001968, |
| "grad_norm": 0.9609816099291726, |
| "learning_rate": 8.534723603962497e-07, |
| "loss": 0.1191, |
| "step": 943 |
| }, |
| { |
| "epoch": 0.7430145611963794, |
| "grad_norm": 1.0149972325544658, |
| "learning_rate": 8.48577461696608e-07, |
| "loss": 0.1192, |
| "step": 944 |
| }, |
| { |
| "epoch": 0.743801652892562, |
| "grad_norm": 0.9584184891745349, |
| "learning_rate": 8.436937693188232e-07, |
| "loss": 0.1267, |
| "step": 945 |
| }, |
| { |
| "epoch": 0.7445887445887446, |
| "grad_norm": 0.9986011121611049, |
| "learning_rate": 8.38821316402946e-07, |
| "loss": 0.1177, |
| "step": 946 |
| }, |
| { |
| "epoch": 0.7453758362849272, |
| "grad_norm": 0.962256278467975, |
| "learning_rate": 8.339601360127592e-07, |
| "loss": 0.1131, |
| "step": 947 |
| }, |
| { |
| "epoch": 0.7461629279811098, |
| "grad_norm": 0.9419406227649391, |
| "learning_rate": 8.291102611355526e-07, |
| "loss": 0.1123, |
| "step": 948 |
| }, |
| { |
| "epoch": 0.7469500196772924, |
| "grad_norm": 0.9728190857016107, |
| "learning_rate": 8.242717246818957e-07, |
| "loss": 0.1197, |
| "step": 949 |
| }, |
| { |
| "epoch": 0.747737111373475, |
| "grad_norm": 1.0169044023539633, |
| "learning_rate": 8.1944455948542e-07, |
| "loss": 0.1219, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.7485242030696576, |
| "grad_norm": 0.9972018368498321, |
| "learning_rate": 8.146287983025902e-07, |
| "loss": 0.1241, |
| "step": 951 |
| }, |
| { |
| "epoch": 0.7493112947658402, |
| "grad_norm": 1.040910663691627, |
| "learning_rate": 8.098244738124888e-07, |
| "loss": 0.1138, |
| "step": 952 |
| }, |
| { |
| "epoch": 0.7500983864620229, |
| "grad_norm": 1.0438538265069202, |
| "learning_rate": 8.050316186165862e-07, |
| "loss": 0.134, |
| "step": 953 |
| }, |
| { |
| "epoch": 0.7508854781582054, |
| "grad_norm": 0.9793759854817412, |
| "learning_rate": 8.002502652385278e-07, |
| "loss": 0.1241, |
| "step": 954 |
| }, |
| { |
| "epoch": 0.751672569854388, |
| "grad_norm": 0.9636283038275181, |
| "learning_rate": 7.954804461239054e-07, |
| "loss": 0.1171, |
| "step": 955 |
| }, |
| { |
| "epoch": 0.7524596615505706, |
| "grad_norm": 0.9416057200961391, |
| "learning_rate": 7.907221936400452e-07, |
| "loss": 0.1194, |
| "step": 956 |
| }, |
| { |
| "epoch": 0.7532467532467533, |
| "grad_norm": 0.9258555685816136, |
| "learning_rate": 7.859755400757793e-07, |
| "loss": 0.1199, |
| "step": 957 |
| }, |
| { |
| "epoch": 0.7540338449429359, |
| "grad_norm": 0.9697084160189383, |
| "learning_rate": 7.812405176412354e-07, |
| "loss": 0.1206, |
| "step": 958 |
| }, |
| { |
| "epoch": 0.7548209366391184, |
| "grad_norm": 1.008937777573116, |
| "learning_rate": 7.76517158467611e-07, |
| "loss": 0.1238, |
| "step": 959 |
| }, |
| { |
| "epoch": 0.755608028335301, |
| "grad_norm": 0.987888023607684, |
| "learning_rate": 7.718054946069589e-07, |
| "loss": 0.1246, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.7563951200314837, |
| "grad_norm": 0.9699505992391279, |
| "learning_rate": 7.671055580319706e-07, |
| "loss": 0.1203, |
| "step": 961 |
| }, |
| { |
| "epoch": 0.7571822117276663, |
| "grad_norm": 0.9382257403962697, |
| "learning_rate": 7.62417380635756e-07, |
| "loss": 0.1151, |
| "step": 962 |
| }, |
| { |
| "epoch": 0.7579693034238488, |
| "grad_norm": 0.9437354430265479, |
| "learning_rate": 7.577409942316305e-07, |
| "loss": 0.1163, |
| "step": 963 |
| }, |
| { |
| "epoch": 0.7587563951200315, |
| "grad_norm": 0.9503047394426882, |
| "learning_rate": 7.530764305528959e-07, |
| "loss": 0.1211, |
| "step": 964 |
| }, |
| { |
| "epoch": 0.7595434868162141, |
| "grad_norm": 1.0356071156065598, |
| "learning_rate": 7.484237212526288e-07, |
| "loss": 0.1273, |
| "step": 965 |
| }, |
| { |
| "epoch": 0.7603305785123967, |
| "grad_norm": 0.9856511020736725, |
| "learning_rate": 7.437828979034606e-07, |
| "loss": 0.1315, |
| "step": 966 |
| }, |
| { |
| "epoch": 0.7611176702085793, |
| "grad_norm": 0.9629717208752256, |
| "learning_rate": 7.391539919973698e-07, |
| "loss": 0.1062, |
| "step": 967 |
| }, |
| { |
| "epoch": 0.7619047619047619, |
| "grad_norm": 0.9607163301231785, |
| "learning_rate": 7.345370349454611e-07, |
| "loss": 0.1189, |
| "step": 968 |
| }, |
| { |
| "epoch": 0.7626918536009445, |
| "grad_norm": 0.9578086152431808, |
| "learning_rate": 7.2993205807776e-07, |
| "loss": 0.1183, |
| "step": 969 |
| }, |
| { |
| "epoch": 0.7634789452971271, |
| "grad_norm": 0.9162347277154375, |
| "learning_rate": 7.253390926429918e-07, |
| "loss": 0.1104, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.7642660369933097, |
| "grad_norm": 0.9822027407988481, |
| "learning_rate": 7.207581698083782e-07, |
| "loss": 0.1304, |
| "step": 971 |
| }, |
| { |
| "epoch": 0.7650531286894924, |
| "grad_norm": 0.940452384125095, |
| "learning_rate": 7.161893206594175e-07, |
| "loss": 0.1168, |
| "step": 972 |
| }, |
| { |
| "epoch": 0.7658402203856749, |
| "grad_norm": 0.9588362662800347, |
| "learning_rate": 7.116325761996818e-07, |
| "loss": 0.1206, |
| "step": 973 |
| }, |
| { |
| "epoch": 0.7666273120818575, |
| "grad_norm": 1.011535036970359, |
| "learning_rate": 7.070879673505976e-07, |
| "loss": 0.1141, |
| "step": 974 |
| }, |
| { |
| "epoch": 0.7674144037780402, |
| "grad_norm": 0.9688050929102817, |
| "learning_rate": 7.025555249512461e-07, |
| "loss": 0.1134, |
| "step": 975 |
| }, |
| { |
| "epoch": 0.7682014954742228, |
| "grad_norm": 0.9177610587932681, |
| "learning_rate": 6.980352797581438e-07, |
| "loss": 0.1089, |
| "step": 976 |
| }, |
| { |
| "epoch": 0.7689885871704053, |
| "grad_norm": 1.0472410615763514, |
| "learning_rate": 6.935272624450432e-07, |
| "loss": 0.1249, |
| "step": 977 |
| }, |
| { |
| "epoch": 0.7697756788665879, |
| "grad_norm": 0.9636506719030409, |
| "learning_rate": 6.890315036027156e-07, |
| "loss": 0.1166, |
| "step": 978 |
| }, |
| { |
| "epoch": 0.7705627705627706, |
| "grad_norm": 0.9412495575321557, |
| "learning_rate": 6.845480337387525e-07, |
| "loss": 0.1195, |
| "step": 979 |
| }, |
| { |
| "epoch": 0.7713498622589532, |
| "grad_norm": 0.9930839323289444, |
| "learning_rate": 6.800768832773505e-07, |
| "loss": 0.1267, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.7721369539551358, |
| "grad_norm": 0.9929049734904327, |
| "learning_rate": 6.756180825591099e-07, |
| "loss": 0.1199, |
| "step": 981 |
| }, |
| { |
| "epoch": 0.7729240456513183, |
| "grad_norm": 0.9842816070024283, |
| "learning_rate": 6.711716618408282e-07, |
| "loss": 0.1179, |
| "step": 982 |
| }, |
| { |
| "epoch": 0.773711137347501, |
| "grad_norm": 0.9944580491304532, |
| "learning_rate": 6.66737651295292e-07, |
| "loss": 0.1198, |
| "step": 983 |
| }, |
| { |
| "epoch": 0.7744982290436836, |
| "grad_norm": 0.9839708888147434, |
| "learning_rate": 6.623160810110765e-07, |
| "loss": 0.1193, |
| "step": 984 |
| }, |
| { |
| "epoch": 0.7752853207398662, |
| "grad_norm": 0.9098159637038072, |
| "learning_rate": 6.579069809923367e-07, |
| "loss": 0.1123, |
| "step": 985 |
| }, |
| { |
| "epoch": 0.7760724124360489, |
| "grad_norm": 0.9606605502093222, |
| "learning_rate": 6.535103811586085e-07, |
| "loss": 0.1174, |
| "step": 986 |
| }, |
| { |
| "epoch": 0.7768595041322314, |
| "grad_norm": 1.0270282467875798, |
| "learning_rate": 6.491263113446005e-07, |
| "loss": 0.1287, |
| "step": 987 |
| }, |
| { |
| "epoch": 0.777646595828414, |
| "grad_norm": 0.9366844684145114, |
| "learning_rate": 6.44754801299998e-07, |
| "loss": 0.1158, |
| "step": 988 |
| }, |
| { |
| "epoch": 0.7784336875245966, |
| "grad_norm": 0.9037352215899601, |
| "learning_rate": 6.403958806892535e-07, |
| "loss": 0.1053, |
| "step": 989 |
| }, |
| { |
| "epoch": 0.7792207792207793, |
| "grad_norm": 0.9440225724017625, |
| "learning_rate": 6.360495790913926e-07, |
| "loss": 0.114, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.7800078709169618, |
| "grad_norm": 0.9351662026461205, |
| "learning_rate": 6.317159259998074e-07, |
| "loss": 0.113, |
| "step": 991 |
| }, |
| { |
| "epoch": 0.7807949626131444, |
| "grad_norm": 0.9872272779542443, |
| "learning_rate": 6.273949508220612e-07, |
| "loss": 0.1217, |
| "step": 992 |
| }, |
| { |
| "epoch": 0.781582054309327, |
| "grad_norm": 1.0021043961378415, |
| "learning_rate": 6.23086682879686e-07, |
| "loss": 0.1194, |
| "step": 993 |
| }, |
| { |
| "epoch": 0.7823691460055097, |
| "grad_norm": 0.9797645648660196, |
| "learning_rate": 6.187911514079834e-07, |
| "loss": 0.1294, |
| "step": 994 |
| }, |
| { |
| "epoch": 0.7831562377016923, |
| "grad_norm": 1.0054784443943467, |
| "learning_rate": 6.14508385555829e-07, |
| "loss": 0.1236, |
| "step": 995 |
| }, |
| { |
| "epoch": 0.7839433293978748, |
| "grad_norm": 0.9433076242026539, |
| "learning_rate": 6.102384143854698e-07, |
| "loss": 0.1147, |
| "step": 996 |
| }, |
| { |
| "epoch": 0.7847304210940574, |
| "grad_norm": 0.9383907844400864, |
| "learning_rate": 6.059812668723336e-07, |
| "loss": 0.115, |
| "step": 997 |
| }, |
| { |
| "epoch": 0.7855175127902401, |
| "grad_norm": 0.9452315722932242, |
| "learning_rate": 6.017369719048255e-07, |
| "loss": 0.1154, |
| "step": 998 |
| }, |
| { |
| "epoch": 0.7863046044864227, |
| "grad_norm": 0.9247930090252802, |
| "learning_rate": 5.975055582841358e-07, |
| "loss": 0.1127, |
| "step": 999 |
| }, |
| { |
| "epoch": 0.7870916961826053, |
| "grad_norm": 1.0061872579787852, |
| "learning_rate": 5.932870547240455e-07, |
| "loss": 0.1183, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.7870916961826053, |
| "eval_loss": 0.11849173903465271, |
| "eval_runtime": 18.0453, |
| "eval_samples_per_second": 45.552, |
| "eval_steps_per_second": 5.708, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.7878787878787878, |
| "grad_norm": 0.9623831636196449, |
| "learning_rate": 5.890814898507277e-07, |
| "loss": 0.1201, |
| "step": 1001 |
| }, |
| { |
| "epoch": 0.7886658795749705, |
| "grad_norm": 1.0229456536544794, |
| "learning_rate": 5.848888922025553e-07, |
| "loss": 0.1223, |
| "step": 1002 |
| }, |
| { |
| "epoch": 0.7894529712711531, |
| "grad_norm": 1.0277580850565635, |
| "learning_rate": 5.8070929022991e-07, |
| "loss": 0.1178, |
| "step": 1003 |
| }, |
| { |
| "epoch": 0.7902400629673357, |
| "grad_norm": 1.0160977169162413, |
| "learning_rate": 5.76542712294983e-07, |
| "loss": 0.1216, |
| "step": 1004 |
| }, |
| { |
| "epoch": 0.7910271546635183, |
| "grad_norm": 0.9181674404236817, |
| "learning_rate": 5.723891866715899e-07, |
| "loss": 0.1118, |
| "step": 1005 |
| }, |
| { |
| "epoch": 0.7918142463597009, |
| "grad_norm": 0.9799718055295829, |
| "learning_rate": 5.682487415449719e-07, |
| "loss": 0.1217, |
| "step": 1006 |
| }, |
| { |
| "epoch": 0.7926013380558835, |
| "grad_norm": 0.980465660075739, |
| "learning_rate": 5.641214050116098e-07, |
| "loss": 0.1252, |
| "step": 1007 |
| }, |
| { |
| "epoch": 0.7933884297520661, |
| "grad_norm": 0.9234542096536653, |
| "learning_rate": 5.600072050790317e-07, |
| "loss": 0.1096, |
| "step": 1008 |
| }, |
| { |
| "epoch": 0.7941755214482488, |
| "grad_norm": 0.8947896913580902, |
| "learning_rate": 5.559061696656199e-07, |
| "loss": 0.1075, |
| "step": 1009 |
| }, |
| { |
| "epoch": 0.7949626131444313, |
| "grad_norm": 0.9790572754851533, |
| "learning_rate": 5.518183266004276e-07, |
| "loss": 0.1171, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.7957497048406139, |
| "grad_norm": 0.9667752845159187, |
| "learning_rate": 5.477437036229832e-07, |
| "loss": 0.1098, |
| "step": 1011 |
| }, |
| { |
| "epoch": 0.7965367965367965, |
| "grad_norm": 1.1238301396219903, |
| "learning_rate": 5.436823283831083e-07, |
| "loss": 0.1373, |
| "step": 1012 |
| }, |
| { |
| "epoch": 0.7973238882329792, |
| "grad_norm": 0.9927017197297952, |
| "learning_rate": 5.396342284407252e-07, |
| "loss": 0.1188, |
| "step": 1013 |
| }, |
| { |
| "epoch": 0.7981109799291618, |
| "grad_norm": 0.9934845633471078, |
| "learning_rate": 5.355994312656734e-07, |
| "loss": 0.1142, |
| "step": 1014 |
| }, |
| { |
| "epoch": 0.7988980716253443, |
| "grad_norm": 0.9432843712008361, |
| "learning_rate": 5.315779642375199e-07, |
| "loss": 0.1158, |
| "step": 1015 |
| }, |
| { |
| "epoch": 0.799685163321527, |
| "grad_norm": 1.0251992909650254, |
| "learning_rate": 5.275698546453775e-07, |
| "loss": 0.1175, |
| "step": 1016 |
| }, |
| { |
| "epoch": 0.8004722550177096, |
| "grad_norm": 1.010003247709555, |
| "learning_rate": 5.235751296877148e-07, |
| "loss": 0.1223, |
| "step": 1017 |
| }, |
| { |
| "epoch": 0.8012593467138922, |
| "grad_norm": 1.0290265946769084, |
| "learning_rate": 5.195938164721767e-07, |
| "loss": 0.1213, |
| "step": 1018 |
| }, |
| { |
| "epoch": 0.8020464384100747, |
| "grad_norm": 1.0065180201235937, |
| "learning_rate": 5.156259420153962e-07, |
| "loss": 0.1238, |
| "step": 1019 |
| }, |
| { |
| "epoch": 0.8028335301062574, |
| "grad_norm": 0.9673994358176322, |
| "learning_rate": 5.116715332428118e-07, |
| "loss": 0.1106, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.80362062180244, |
| "grad_norm": 0.9855226309577549, |
| "learning_rate": 5.077306169884888e-07, |
| "loss": 0.1107, |
| "step": 1021 |
| }, |
| { |
| "epoch": 0.8044077134986226, |
| "grad_norm": 0.9719491378095487, |
| "learning_rate": 5.038032199949313e-07, |
| "loss": 0.1241, |
| "step": 1022 |
| }, |
| { |
| "epoch": 0.8051948051948052, |
| "grad_norm": 0.9767976318784359, |
| "learning_rate": 4.998893689129061e-07, |
| "loss": 0.1165, |
| "step": 1023 |
| }, |
| { |
| "epoch": 0.8059818968909878, |
| "grad_norm": 0.9563874747183178, |
| "learning_rate": 4.959890903012568e-07, |
| "loss": 0.1084, |
| "step": 1024 |
| }, |
| { |
| "epoch": 0.8067689885871704, |
| "grad_norm": 0.9724824504692731, |
| "learning_rate": 4.921024106267283e-07, |
| "loss": 0.1199, |
| "step": 1025 |
| }, |
| { |
| "epoch": 0.807556080283353, |
| "grad_norm": 0.9939320935755256, |
| "learning_rate": 4.882293562637827e-07, |
| "loss": 0.1221, |
| "step": 1026 |
| }, |
| { |
| "epoch": 0.8083431719795356, |
| "grad_norm": 0.9407936084194475, |
| "learning_rate": 4.843699534944258e-07, |
| "loss": 0.114, |
| "step": 1027 |
| }, |
| { |
| "epoch": 0.8091302636757183, |
| "grad_norm": 1.00154528232707, |
| "learning_rate": 4.805242285080222e-07, |
| "loss": 0.116, |
| "step": 1028 |
| }, |
| { |
| "epoch": 0.8099173553719008, |
| "grad_norm": 0.9808095624927836, |
| "learning_rate": 4.7669220740112376e-07, |
| "loss": 0.1166, |
| "step": 1029 |
| }, |
| { |
| "epoch": 0.8107044470680834, |
| "grad_norm": 0.9094800308143445, |
| "learning_rate": 4.728739161772874e-07, |
| "loss": 0.1091, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.811491538764266, |
| "grad_norm": 0.8938814168806941, |
| "learning_rate": 4.690693807469035e-07, |
| "loss": 0.1032, |
| "step": 1031 |
| }, |
| { |
| "epoch": 0.8122786304604487, |
| "grad_norm": 0.9946966466940443, |
| "learning_rate": 4.6527862692701487e-07, |
| "loss": 0.1158, |
| "step": 1032 |
| }, |
| { |
| "epoch": 0.8130657221566312, |
| "grad_norm": 0.9504344369313477, |
| "learning_rate": 4.615016804411465e-07, |
| "loss": 0.1114, |
| "step": 1033 |
| }, |
| { |
| "epoch": 0.8138528138528138, |
| "grad_norm": 0.9453878426287087, |
| "learning_rate": 4.5773856691912726e-07, |
| "loss": 0.1208, |
| "step": 1034 |
| }, |
| { |
| "epoch": 0.8146399055489965, |
| "grad_norm": 1.0239509589587361, |
| "learning_rate": 4.53989311896918e-07, |
| "loss": 0.1186, |
| "step": 1035 |
| }, |
| { |
| "epoch": 0.8154269972451791, |
| "grad_norm": 0.915087834897205, |
| "learning_rate": 4.502539408164386e-07, |
| "loss": 0.106, |
| "step": 1036 |
| }, |
| { |
| "epoch": 0.8162140889413617, |
| "grad_norm": 0.936045456280838, |
| "learning_rate": 4.465324790253922e-07, |
| "loss": 0.1104, |
| "step": 1037 |
| }, |
| { |
| "epoch": 0.8170011806375442, |
| "grad_norm": 0.986165452237028, |
| "learning_rate": 4.428249517770986e-07, |
| "loss": 0.1137, |
| "step": 1038 |
| }, |
| { |
| "epoch": 0.8177882723337269, |
| "grad_norm": 0.9641211589952483, |
| "learning_rate": 4.391313842303166e-07, |
| "loss": 0.1164, |
| "step": 1039 |
| }, |
| { |
| "epoch": 0.8185753640299095, |
| "grad_norm": 0.9334697738707801, |
| "learning_rate": 4.3545180144907857e-07, |
| "loss": 0.113, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.8193624557260921, |
| "grad_norm": 0.9871068120631671, |
| "learning_rate": 4.3178622840251647e-07, |
| "loss": 0.1241, |
| "step": 1041 |
| }, |
| { |
| "epoch": 0.8201495474222747, |
| "grad_norm": 0.9529229208831895, |
| "learning_rate": 4.2813468996469654e-07, |
| "loss": 0.121, |
| "step": 1042 |
| }, |
| { |
| "epoch": 0.8209366391184573, |
| "grad_norm": 0.9336358260352323, |
| "learning_rate": 4.2449721091444545e-07, |
| "loss": 0.1134, |
| "step": 1043 |
| }, |
| { |
| "epoch": 0.8217237308146399, |
| "grad_norm": 1.0307131336178375, |
| "learning_rate": 4.2087381593518716e-07, |
| "loss": 0.1274, |
| "step": 1044 |
| }, |
| { |
| "epoch": 0.8225108225108225, |
| "grad_norm": 0.9433490812621332, |
| "learning_rate": 4.1726452961477147e-07, |
| "loss": 0.116, |
| "step": 1045 |
| }, |
| { |
| "epoch": 0.8232979142070052, |
| "grad_norm": 0.9751019494649681, |
| "learning_rate": 4.136693764453101e-07, |
| "loss": 0.1129, |
| "step": 1046 |
| }, |
| { |
| "epoch": 0.8240850059031877, |
| "grad_norm": 1.056328093801445, |
| "learning_rate": 4.1008838082300743e-07, |
| "loss": 0.1168, |
| "step": 1047 |
| }, |
| { |
| "epoch": 0.8248720975993703, |
| "grad_norm": 1.0079092402082175, |
| "learning_rate": 4.065215670479991e-07, |
| "loss": 0.1258, |
| "step": 1048 |
| }, |
| { |
| "epoch": 0.8256591892955529, |
| "grad_norm": 0.9852819205932097, |
| "learning_rate": 4.02968959324182e-07, |
| "loss": 0.1161, |
| "step": 1049 |
| }, |
| { |
| "epoch": 0.8264462809917356, |
| "grad_norm": 0.9840178810234324, |
| "learning_rate": 3.9943058175905493e-07, |
| "loss": 0.1184, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.8272333726879182, |
| "grad_norm": 1.0203308364665442, |
| "learning_rate": 3.9590645836355275e-07, |
| "loss": 0.1232, |
| "step": 1051 |
| }, |
| { |
| "epoch": 0.8280204643841007, |
| "grad_norm": 0.967005062015959, |
| "learning_rate": 3.923966130518814e-07, |
| "loss": 0.1209, |
| "step": 1052 |
| }, |
| { |
| "epoch": 0.8288075560802833, |
| "grad_norm": 0.9434419280443521, |
| "learning_rate": 3.889010696413606e-07, |
| "loss": 0.1211, |
| "step": 1053 |
| }, |
| { |
| "epoch": 0.829594647776466, |
| "grad_norm": 0.9442684717641329, |
| "learning_rate": 3.8541985185225645e-07, |
| "loss": 0.1078, |
| "step": 1054 |
| }, |
| { |
| "epoch": 0.8303817394726486, |
| "grad_norm": 0.9927726007062886, |
| "learning_rate": 3.819529833076263e-07, |
| "loss": 0.1214, |
| "step": 1055 |
| }, |
| { |
| "epoch": 0.8311688311688312, |
| "grad_norm": 0.9486732370194126, |
| "learning_rate": 3.7850048753315274e-07, |
| "loss": 0.1087, |
| "step": 1056 |
| }, |
| { |
| "epoch": 0.8319559228650137, |
| "grad_norm": 0.9803070634979109, |
| "learning_rate": 3.750623879569895e-07, |
| "loss": 0.1181, |
| "step": 1057 |
| }, |
| { |
| "epoch": 0.8327430145611964, |
| "grad_norm": 1.0010555667628969, |
| "learning_rate": 3.716387079095973e-07, |
| "loss": 0.1172, |
| "step": 1058 |
| }, |
| { |
| "epoch": 0.833530106257379, |
| "grad_norm": 0.9548728126471466, |
| "learning_rate": 3.6822947062359004e-07, |
| "loss": 0.1125, |
| "step": 1059 |
| }, |
| { |
| "epoch": 0.8343171979535616, |
| "grad_norm": 1.0333251380116057, |
| "learning_rate": 3.6483469923357327e-07, |
| "loss": 0.1119, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.8351042896497441, |
| "grad_norm": 0.9473784570136893, |
| "learning_rate": 3.614544167759901e-07, |
| "loss": 0.1136, |
| "step": 1061 |
| }, |
| { |
| "epoch": 0.8358913813459268, |
| "grad_norm": 0.9283902827668026, |
| "learning_rate": 3.5808864618896295e-07, |
| "loss": 0.1004, |
| "step": 1062 |
| }, |
| { |
| "epoch": 0.8366784730421094, |
| "grad_norm": 0.9856546672764643, |
| "learning_rate": 3.5473741031213983e-07, |
| "loss": 0.1136, |
| "step": 1063 |
| }, |
| { |
| "epoch": 0.837465564738292, |
| "grad_norm": 0.955024488651013, |
| "learning_rate": 3.51400731886537e-07, |
| "loss": 0.1199, |
| "step": 1064 |
| }, |
| { |
| "epoch": 0.8382526564344747, |
| "grad_norm": 0.9527578800496054, |
| "learning_rate": 3.4807863355438703e-07, |
| "loss": 0.1178, |
| "step": 1065 |
| }, |
| { |
| "epoch": 0.8390397481306572, |
| "grad_norm": 0.9748866277343534, |
| "learning_rate": 3.447711378589841e-07, |
| "loss": 0.1126, |
| "step": 1066 |
| }, |
| { |
| "epoch": 0.8398268398268398, |
| "grad_norm": 0.9722315190803439, |
| "learning_rate": 3.414782672445291e-07, |
| "loss": 0.1143, |
| "step": 1067 |
| }, |
| { |
| "epoch": 0.8406139315230224, |
| "grad_norm": 1.0223007557494088, |
| "learning_rate": 3.3820004405598157e-07, |
| "loss": 0.1141, |
| "step": 1068 |
| }, |
| { |
| "epoch": 0.8414010232192051, |
| "grad_norm": 0.969999076611352, |
| "learning_rate": 3.3493649053890325e-07, |
| "loss": 0.1161, |
| "step": 1069 |
| }, |
| { |
| "epoch": 0.8421881149153877, |
| "grad_norm": 0.9878840165050939, |
| "learning_rate": 3.3168762883931256e-07, |
| "loss": 0.1164, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.8429752066115702, |
| "grad_norm": 1.030579753477139, |
| "learning_rate": 3.284534810035278e-07, |
| "loss": 0.1258, |
| "step": 1071 |
| }, |
| { |
| "epoch": 0.8437622983077528, |
| "grad_norm": 0.987395029549749, |
| "learning_rate": 3.252340689780245e-07, |
| "loss": 0.1219, |
| "step": 1072 |
| }, |
| { |
| "epoch": 0.8445493900039355, |
| "grad_norm": 0.9750378310364627, |
| "learning_rate": 3.2202941460927977e-07, |
| "loss": 0.1275, |
| "step": 1073 |
| }, |
| { |
| "epoch": 0.8453364817001181, |
| "grad_norm": 0.9446071297273908, |
| "learning_rate": 3.1883953964363057e-07, |
| "loss": 0.1177, |
| "step": 1074 |
| }, |
| { |
| "epoch": 0.8461235733963006, |
| "grad_norm": 0.9469617356782836, |
| "learning_rate": 3.156644657271196e-07, |
| "loss": 0.1128, |
| "step": 1075 |
| }, |
| { |
| "epoch": 0.8469106650924833, |
| "grad_norm": 0.951509014508041, |
| "learning_rate": 3.12504214405355e-07, |
| "loss": 0.108, |
| "step": 1076 |
| }, |
| { |
| "epoch": 0.8476977567886659, |
| "grad_norm": 0.9472516964934904, |
| "learning_rate": 3.093588071233578e-07, |
| "loss": 0.1141, |
| "step": 1077 |
| }, |
| { |
| "epoch": 0.8484848484848485, |
| "grad_norm": 0.9084748554063148, |
| "learning_rate": 3.06228265225422e-07, |
| "loss": 0.1116, |
| "step": 1078 |
| }, |
| { |
| "epoch": 0.8492719401810311, |
| "grad_norm": 0.9337280963011981, |
| "learning_rate": 3.031126099549653e-07, |
| "loss": 0.1119, |
| "step": 1079 |
| }, |
| { |
| "epoch": 0.8500590318772137, |
| "grad_norm": 0.9417103223229273, |
| "learning_rate": 3.000118624543888e-07, |
| "loss": 0.1117, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.8508461235733963, |
| "grad_norm": 0.9259525236801444, |
| "learning_rate": 2.9692604376492935e-07, |
| "loss": 0.1067, |
| "step": 1081 |
| }, |
| { |
| "epoch": 0.8516332152695789, |
| "grad_norm": 0.98856338356212, |
| "learning_rate": 2.9385517482651974e-07, |
| "loss": 0.1218, |
| "step": 1082 |
| }, |
| { |
| "epoch": 0.8524203069657615, |
| "grad_norm": 0.9208981942587281, |
| "learning_rate": 2.907992764776471e-07, |
| "loss": 0.1078, |
| "step": 1083 |
| }, |
| { |
| "epoch": 0.8532073986619442, |
| "grad_norm": 1.0070559943311361, |
| "learning_rate": 2.877583694552083e-07, |
| "loss": 0.1236, |
| "step": 1084 |
| }, |
| { |
| "epoch": 0.8539944903581267, |
| "grad_norm": 1.0056266955151931, |
| "learning_rate": 2.847324743943733e-07, |
| "loss": 0.1168, |
| "step": 1085 |
| }, |
| { |
| "epoch": 0.8547815820543093, |
| "grad_norm": 0.9843297655404193, |
| "learning_rate": 2.8172161182844076e-07, |
| "loss": 0.1179, |
| "step": 1086 |
| }, |
| { |
| "epoch": 0.855568673750492, |
| "grad_norm": 0.9794092387444499, |
| "learning_rate": 2.7872580218870293e-07, |
| "loss": 0.1143, |
| "step": 1087 |
| }, |
| { |
| "epoch": 0.8563557654466746, |
| "grad_norm": 0.9193678895589191, |
| "learning_rate": 2.757450658043029e-07, |
| "loss": 0.1033, |
| "step": 1088 |
| }, |
| { |
| "epoch": 0.8571428571428571, |
| "grad_norm": 0.943695053791707, |
| "learning_rate": 2.7277942290210105e-07, |
| "loss": 0.1197, |
| "step": 1089 |
| }, |
| { |
| "epoch": 0.8579299488390397, |
| "grad_norm": 0.9996493471626344, |
| "learning_rate": 2.698288936065338e-07, |
| "loss": 0.1121, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.8587170405352224, |
| "grad_norm": 0.9457549280655658, |
| "learning_rate": 2.6689349793947993e-07, |
| "loss": 0.1135, |
| "step": 1091 |
| }, |
| { |
| "epoch": 0.859504132231405, |
| "grad_norm": 0.9370356637509019, |
| "learning_rate": 2.639732558201219e-07, |
| "loss": 0.1142, |
| "step": 1092 |
| }, |
| { |
| "epoch": 0.8602912239275876, |
| "grad_norm": 0.9355392394517238, |
| "learning_rate": 2.610681870648149e-07, |
| "loss": 0.1101, |
| "step": 1093 |
| }, |
| { |
| "epoch": 0.8610783156237701, |
| "grad_norm": 0.9395826896807081, |
| "learning_rate": 2.5817831138694685e-07, |
| "loss": 0.1143, |
| "step": 1094 |
| }, |
| { |
| "epoch": 0.8618654073199528, |
| "grad_norm": 0.9169613408255519, |
| "learning_rate": 2.553036483968094e-07, |
| "loss": 0.1125, |
| "step": 1095 |
| }, |
| { |
| "epoch": 0.8626524990161354, |
| "grad_norm": 0.9214498686393413, |
| "learning_rate": 2.5244421760146354e-07, |
| "loss": 0.1061, |
| "step": 1096 |
| }, |
| { |
| "epoch": 0.863439590712318, |
| "grad_norm": 0.9759019789855492, |
| "learning_rate": 2.496000384046046e-07, |
| "loss": 0.1134, |
| "step": 1097 |
| }, |
| { |
| "epoch": 0.8642266824085005, |
| "grad_norm": 1.002942267675624, |
| "learning_rate": 2.467711301064349e-07, |
| "loss": 0.1249, |
| "step": 1098 |
| }, |
| { |
| "epoch": 0.8650137741046832, |
| "grad_norm": 0.9774180094397517, |
| "learning_rate": 2.4395751190352924e-07, |
| "loss": 0.1192, |
| "step": 1099 |
| }, |
| { |
| "epoch": 0.8658008658008658, |
| "grad_norm": 0.9977028529540362, |
| "learning_rate": 2.411592028887058e-07, |
| "loss": 0.1189, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.8665879574970484, |
| "grad_norm": 0.9904872465257003, |
| "learning_rate": 2.383762220508984e-07, |
| "loss": 0.1183, |
| "step": 1101 |
| }, |
| { |
| "epoch": 0.867375049193231, |
| "grad_norm": 0.9148412123541501, |
| "learning_rate": 2.356085882750242e-07, |
| "loss": 0.1078, |
| "step": 1102 |
| }, |
| { |
| "epoch": 0.8681621408894136, |
| "grad_norm": 1.0117176122067204, |
| "learning_rate": 2.328563203418574e-07, |
| "loss": 0.1217, |
| "step": 1103 |
| }, |
| { |
| "epoch": 0.8689492325855962, |
| "grad_norm": 0.9872081797420905, |
| "learning_rate": 2.3011943692790389e-07, |
| "loss": 0.117, |
| "step": 1104 |
| }, |
| { |
| "epoch": 0.8697363242817788, |
| "grad_norm": 0.9476791655485511, |
| "learning_rate": 2.2739795660526948e-07, |
| "loss": 0.1157, |
| "step": 1105 |
| }, |
| { |
| "epoch": 0.8705234159779615, |
| "grad_norm": 0.9183530530163464, |
| "learning_rate": 2.246918978415394e-07, |
| "loss": 0.1108, |
| "step": 1106 |
| }, |
| { |
| "epoch": 0.8713105076741441, |
| "grad_norm": 0.9622583408924335, |
| "learning_rate": 2.2200127899964786e-07, |
| "loss": 0.1188, |
| "step": 1107 |
| }, |
| { |
| "epoch": 0.8720975993703266, |
| "grad_norm": 0.9915067004751748, |
| "learning_rate": 2.1932611833775846e-07, |
| "loss": 0.1151, |
| "step": 1108 |
| }, |
| { |
| "epoch": 0.8728846910665092, |
| "grad_norm": 0.9404810815181894, |
| "learning_rate": 2.1666643400913512e-07, |
| "loss": 0.1133, |
| "step": 1109 |
| }, |
| { |
| "epoch": 0.8736717827626919, |
| "grad_norm": 0.9750904254975121, |
| "learning_rate": 2.1402224406202377e-07, |
| "loss": 0.1187, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.8744588744588745, |
| "grad_norm": 0.942666742797311, |
| "learning_rate": 2.1139356643952667e-07, |
| "loss": 0.1133, |
| "step": 1111 |
| }, |
| { |
| "epoch": 0.875245966155057, |
| "grad_norm": 0.9261400322366565, |
| "learning_rate": 2.0878041897948121e-07, |
| "loss": 0.1095, |
| "step": 1112 |
| }, |
| { |
| "epoch": 0.8760330578512396, |
| "grad_norm": 1.0714254257987408, |
| "learning_rate": 2.0618281941434058e-07, |
| "loss": 0.1197, |
| "step": 1113 |
| }, |
| { |
| "epoch": 0.8768201495474223, |
| "grad_norm": 0.9322738584358286, |
| "learning_rate": 2.036007853710503e-07, |
| "loss": 0.114, |
| "step": 1114 |
| }, |
| { |
| "epoch": 0.8776072412436049, |
| "grad_norm": 0.9346649367642453, |
| "learning_rate": 2.0103433437093256e-07, |
| "loss": 0.1027, |
| "step": 1115 |
| }, |
| { |
| "epoch": 0.8783943329397875, |
| "grad_norm": 0.9499461297298013, |
| "learning_rate": 1.9848348382956294e-07, |
| "loss": 0.1228, |
| "step": 1116 |
| }, |
| { |
| "epoch": 0.87918142463597, |
| "grad_norm": 0.9811562591520676, |
| "learning_rate": 1.9594825105665654e-07, |
| "loss": 0.1168, |
| "step": 1117 |
| }, |
| { |
| "epoch": 0.8799685163321527, |
| "grad_norm": 0.901833893893408, |
| "learning_rate": 1.934286532559468e-07, |
| "loss": 0.0992, |
| "step": 1118 |
| }, |
| { |
| "epoch": 0.8807556080283353, |
| "grad_norm": 0.9566664879297264, |
| "learning_rate": 1.9092470752507225e-07, |
| "loss": 0.1114, |
| "step": 1119 |
| }, |
| { |
| "epoch": 0.8815426997245179, |
| "grad_norm": 0.8992036132523128, |
| "learning_rate": 1.8843643085545677e-07, |
| "loss": 0.1113, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.8823297914207006, |
| "grad_norm": 1.031563578089281, |
| "learning_rate": 1.8596384013219726e-07, |
| "loss": 0.1168, |
| "step": 1121 |
| }, |
| { |
| "epoch": 0.8831168831168831, |
| "grad_norm": 0.942026348911581, |
| "learning_rate": 1.8350695213394777e-07, |
| "loss": 0.1206, |
| "step": 1122 |
| }, |
| { |
| "epoch": 0.8839039748130657, |
| "grad_norm": 0.9439190891698341, |
| "learning_rate": 1.8106578353280585e-07, |
| "loss": 0.1138, |
| "step": 1123 |
| }, |
| { |
| "epoch": 0.8846910665092483, |
| "grad_norm": 0.9261746330624396, |
| "learning_rate": 1.7864035089419973e-07, |
| "loss": 0.1118, |
| "step": 1124 |
| }, |
| { |
| "epoch": 0.885478158205431, |
| "grad_norm": 0.9958265328548447, |
| "learning_rate": 1.7623067067677467e-07, |
| "loss": 0.1235, |
| "step": 1125 |
| }, |
| { |
| "epoch": 0.8862652499016135, |
| "grad_norm": 1.0264812939307284, |
| "learning_rate": 1.7383675923228372e-07, |
| "loss": 0.1221, |
| "step": 1126 |
| }, |
| { |
| "epoch": 0.8870523415977961, |
| "grad_norm": 1.0457105898882355, |
| "learning_rate": 1.7145863280547348e-07, |
| "loss": 0.1183, |
| "step": 1127 |
| }, |
| { |
| "epoch": 0.8878394332939787, |
| "grad_norm": 0.9616200935637597, |
| "learning_rate": 1.6909630753397716e-07, |
| "loss": 0.1055, |
| "step": 1128 |
| }, |
| { |
| "epoch": 0.8886265249901614, |
| "grad_norm": 1.0114965743550393, |
| "learning_rate": 1.6674979944820258e-07, |
| "loss": 0.1247, |
| "step": 1129 |
| }, |
| { |
| "epoch": 0.889413616686344, |
| "grad_norm": 1.0002918205099012, |
| "learning_rate": 1.644191244712251e-07, |
| "loss": 0.1245, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.8902007083825265, |
| "grad_norm": 1.0106011414793612, |
| "learning_rate": 1.621042984186777e-07, |
| "loss": 0.1222, |
| "step": 1131 |
| }, |
| { |
| "epoch": 0.8909878000787091, |
| "grad_norm": 0.9663301334412446, |
| "learning_rate": 1.598053369986463e-07, |
| "loss": 0.1194, |
| "step": 1132 |
| }, |
| { |
| "epoch": 0.8917748917748918, |
| "grad_norm": 0.9465191160189073, |
| "learning_rate": 1.5752225581155995e-07, |
| "loss": 0.1085, |
| "step": 1133 |
| }, |
| { |
| "epoch": 0.8925619834710744, |
| "grad_norm": 1.0466368501224192, |
| "learning_rate": 1.5525507035008852e-07, |
| "loss": 0.1306, |
| "step": 1134 |
| }, |
| { |
| "epoch": 0.893349075167257, |
| "grad_norm": 0.9494535695924311, |
| "learning_rate": 1.5300379599903408e-07, |
| "loss": 0.1172, |
| "step": 1135 |
| }, |
| { |
| "epoch": 0.8941361668634396, |
| "grad_norm": 0.9062893885912572, |
| "learning_rate": 1.507684480352292e-07, |
| "loss": 0.109, |
| "step": 1136 |
| }, |
| { |
| "epoch": 0.8949232585596222, |
| "grad_norm": 0.9371709714503786, |
| "learning_rate": 1.4854904162743127e-07, |
| "loss": 0.1047, |
| "step": 1137 |
| }, |
| { |
| "epoch": 0.8957103502558048, |
| "grad_norm": 1.0110155118454318, |
| "learning_rate": 1.4634559183622193e-07, |
| "loss": 0.126, |
| "step": 1138 |
| }, |
| { |
| "epoch": 0.8964974419519874, |
| "grad_norm": 1.0367843431150414, |
| "learning_rate": 1.4415811361390142e-07, |
| "loss": 0.1251, |
| "step": 1139 |
| }, |
| { |
| "epoch": 0.89728453364817, |
| "grad_norm": 0.9836901063346848, |
| "learning_rate": 1.4198662180439166e-07, |
| "loss": 0.1225, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.8980716253443526, |
| "grad_norm": 1.0729047805019176, |
| "learning_rate": 1.3983113114313078e-07, |
| "loss": 0.1321, |
| "step": 1141 |
| }, |
| { |
| "epoch": 0.8988587170405352, |
| "grad_norm": 0.9172308694693998, |
| "learning_rate": 1.3769165625697633e-07, |
| "loss": 0.1094, |
| "step": 1142 |
| }, |
| { |
| "epoch": 0.8996458087367178, |
| "grad_norm": 0.9246235655173746, |
| "learning_rate": 1.355682116641052e-07, |
| "loss": 0.1098, |
| "step": 1143 |
| }, |
| { |
| "epoch": 0.9004329004329005, |
| "grad_norm": 0.9671203173685715, |
| "learning_rate": 1.3346081177391474e-07, |
| "loss": 0.1084, |
| "step": 1144 |
| }, |
| { |
| "epoch": 0.901219992129083, |
| "grad_norm": 0.9683781341379243, |
| "learning_rate": 1.3136947088692537e-07, |
| "loss": 0.119, |
| "step": 1145 |
| }, |
| { |
| "epoch": 0.9020070838252656, |
| "grad_norm": 0.9395787949515245, |
| "learning_rate": 1.2929420319468254e-07, |
| "loss": 0.1135, |
| "step": 1146 |
| }, |
| { |
| "epoch": 0.9027941755214482, |
| "grad_norm": 0.9442679409050593, |
| "learning_rate": 1.272350227796626e-07, |
| "loss": 0.1215, |
| "step": 1147 |
| }, |
| { |
| "epoch": 0.9035812672176309, |
| "grad_norm": 0.9549776744266577, |
| "learning_rate": 1.2519194361517468e-07, |
| "loss": 0.1065, |
| "step": 1148 |
| }, |
| { |
| "epoch": 0.9043683589138135, |
| "grad_norm": 0.867426121426782, |
| "learning_rate": 1.231649795652684e-07, |
| "loss": 0.0945, |
| "step": 1149 |
| }, |
| { |
| "epoch": 0.905155450609996, |
| "grad_norm": 0.9448412697453261, |
| "learning_rate": 1.2115414438463646e-07, |
| "loss": 0.1101, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.9059425423061787, |
| "grad_norm": 0.9917169805460462, |
| "learning_rate": 1.1915945171852572e-07, |
| "loss": 0.1163, |
| "step": 1151 |
| }, |
| { |
| "epoch": 0.9067296340023613, |
| "grad_norm": 0.943816804931634, |
| "learning_rate": 1.171809151026404e-07, |
| "loss": 0.1079, |
| "step": 1152 |
| }, |
| { |
| "epoch": 0.9075167256985439, |
| "grad_norm": 0.9275143855618218, |
| "learning_rate": 1.1521854796305243e-07, |
| "loss": 0.1075, |
| "step": 1153 |
| }, |
| { |
| "epoch": 0.9083038173947264, |
| "grad_norm": 1.0135210788160407, |
| "learning_rate": 1.1327236361611066e-07, |
| "loss": 0.1267, |
| "step": 1154 |
| }, |
| { |
| "epoch": 0.9090909090909091, |
| "grad_norm": 0.918433915733085, |
| "learning_rate": 1.1134237526834901e-07, |
| "loss": 0.1134, |
| "step": 1155 |
| }, |
| { |
| "epoch": 0.9098780007870917, |
| "grad_norm": 1.0485642131970048, |
| "learning_rate": 1.0942859601639793e-07, |
| "loss": 0.1181, |
| "step": 1156 |
| }, |
| { |
| "epoch": 0.9106650924832743, |
| "grad_norm": 0.9666449795771067, |
| "learning_rate": 1.0753103884689503e-07, |
| "loss": 0.1142, |
| "step": 1157 |
| }, |
| { |
| "epoch": 0.911452184179457, |
| "grad_norm": 0.9155262976479461, |
| "learning_rate": 1.0564971663639761e-07, |
| "loss": 0.1079, |
| "step": 1158 |
| }, |
| { |
| "epoch": 0.9122392758756395, |
| "grad_norm": 0.9136247354670238, |
| "learning_rate": 1.0378464215129419e-07, |
| "loss": 0.1131, |
| "step": 1159 |
| }, |
| { |
| "epoch": 0.9130263675718221, |
| "grad_norm": 0.9434235785738653, |
| "learning_rate": 1.0193582804771868e-07, |
| "loss": 0.1126, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.9138134592680047, |
| "grad_norm": 0.9320439540121583, |
| "learning_rate": 1.0010328687146464e-07, |
| "loss": 0.1128, |
| "step": 1161 |
| }, |
| { |
| "epoch": 0.9146005509641874, |
| "grad_norm": 0.9883313764295104, |
| "learning_rate": 9.828703105789983e-08, |
| "loss": 0.1189, |
| "step": 1162 |
| }, |
| { |
| "epoch": 0.91538764266037, |
| "grad_norm": 0.962978283773575, |
| "learning_rate": 9.648707293188092e-08, |
| "loss": 0.1181, |
| "step": 1163 |
| }, |
| { |
| "epoch": 0.9161747343565525, |
| "grad_norm": 0.9154295867986278, |
| "learning_rate": 9.470342470767197e-08, |
| "loss": 0.1077, |
| "step": 1164 |
| }, |
| { |
| "epoch": 0.9169618260527351, |
| "grad_norm": 0.9190343113758656, |
| "learning_rate": 9.293609848885971e-08, |
| "loss": 0.1101, |
| "step": 1165 |
| }, |
| { |
| "epoch": 0.9177489177489178, |
| "grad_norm": 0.8935174011429071, |
| "learning_rate": 9.118510626827198e-08, |
| "loss": 0.1112, |
| "step": 1166 |
| }, |
| { |
| "epoch": 0.9185360094451004, |
| "grad_norm": 0.9304289493526803, |
| "learning_rate": 8.945045992789669e-08, |
| "loss": 0.1037, |
| "step": 1167 |
| }, |
| { |
| "epoch": 0.9193231011412829, |
| "grad_norm": 0.9708158046423621, |
| "learning_rate": 8.773217123880074e-08, |
| "loss": 0.1255, |
| "step": 1168 |
| }, |
| { |
| "epoch": 0.9201101928374655, |
| "grad_norm": 0.95525649253936, |
| "learning_rate": 8.603025186105064e-08, |
| "loss": 0.1119, |
| "step": 1169 |
| }, |
| { |
| "epoch": 0.9208972845336482, |
| "grad_norm": 1.0094573892269945, |
| "learning_rate": 8.434471334363204e-08, |
| "loss": 0.1201, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.9216843762298308, |
| "grad_norm": 0.9367157782024292, |
| "learning_rate": 8.267556712437342e-08, |
| "loss": 0.1057, |
| "step": 1171 |
| }, |
| { |
| "epoch": 0.9224714679260134, |
| "grad_norm": 0.9747763894177717, |
| "learning_rate": 8.102282452986693e-08, |
| "loss": 0.1098, |
| "step": 1172 |
| }, |
| { |
| "epoch": 0.9232585596221959, |
| "grad_norm": 0.9372662016679384, |
| "learning_rate": 7.938649677539268e-08, |
| "loss": 0.1081, |
| "step": 1173 |
| }, |
| { |
| "epoch": 0.9240456513183786, |
| "grad_norm": 0.9887803764792047, |
| "learning_rate": 7.77665949648404e-08, |
| "loss": 0.1199, |
| "step": 1174 |
| }, |
| { |
| "epoch": 0.9248327430145612, |
| "grad_norm": 0.948252366615409, |
| "learning_rate": 7.616313009063791e-08, |
| "loss": 0.1064, |
| "step": 1175 |
| }, |
| { |
| "epoch": 0.9256198347107438, |
| "grad_norm": 0.948272373380358, |
| "learning_rate": 7.457611303367196e-08, |
| "loss": 0.1153, |
| "step": 1176 |
| }, |
| { |
| "epoch": 0.9264069264069265, |
| "grad_norm": 0.951666041729817, |
| "learning_rate": 7.300555456321884e-08, |
| "loss": 0.1175, |
| "step": 1177 |
| }, |
| { |
| "epoch": 0.927194018103109, |
| "grad_norm": 0.982566934759444, |
| "learning_rate": 7.145146533686725e-08, |
| "loss": 0.1164, |
| "step": 1178 |
| }, |
| { |
| "epoch": 0.9279811097992916, |
| "grad_norm": 0.9553483812186222, |
| "learning_rate": 6.991385590044947e-08, |
| "loss": 0.1169, |
| "step": 1179 |
| }, |
| { |
| "epoch": 0.9287682014954742, |
| "grad_norm": 0.9724385774768447, |
| "learning_rate": 6.839273668796747e-08, |
| "loss": 0.1078, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.9295552931916569, |
| "grad_norm": 0.9324047746069145, |
| "learning_rate": 6.688811802152279e-08, |
| "loss": 0.1162, |
| "step": 1181 |
| }, |
| { |
| "epoch": 0.9303423848878394, |
| "grad_norm": 0.9711188611376046, |
| "learning_rate": 6.540001011124703e-08, |
| "loss": 0.1089, |
| "step": 1182 |
| }, |
| { |
| "epoch": 0.931129476584022, |
| "grad_norm": 1.0007682860058293, |
| "learning_rate": 6.392842305523172e-08, |
| "loss": 0.1225, |
| "step": 1183 |
| }, |
| { |
| "epoch": 0.9319165682802046, |
| "grad_norm": 0.9074164360304593, |
| "learning_rate": 6.247336683946031e-08, |
| "loss": 0.1086, |
| "step": 1184 |
| }, |
| { |
| "epoch": 0.9327036599763873, |
| "grad_norm": 0.9132051814101239, |
| "learning_rate": 6.103485133774039e-08, |
| "loss": 0.1168, |
| "step": 1185 |
| }, |
| { |
| "epoch": 0.9334907516725699, |
| "grad_norm": 0.9362633318018305, |
| "learning_rate": 5.961288631163687e-08, |
| "loss": 0.1162, |
| "step": 1186 |
| }, |
| { |
| "epoch": 0.9342778433687524, |
| "grad_norm": 1.0037784789548483, |
| "learning_rate": 5.820748141040444e-08, |
| "loss": 0.1246, |
| "step": 1187 |
| }, |
| { |
| "epoch": 0.935064935064935, |
| "grad_norm": 0.9109713869964553, |
| "learning_rate": 5.681864617092414e-08, |
| "loss": 0.1062, |
| "step": 1188 |
| }, |
| { |
| "epoch": 0.9358520267611177, |
| "grad_norm": 0.9353212505070928, |
| "learning_rate": 5.544639001763719e-08, |
| "loss": 0.1116, |
| "step": 1189 |
| }, |
| { |
| "epoch": 0.9366391184573003, |
| "grad_norm": 0.9763975248080838, |
| "learning_rate": 5.4090722262481463e-08, |
| "loss": 0.1183, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.9374262101534829, |
| "grad_norm": 0.9389308712439575, |
| "learning_rate": 5.2751652104828245e-08, |
| "loss": 0.1125, |
| "step": 1191 |
| }, |
| { |
| "epoch": 0.9382133018496654, |
| "grad_norm": 0.9219409357377748, |
| "learning_rate": 5.142918863141999e-08, |
| "loss": 0.1045, |
| "step": 1192 |
| }, |
| { |
| "epoch": 0.9390003935458481, |
| "grad_norm": 0.9112274236581333, |
| "learning_rate": 5.012334081630821e-08, |
| "loss": 0.1154, |
| "step": 1193 |
| }, |
| { |
| "epoch": 0.9397874852420307, |
| "grad_norm": 1.0413638466805462, |
| "learning_rate": 4.8834117520793754e-08, |
| "loss": 0.1235, |
| "step": 1194 |
| }, |
| { |
| "epoch": 0.9405745769382133, |
| "grad_norm": 0.929473983088696, |
| "learning_rate": 4.756152749336468e-08, |
| "loss": 0.1216, |
| "step": 1195 |
| }, |
| { |
| "epoch": 0.9413616686343959, |
| "grad_norm": 0.9499785152690334, |
| "learning_rate": 4.6305579369638474e-08, |
| "loss": 0.119, |
| "step": 1196 |
| }, |
| { |
| "epoch": 0.9421487603305785, |
| "grad_norm": 0.9512339302739883, |
| "learning_rate": 4.506628167230326e-08, |
| "loss": 0.1128, |
| "step": 1197 |
| }, |
| { |
| "epoch": 0.9429358520267611, |
| "grad_norm": 0.9789643064479855, |
| "learning_rate": 4.384364281105974e-08, |
| "loss": 0.1156, |
| "step": 1198 |
| }, |
| { |
| "epoch": 0.9437229437229437, |
| "grad_norm": 0.9338458352272411, |
| "learning_rate": 4.2637671082563225e-08, |
| "loss": 0.1097, |
| "step": 1199 |
| }, |
| { |
| "epoch": 0.9445100354191264, |
| "grad_norm": 0.9477658752462017, |
| "learning_rate": 4.144837467036922e-08, |
| "loss": 0.1062, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.9452971271153089, |
| "grad_norm": 0.9108658408708349, |
| "learning_rate": 4.0275761644876785e-08, |
| "loss": 0.113, |
| "step": 1201 |
| }, |
| { |
| "epoch": 0.9460842188114915, |
| "grad_norm": 0.9670323890311822, |
| "learning_rate": 3.911983996327251e-08, |
| "loss": 0.1159, |
| "step": 1202 |
| }, |
| { |
| "epoch": 0.9468713105076741, |
| "grad_norm": 0.9508320183305409, |
| "learning_rate": 3.798061746947995e-08, |
| "loss": 0.1183, |
| "step": 1203 |
| }, |
| { |
| "epoch": 0.9476584022038568, |
| "grad_norm": 0.9446280692518585, |
| "learning_rate": 3.6858101894102774e-08, |
| "loss": 0.1039, |
| "step": 1204 |
| }, |
| { |
| "epoch": 0.9484454939000394, |
| "grad_norm": 0.8915603321077954, |
| "learning_rate": 3.575230085437448e-08, |
| "loss": 0.1131, |
| "step": 1205 |
| }, |
| { |
| "epoch": 0.9492325855962219, |
| "grad_norm": 1.0049928450920236, |
| "learning_rate": 3.466322185410542e-08, |
| "loss": 0.1075, |
| "step": 1206 |
| }, |
| { |
| "epoch": 0.9500196772924046, |
| "grad_norm": 0.9377285133019151, |
| "learning_rate": 3.3590872283633944e-08, |
| "loss": 0.1047, |
| "step": 1207 |
| }, |
| { |
| "epoch": 0.9508067689885872, |
| "grad_norm": 1.0283573568735918, |
| "learning_rate": 3.253525941977309e-08, |
| "loss": 0.1207, |
| "step": 1208 |
| }, |
| { |
| "epoch": 0.9515938606847698, |
| "grad_norm": 0.8716127646526632, |
| "learning_rate": 3.1496390425764246e-08, |
| "loss": 0.1034, |
| "step": 1209 |
| }, |
| { |
| "epoch": 0.9523809523809523, |
| "grad_norm": 0.9413176172911034, |
| "learning_rate": 3.047427235122663e-08, |
| "loss": 0.1094, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.953168044077135, |
| "grad_norm": 0.9235158713031617, |
| "learning_rate": 2.9468912132110117e-08, |
| "loss": 0.1112, |
| "step": 1211 |
| }, |
| { |
| "epoch": 0.9539551357733176, |
| "grad_norm": 1.0172961607730988, |
| "learning_rate": 2.8480316590648315e-08, |
| "loss": 0.1174, |
| "step": 1212 |
| }, |
| { |
| "epoch": 0.9547422274695002, |
| "grad_norm": 0.9633734296197954, |
| "learning_rate": 2.750849243531223e-08, |
| "loss": 0.119, |
| "step": 1213 |
| }, |
| { |
| "epoch": 0.9555293191656828, |
| "grad_norm": 0.8730009976362983, |
| "learning_rate": 2.655344626076417e-08, |
| "loss": 0.1012, |
| "step": 1214 |
| }, |
| { |
| "epoch": 0.9563164108618654, |
| "grad_norm": 0.9472387317654097, |
| "learning_rate": 2.5615184547813364e-08, |
| "loss": 0.1169, |
| "step": 1215 |
| }, |
| { |
| "epoch": 0.957103502558048, |
| "grad_norm": 1.009282211648514, |
| "learning_rate": 2.4693713663372643e-08, |
| "loss": 0.1193, |
| "step": 1216 |
| }, |
| { |
| "epoch": 0.9578905942542306, |
| "grad_norm": 0.9320555090575322, |
| "learning_rate": 2.378903986041403e-08, |
| "loss": 0.1134, |
| "step": 1217 |
| }, |
| { |
| "epoch": 0.9586776859504132, |
| "grad_norm": 0.9964824551488268, |
| "learning_rate": 2.2901169277927126e-08, |
| "loss": 0.123, |
| "step": 1218 |
| }, |
| { |
| "epoch": 0.9594647776465959, |
| "grad_norm": 0.9252548740156445, |
| "learning_rate": 2.2030107940877733e-08, |
| "loss": 0.1145, |
| "step": 1219 |
| }, |
| { |
| "epoch": 0.9602518693427784, |
| "grad_norm": 1.0057993388132023, |
| "learning_rate": 2.117586176016512e-08, |
| "loss": 0.1246, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.961038961038961, |
| "grad_norm": 0.953596307625609, |
| "learning_rate": 2.0338436532584826e-08, |
| "loss": 0.1113, |
| "step": 1221 |
| }, |
| { |
| "epoch": 0.9618260527351437, |
| "grad_norm": 0.8951377954636907, |
| "learning_rate": 1.9517837940786767e-08, |
| "loss": 0.1033, |
| "step": 1222 |
| }, |
| { |
| "epoch": 0.9626131444313263, |
| "grad_norm": 1.0221216536124687, |
| "learning_rate": 1.8714071553238012e-08, |
| "loss": 0.125, |
| "step": 1223 |
| }, |
| { |
| "epoch": 0.9634002361275088, |
| "grad_norm": 0.9387870706305922, |
| "learning_rate": 1.7927142824184784e-08, |
| "loss": 0.1144, |
| "step": 1224 |
| }, |
| { |
| "epoch": 0.9641873278236914, |
| "grad_norm": 0.9321486745496109, |
| "learning_rate": 1.7157057093614704e-08, |
| "loss": 0.1094, |
| "step": 1225 |
| }, |
| { |
| "epoch": 0.9649744195198741, |
| "grad_norm": 1.0128840378759991, |
| "learning_rate": 1.6403819587221814e-08, |
| "loss": 0.1217, |
| "step": 1226 |
| }, |
| { |
| "epoch": 0.9657615112160567, |
| "grad_norm": 0.924640460993744, |
| "learning_rate": 1.5667435416370226e-08, |
| "loss": 0.1115, |
| "step": 1227 |
| }, |
| { |
| "epoch": 0.9665486029122393, |
| "grad_norm": 0.9932790569946806, |
| "learning_rate": 1.494790957805997e-08, |
| "loss": 0.115, |
| "step": 1228 |
| }, |
| { |
| "epoch": 0.9673356946084218, |
| "grad_norm": 0.9381951620042324, |
| "learning_rate": 1.4245246954892323e-08, |
| "loss": 0.1096, |
| "step": 1229 |
| }, |
| { |
| "epoch": 0.9681227863046045, |
| "grad_norm": 0.9070754098607924, |
| "learning_rate": 1.3559452315037025e-08, |
| "loss": 0.107, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.9689098780007871, |
| "grad_norm": 0.9138545833001099, |
| "learning_rate": 1.2890530312200944e-08, |
| "loss": 0.1027, |
| "step": 1231 |
| }, |
| { |
| "epoch": 0.9696969696969697, |
| "grad_norm": 0.959458803400461, |
| "learning_rate": 1.2238485485594753e-08, |
| "loss": 0.1163, |
| "step": 1232 |
| }, |
| { |
| "epoch": 0.9704840613931524, |
| "grad_norm": 0.9577614004721761, |
| "learning_rate": 1.160332225990296e-08, |
| "loss": 0.1148, |
| "step": 1233 |
| }, |
| { |
| "epoch": 0.9712711530893349, |
| "grad_norm": 0.9713054396060389, |
| "learning_rate": 1.0985044945254763e-08, |
| "loss": 0.1249, |
| "step": 1234 |
| }, |
| { |
| "epoch": 0.9720582447855175, |
| "grad_norm": 0.9727071823355634, |
| "learning_rate": 1.0383657737192964e-08, |
| "loss": 0.1188, |
| "step": 1235 |
| }, |
| { |
| "epoch": 0.9728453364817001, |
| "grad_norm": 0.9525541993565831, |
| "learning_rate": 9.79916471664677e-09, |
| "loss": 0.1129, |
| "step": 1236 |
| }, |
| { |
| "epoch": 0.9736324281778828, |
| "grad_norm": 0.9792192000262094, |
| "learning_rate": 9.231569849904309e-09, |
| "loss": 0.124, |
| "step": 1237 |
| }, |
| { |
| "epoch": 0.9744195198740653, |
| "grad_norm": 0.9711922853486228, |
| "learning_rate": 8.680876988584607e-09, |
| "loss": 0.1114, |
| "step": 1238 |
| }, |
| { |
| "epoch": 0.9752066115702479, |
| "grad_norm": 0.9623623423156289, |
| "learning_rate": 8.147089869612045e-09, |
| "loss": 0.1064, |
| "step": 1239 |
| }, |
| { |
| "epoch": 0.9759937032664305, |
| "grad_norm": 0.9583105118384698, |
| "learning_rate": 7.630212115191381e-09, |
| "loss": 0.1128, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.9767807949626132, |
| "grad_norm": 0.958416495157565, |
| "learning_rate": 7.130247232782217e-09, |
| "loss": 0.1124, |
| "step": 1241 |
| }, |
| { |
| "epoch": 0.9775678866587958, |
| "grad_norm": 0.9739671638616083, |
| "learning_rate": 6.647198615076789e-09, |
| "loss": 0.1239, |
| "step": 1242 |
| }, |
| { |
| "epoch": 0.9783549783549783, |
| "grad_norm": 0.9831503405039121, |
| "learning_rate": 6.181069539974716e-09, |
| "loss": 0.1152, |
| "step": 1243 |
| }, |
| { |
| "epoch": 0.9791420700511609, |
| "grad_norm": 0.9837400272891019, |
| "learning_rate": 5.7318631705630126e-09, |
| "loss": 0.1196, |
| "step": 1244 |
| }, |
| { |
| "epoch": 0.9799291617473436, |
| "grad_norm": 0.9663876709057511, |
| "learning_rate": 5.299582555093052e-09, |
| "loss": 0.1174, |
| "step": 1245 |
| }, |
| { |
| "epoch": 0.9807162534435262, |
| "grad_norm": 0.9077428753737183, |
| "learning_rate": 4.884230626960307e-09, |
| "loss": 0.1046, |
| "step": 1246 |
| }, |
| { |
| "epoch": 0.9815033451397088, |
| "grad_norm": 0.9624717672564354, |
| "learning_rate": 4.485810204684638e-09, |
| "loss": 0.1143, |
| "step": 1247 |
| }, |
| { |
| "epoch": 0.9822904368358913, |
| "grad_norm": 0.9768731662603329, |
| "learning_rate": 4.104323991891424e-09, |
| "loss": 0.1111, |
| "step": 1248 |
| }, |
| { |
| "epoch": 0.983077528532074, |
| "grad_norm": 0.9918364976987204, |
| "learning_rate": 3.739774577292688e-09, |
| "loss": 0.1146, |
| "step": 1249 |
| }, |
| { |
| "epoch": 0.9838646202282566, |
| "grad_norm": 0.9235361594154657, |
| "learning_rate": 3.392164434669609e-09, |
| "loss": 0.115, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.9846517119244392, |
| "grad_norm": 0.9407569711042593, |
| "learning_rate": 3.0614959228558728e-09, |
| "loss": 0.1048, |
| "step": 1251 |
| }, |
| { |
| "epoch": 0.9854388036206218, |
| "grad_norm": 0.9503971642912823, |
| "learning_rate": 2.7477712857215676e-09, |
| "loss": 0.1118, |
| "step": 1252 |
| }, |
| { |
| "epoch": 0.9862258953168044, |
| "grad_norm": 0.9423931683522886, |
| "learning_rate": 2.450992652157924e-09, |
| "loss": 0.1052, |
| "step": 1253 |
| }, |
| { |
| "epoch": 0.987012987012987, |
| "grad_norm": 0.9692730394836648, |
| "learning_rate": 2.1711620360634344e-09, |
| "loss": 0.1134, |
| "step": 1254 |
| }, |
| { |
| "epoch": 0.9878000787091696, |
| "grad_norm": 0.8578557269739953, |
| "learning_rate": 1.9082813363294205e-09, |
| "loss": 0.1062, |
| "step": 1255 |
| }, |
| { |
| "epoch": 0.9885871704053523, |
| "grad_norm": 0.9417059944579995, |
| "learning_rate": 1.662352336827544e-09, |
| "loss": 0.1198, |
| "step": 1256 |
| }, |
| { |
| "epoch": 0.9893742621015348, |
| "grad_norm": 1.0199292836158018, |
| "learning_rate": 1.4333767063973159e-09, |
| "loss": 0.1085, |
| "step": 1257 |
| }, |
| { |
| "epoch": 0.9901613537977174, |
| "grad_norm": 0.9283942300248242, |
| "learning_rate": 1.221355998835272e-09, |
| "loss": 0.1069, |
| "step": 1258 |
| }, |
| { |
| "epoch": 0.9909484454939, |
| "grad_norm": 0.9097063653552931, |
| "learning_rate": 1.0262916528841483e-09, |
| "loss": 0.1074, |
| "step": 1259 |
| }, |
| { |
| "epoch": 0.9917355371900827, |
| "grad_norm": 0.9806238353833303, |
| "learning_rate": 8.481849922237217e-10, |
| "loss": 0.1136, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.9925226288862653, |
| "grad_norm": 0.9354892517698297, |
| "learning_rate": 6.870372254602631e-10, |
| "loss": 0.1048, |
| "step": 1261 |
| }, |
| { |
| "epoch": 0.9933097205824478, |
| "grad_norm": 0.9045069809298675, |
| "learning_rate": 5.428494461201527e-10, |
| "loss": 0.109, |
| "step": 1262 |
| }, |
| { |
| "epoch": 0.9940968122786304, |
| "grad_norm": 0.9462573748322036, |
| "learning_rate": 4.156226326415547e-10, |
| "loss": 0.1156, |
| "step": 1263 |
| }, |
| { |
| "epoch": 0.9948839039748131, |
| "grad_norm": 0.9123815541723352, |
| "learning_rate": 3.0535764836747696e-10, |
| "loss": 0.1078, |
| "step": 1264 |
| }, |
| { |
| "epoch": 0.9956709956709957, |
| "grad_norm": 0.960544516493779, |
| "learning_rate": 2.1205524154105372e-10, |
| "loss": 0.1088, |
| "step": 1265 |
| }, |
| { |
| "epoch": 0.9964580873671782, |
| "grad_norm": 0.8887242195384208, |
| "learning_rate": 1.357160452988837e-10, |
| "loss": 0.1031, |
| "step": 1266 |
| }, |
| { |
| "epoch": 0.9972451790633609, |
| "grad_norm": 0.9502744942086568, |
| "learning_rate": 7.63405776685322e-11, |
| "loss": 0.1124, |
| "step": 1267 |
| }, |
| { |
| "epoch": 0.9980322707595435, |
| "grad_norm": 0.8901953170723566, |
| "learning_rate": 3.3929241563535056e-11, |
| "loss": 0.1001, |
| "step": 1268 |
| }, |
| { |
| "epoch": 0.9988193624557261, |
| "grad_norm": 0.9602428624879035, |
| "learning_rate": 8.482324780900718e-12, |
| "loss": 0.1116, |
| "step": 1269 |
| }, |
| { |
| "epoch": 0.9996064541519087, |
| "grad_norm": 0.9779040145681426, |
| "learning_rate": 0.0, |
| "loss": 0.119, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.9996064541519087, |
| "step": 1270, |
| "total_flos": 223330201436160.0, |
| "train_loss": 0.14531472616308316, |
| "train_runtime": 7100.3282, |
| "train_samples_per_second": 11.452, |
| "train_steps_per_second": 0.179 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 1270, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": false, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 223330201436160.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|