diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,65844 @@ +{ + "best_global_step": 24200, + "best_metric": 0.24659645557403564, + "best_model_checkpoint": "saves/prompt-tuning/llama-3-8b-instruct/train_cola_1744902672/checkpoint-24200", + "epoch": 83.16008316008316, + "eval_steps": 200, + "global_step": 40000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.010395010395010396, + "grad_norm": 0.13502776622772217, + "learning_rate": 0.29999999259779675, + "loss": 11.1472, + "num_input_tokens_seen": 3680, + "step": 5 + }, + { + "epoch": 0.02079002079002079, + "grad_norm": 0.06767003238201141, + "learning_rate": 0.29999996252634736, + "loss": 8.1576, + "num_input_tokens_seen": 7584, + "step": 10 + }, + { + "epoch": 0.031185031185031187, + "grad_norm": 0.0790955126285553, + "learning_rate": 0.2999999093230187, + "loss": 6.3339, + "num_input_tokens_seen": 11264, + "step": 15 + }, + { + "epoch": 0.04158004158004158, + "grad_norm": 0.07418269664049149, + "learning_rate": 0.299999832987819, + "loss": 5.6189, + "num_input_tokens_seen": 15232, + "step": 20 + }, + { + "epoch": 0.05197505197505198, + "grad_norm": 0.020311672240495682, + "learning_rate": 0.29999973352076004, + "loss": 4.9144, + "num_input_tokens_seen": 18976, + "step": 25 + }, + { + "epoch": 0.062370062370062374, + "grad_norm": 0.1048390343785286, + "learning_rate": 0.2999996109218572, + "loss": 4.2665, + "num_input_tokens_seen": 22784, + "step": 30 + }, + { + "epoch": 0.07276507276507277, + "grad_norm": 0.02579992264509201, + "learning_rate": 0.2999994651911293, + "loss": 3.5678, + "num_input_tokens_seen": 26560, + "step": 35 + }, + { + "epoch": 0.08316008316008316, + "grad_norm": 0.02307146228849888, + "learning_rate": 0.2999992963285989, + "loss": 2.935, + "num_input_tokens_seen": 30432, + "step": 40 + }, + { + "epoch": 0.09355509355509356, + "grad_norm": 0.017982393503189087, + "learning_rate": 0.29999910433429194, + "loss": 2.2972, + "num_input_tokens_seen": 34208, + "step": 45 + }, + { + "epoch": 0.10395010395010396, + "grad_norm": 0.03599481284618378, + "learning_rate": 0.29999888920823814, + "loss": 1.7707, + "num_input_tokens_seen": 37856, + "step": 50 + }, + { + "epoch": 0.11434511434511435, + "grad_norm": 0.036551497876644135, + "learning_rate": 0.29999865095047057, + "loss": 1.3339, + "num_input_tokens_seen": 41760, + "step": 55 + }, + { + "epoch": 0.12474012474012475, + "grad_norm": 0.024816714227199554, + "learning_rate": 0.29999838956102604, + "loss": 0.7955, + "num_input_tokens_seen": 45568, + "step": 60 + }, + { + "epoch": 0.13513513513513514, + "grad_norm": 0.013269827701151371, + "learning_rate": 0.29999810503994484, + "loss": 0.4969, + "num_input_tokens_seen": 49376, + "step": 65 + }, + { + "epoch": 0.14553014553014554, + "grad_norm": 0.01420884020626545, + "learning_rate": 0.29999779738727084, + "loss": 0.3892, + "num_input_tokens_seen": 53152, + "step": 70 + }, + { + "epoch": 0.15592515592515593, + "grad_norm": 0.007514605764299631, + "learning_rate": 0.29999746660305154, + "loss": 0.3384, + "num_input_tokens_seen": 57024, + "step": 75 + }, + { + "epoch": 0.16632016632016633, + "grad_norm": 0.01797129213809967, + "learning_rate": 0.2999971126873379, + "loss": 0.2932, + "num_input_tokens_seen": 60928, + "step": 80 + }, + { + "epoch": 0.17671517671517672, + "grad_norm": 0.09406391531229019, + "learning_rate": 0.2999967356401845, + "loss": 0.3492, + "num_input_tokens_seen": 64672, + "step": 85 + }, + { + "epoch": 0.18711018711018712, + "grad_norm": 0.051161929965019226, + "learning_rate": 0.29999633546164944, + "loss": 2.7805, + "num_input_tokens_seen": 68416, + "step": 90 + }, + { + "epoch": 0.19750519750519752, + "grad_norm": 0.047030698508024216, + "learning_rate": 0.29999591215179444, + "loss": 2.9492, + "num_input_tokens_seen": 72288, + "step": 95 + }, + { + "epoch": 0.2079002079002079, + "grad_norm": 0.029817547649145126, + "learning_rate": 0.2999954657106849, + "loss": 1.4902, + "num_input_tokens_seen": 76224, + "step": 100 + }, + { + "epoch": 0.2182952182952183, + "grad_norm": 0.019339989870786667, + "learning_rate": 0.2999949961383896, + "loss": 0.8145, + "num_input_tokens_seen": 80032, + "step": 105 + }, + { + "epoch": 0.2286902286902287, + "grad_norm": 0.03233005106449127, + "learning_rate": 0.2999945034349809, + "loss": 0.5747, + "num_input_tokens_seen": 84000, + "step": 110 + }, + { + "epoch": 0.2390852390852391, + "grad_norm": 0.04136732965707779, + "learning_rate": 0.2999939876005348, + "loss": 0.4918, + "num_input_tokens_seen": 87840, + "step": 115 + }, + { + "epoch": 0.2494802494802495, + "grad_norm": 0.008512151427567005, + "learning_rate": 0.29999344863513094, + "loss": 0.4423, + "num_input_tokens_seen": 91680, + "step": 120 + }, + { + "epoch": 0.2598752598752599, + "grad_norm": 0.02633056789636612, + "learning_rate": 0.2999928865388523, + "loss": 0.3913, + "num_input_tokens_seen": 95648, + "step": 125 + }, + { + "epoch": 0.2702702702702703, + "grad_norm": 0.029025232419371605, + "learning_rate": 0.29999230131178567, + "loss": 0.4051, + "num_input_tokens_seen": 99392, + "step": 130 + }, + { + "epoch": 0.2806652806652807, + "grad_norm": 0.01705402508378029, + "learning_rate": 0.2999916929540212, + "loss": 0.397, + "num_input_tokens_seen": 103424, + "step": 135 + }, + { + "epoch": 0.2910602910602911, + "grad_norm": 0.03655802831053734, + "learning_rate": 0.29999106146565285, + "loss": 0.4193, + "num_input_tokens_seen": 107168, + "step": 140 + }, + { + "epoch": 0.30145530145530147, + "grad_norm": 0.021294977515935898, + "learning_rate": 0.29999040684677786, + "loss": 0.3571, + "num_input_tokens_seen": 111104, + "step": 145 + }, + { + "epoch": 0.31185031185031187, + "grad_norm": 0.007375025190412998, + "learning_rate": 0.2999897290974972, + "loss": 0.3344, + "num_input_tokens_seen": 114944, + "step": 150 + }, + { + "epoch": 0.32224532224532226, + "grad_norm": 0.009771570563316345, + "learning_rate": 0.2999890282179155, + "loss": 0.3405, + "num_input_tokens_seen": 118944, + "step": 155 + }, + { + "epoch": 0.33264033264033266, + "grad_norm": 0.033297546207904816, + "learning_rate": 0.29998830420814077, + "loss": 0.3741, + "num_input_tokens_seen": 122848, + "step": 160 + }, + { + "epoch": 0.34303534303534305, + "grad_norm": 0.027968930080533028, + "learning_rate": 0.2999875570682846, + "loss": 0.3409, + "num_input_tokens_seen": 126656, + "step": 165 + }, + { + "epoch": 0.35343035343035345, + "grad_norm": 0.006693217437714338, + "learning_rate": 0.2999867867984623, + "loss": 0.3116, + "num_input_tokens_seen": 130272, + "step": 170 + }, + { + "epoch": 0.36382536382536385, + "grad_norm": 0.025718048214912415, + "learning_rate": 0.29998599339879267, + "loss": 0.3441, + "num_input_tokens_seen": 134144, + "step": 175 + }, + { + "epoch": 0.37422037422037424, + "grad_norm": 0.02536095678806305, + "learning_rate": 0.29998517686939796, + "loss": 0.3092, + "num_input_tokens_seen": 137984, + "step": 180 + }, + { + "epoch": 0.38461538461538464, + "grad_norm": 0.02042613923549652, + "learning_rate": 0.29998433721040413, + "loss": 0.3889, + "num_input_tokens_seen": 141728, + "step": 185 + }, + { + "epoch": 0.39501039501039503, + "grad_norm": 0.03502725064754486, + "learning_rate": 0.29998347442194073, + "loss": 0.3761, + "num_input_tokens_seen": 145504, + "step": 190 + }, + { + "epoch": 0.40540540540540543, + "grad_norm": 0.010042614303529263, + "learning_rate": 0.2999825885041407, + "loss": 0.3479, + "num_input_tokens_seen": 149216, + "step": 195 + }, + { + "epoch": 0.4158004158004158, + "grad_norm": 0.004226129036396742, + "learning_rate": 0.29998167945714077, + "loss": 0.3225, + "num_input_tokens_seen": 153120, + "step": 200 + }, + { + "epoch": 0.4158004158004158, + "eval_loss": 0.30661338567733765, + "eval_runtime": 13.3817, + "eval_samples_per_second": 63.968, + "eval_steps_per_second": 15.992, + "num_input_tokens_seen": 153120, + "step": 200 + }, + { + "epoch": 0.4261954261954262, + "grad_norm": 0.029882246628403664, + "learning_rate": 0.2999807472810811, + "loss": 0.3066, + "num_input_tokens_seen": 156992, + "step": 205 + }, + { + "epoch": 0.4365904365904366, + "grad_norm": 0.01448049210011959, + "learning_rate": 0.29997979197610536, + "loss": 0.3289, + "num_input_tokens_seen": 160672, + "step": 210 + }, + { + "epoch": 0.446985446985447, + "grad_norm": 0.008233333937823772, + "learning_rate": 0.299978813542361, + "loss": 0.3489, + "num_input_tokens_seen": 164448, + "step": 215 + }, + { + "epoch": 0.4573804573804574, + "grad_norm": 0.010366841219365597, + "learning_rate": 0.2999778119799988, + "loss": 0.3181, + "num_input_tokens_seen": 168416, + "step": 220 + }, + { + "epoch": 0.4677754677754678, + "grad_norm": 0.002933696610853076, + "learning_rate": 0.29997678728917326, + "loss": 0.3125, + "num_input_tokens_seen": 172448, + "step": 225 + }, + { + "epoch": 0.4781704781704782, + "grad_norm": 0.021501148119568825, + "learning_rate": 0.2999757394700424, + "loss": 0.3142, + "num_input_tokens_seen": 176256, + "step": 230 + }, + { + "epoch": 0.4885654885654886, + "grad_norm": 0.005177576560527086, + "learning_rate": 0.29997466852276783, + "loss": 0.3367, + "num_input_tokens_seen": 179968, + "step": 235 + }, + { + "epoch": 0.498960498960499, + "grad_norm": 0.020754391327500343, + "learning_rate": 0.29997357444751466, + "loss": 0.328, + "num_input_tokens_seen": 183808, + "step": 240 + }, + { + "epoch": 0.5093555093555093, + "grad_norm": 0.0068879518657922745, + "learning_rate": 0.2999724572444516, + "loss": 0.3212, + "num_input_tokens_seen": 187648, + "step": 245 + }, + { + "epoch": 0.5197505197505198, + "grad_norm": 0.020043136551976204, + "learning_rate": 0.29997131691375095, + "loss": 0.3052, + "num_input_tokens_seen": 191488, + "step": 250 + }, + { + "epoch": 0.5301455301455301, + "grad_norm": 0.013065611943602562, + "learning_rate": 0.2999701534555886, + "loss": 0.3754, + "num_input_tokens_seen": 195328, + "step": 255 + }, + { + "epoch": 0.5405405405405406, + "grad_norm": 0.03995097428560257, + "learning_rate": 0.2999689668701439, + "loss": 0.4291, + "num_input_tokens_seen": 199008, + "step": 260 + }, + { + "epoch": 0.5509355509355509, + "grad_norm": 0.028009936213493347, + "learning_rate": 0.29996775715759993, + "loss": 0.4939, + "num_input_tokens_seen": 202848, + "step": 265 + }, + { + "epoch": 0.5613305613305614, + "grad_norm": 0.01900005154311657, + "learning_rate": 0.2999665243181432, + "loss": 0.4212, + "num_input_tokens_seen": 206816, + "step": 270 + }, + { + "epoch": 0.5717255717255717, + "grad_norm": 0.02086745575070381, + "learning_rate": 0.2999652683519638, + "loss": 0.4318, + "num_input_tokens_seen": 210496, + "step": 275 + }, + { + "epoch": 0.5821205821205822, + "grad_norm": 0.03029743582010269, + "learning_rate": 0.29996398925925544, + "loss": 0.3949, + "num_input_tokens_seen": 214432, + "step": 280 + }, + { + "epoch": 0.5925155925155925, + "grad_norm": 0.009148284792900085, + "learning_rate": 0.2999626870402154, + "loss": 0.3509, + "num_input_tokens_seen": 218144, + "step": 285 + }, + { + "epoch": 0.6029106029106029, + "grad_norm": 0.003327825805172324, + "learning_rate": 0.29996136169504445, + "loss": 0.2992, + "num_input_tokens_seen": 221920, + "step": 290 + }, + { + "epoch": 0.6133056133056133, + "grad_norm": 0.011896105483174324, + "learning_rate": 0.29996001322394694, + "loss": 0.3451, + "num_input_tokens_seen": 225728, + "step": 295 + }, + { + "epoch": 0.6237006237006237, + "grad_norm": 0.012294461019337177, + "learning_rate": 0.29995864162713093, + "loss": 0.3668, + "num_input_tokens_seen": 229376, + "step": 300 + }, + { + "epoch": 0.6340956340956341, + "grad_norm": 0.01129826158285141, + "learning_rate": 0.2999572469048079, + "loss": 0.3196, + "num_input_tokens_seen": 233408, + "step": 305 + }, + { + "epoch": 0.6444906444906445, + "grad_norm": 0.022352736443281174, + "learning_rate": 0.29995582905719287, + "loss": 0.3337, + "num_input_tokens_seen": 237152, + "step": 310 + }, + { + "epoch": 0.6548856548856549, + "grad_norm": 0.018420910462737083, + "learning_rate": 0.2999543880845046, + "loss": 0.4146, + "num_input_tokens_seen": 241024, + "step": 315 + }, + { + "epoch": 0.6652806652806653, + "grad_norm": 0.007515036966651678, + "learning_rate": 0.2999529239869652, + "loss": 0.3348, + "num_input_tokens_seen": 244768, + "step": 320 + }, + { + "epoch": 0.6756756756756757, + "grad_norm": 0.008488037623465061, + "learning_rate": 0.2999514367648005, + "loss": 0.3123, + "num_input_tokens_seen": 248832, + "step": 325 + }, + { + "epoch": 0.6860706860706861, + "grad_norm": 0.01192617416381836, + "learning_rate": 0.29994992641823987, + "loss": 0.2993, + "num_input_tokens_seen": 252704, + "step": 330 + }, + { + "epoch": 0.6964656964656964, + "grad_norm": 0.006763458251953125, + "learning_rate": 0.29994839294751613, + "loss": 0.282, + "num_input_tokens_seen": 256576, + "step": 335 + }, + { + "epoch": 0.7068607068607069, + "grad_norm": 0.019589852541685104, + "learning_rate": 0.29994683635286584, + "loss": 0.3058, + "num_input_tokens_seen": 260384, + "step": 340 + }, + { + "epoch": 0.7172557172557172, + "grad_norm": 0.011946571990847588, + "learning_rate": 0.2999452566345291, + "loss": 0.3021, + "num_input_tokens_seen": 264224, + "step": 345 + }, + { + "epoch": 0.7276507276507277, + "grad_norm": 0.004684171173721552, + "learning_rate": 0.2999436537927494, + "loss": 0.3145, + "num_input_tokens_seen": 268032, + "step": 350 + }, + { + "epoch": 0.738045738045738, + "grad_norm": 0.008012689650058746, + "learning_rate": 0.299942027827774, + "loss": 0.3055, + "num_input_tokens_seen": 271648, + "step": 355 + }, + { + "epoch": 0.7484407484407485, + "grad_norm": 0.0197573471814394, + "learning_rate": 0.29994037873985363, + "loss": 0.3179, + "num_input_tokens_seen": 275456, + "step": 360 + }, + { + "epoch": 0.7588357588357588, + "grad_norm": 0.0026678459253162146, + "learning_rate": 0.29993870652924254, + "loss": 0.2814, + "num_input_tokens_seen": 279232, + "step": 365 + }, + { + "epoch": 0.7692307692307693, + "grad_norm": 0.0012401110725477338, + "learning_rate": 0.29993701119619876, + "loss": 0.3143, + "num_input_tokens_seen": 282880, + "step": 370 + }, + { + "epoch": 0.7796257796257796, + "grad_norm": 0.016871865838766098, + "learning_rate": 0.2999352927409835, + "loss": 0.3419, + "num_input_tokens_seen": 286560, + "step": 375 + }, + { + "epoch": 0.7900207900207901, + "grad_norm": 0.01477054599672556, + "learning_rate": 0.29993355116386194, + "loss": 0.374, + "num_input_tokens_seen": 290400, + "step": 380 + }, + { + "epoch": 0.8004158004158004, + "grad_norm": 0.018198197707533836, + "learning_rate": 0.29993178646510266, + "loss": 0.3913, + "num_input_tokens_seen": 294176, + "step": 385 + }, + { + "epoch": 0.8108108108108109, + "grad_norm": 0.012671583332121372, + "learning_rate": 0.2999299986449777, + "loss": 0.3133, + "num_input_tokens_seen": 298016, + "step": 390 + }, + { + "epoch": 0.8212058212058212, + "grad_norm": 0.011242128908634186, + "learning_rate": 0.29992818770376284, + "loss": 0.2792, + "num_input_tokens_seen": 301760, + "step": 395 + }, + { + "epoch": 0.8316008316008316, + "grad_norm": 0.007946498692035675, + "learning_rate": 0.29992635364173725, + "loss": 0.2547, + "num_input_tokens_seen": 305504, + "step": 400 + }, + { + "epoch": 0.8316008316008316, + "eval_loss": 0.2710922956466675, + "eval_runtime": 13.4137, + "eval_samples_per_second": 63.815, + "eval_steps_per_second": 15.954, + "num_input_tokens_seen": 305504, + "step": 400 + }, + { + "epoch": 0.841995841995842, + "grad_norm": 0.006660632789134979, + "learning_rate": 0.2999244964591839, + "loss": 0.3168, + "num_input_tokens_seen": 309280, + "step": 405 + }, + { + "epoch": 0.8523908523908524, + "grad_norm": 0.02225935272872448, + "learning_rate": 0.2999226161563891, + "loss": 0.2954, + "num_input_tokens_seen": 313184, + "step": 410 + }, + { + "epoch": 0.8627858627858628, + "grad_norm": 0.004138450603932142, + "learning_rate": 0.2999207127336429, + "loss": 0.3142, + "num_input_tokens_seen": 316960, + "step": 415 + }, + { + "epoch": 0.8731808731808732, + "grad_norm": 0.018559657037258148, + "learning_rate": 0.2999187861912387, + "loss": 0.3172, + "num_input_tokens_seen": 320704, + "step": 420 + }, + { + "epoch": 0.8835758835758836, + "grad_norm": 0.010357167571783066, + "learning_rate": 0.2999168365294737, + "loss": 0.3592, + "num_input_tokens_seen": 324512, + "step": 425 + }, + { + "epoch": 0.893970893970894, + "grad_norm": 0.021069606766104698, + "learning_rate": 0.29991486374864856, + "loss": 0.339, + "num_input_tokens_seen": 328448, + "step": 430 + }, + { + "epoch": 0.9043659043659044, + "grad_norm": 0.018024256452918053, + "learning_rate": 0.29991286784906745, + "loss": 0.3239, + "num_input_tokens_seen": 332192, + "step": 435 + }, + { + "epoch": 0.9147609147609148, + "grad_norm": 0.003186495741829276, + "learning_rate": 0.2999108488310382, + "loss": 0.2863, + "num_input_tokens_seen": 335936, + "step": 440 + }, + { + "epoch": 0.9251559251559252, + "grad_norm": 0.026957403868436813, + "learning_rate": 0.29990880669487213, + "loss": 0.3209, + "num_input_tokens_seen": 339744, + "step": 445 + }, + { + "epoch": 0.9355509355509356, + "grad_norm": 0.0018293571192771196, + "learning_rate": 0.29990674144088425, + "loss": 0.2633, + "num_input_tokens_seen": 343552, + "step": 450 + }, + { + "epoch": 0.9459459459459459, + "grad_norm": 0.03124442882835865, + "learning_rate": 0.299904653069393, + "loss": 0.353, + "num_input_tokens_seen": 347552, + "step": 455 + }, + { + "epoch": 0.9563409563409564, + "grad_norm": 0.01398311834782362, + "learning_rate": 0.29990254158072044, + "loss": 0.3451, + "num_input_tokens_seen": 351392, + "step": 460 + }, + { + "epoch": 0.9667359667359667, + "grad_norm": 0.006803765427321196, + "learning_rate": 0.2999004069751921, + "loss": 0.2995, + "num_input_tokens_seen": 355168, + "step": 465 + }, + { + "epoch": 0.9771309771309772, + "grad_norm": 0.010371006093919277, + "learning_rate": 0.2998982492531373, + "loss": 0.2956, + "num_input_tokens_seen": 358976, + "step": 470 + }, + { + "epoch": 0.9875259875259875, + "grad_norm": 0.015438636764883995, + "learning_rate": 0.2998960684148887, + "loss": 0.3037, + "num_input_tokens_seen": 362720, + "step": 475 + }, + { + "epoch": 0.997920997920998, + "grad_norm": 0.008887200616300106, + "learning_rate": 0.29989386446078264, + "loss": 0.2824, + "num_input_tokens_seen": 366592, + "step": 480 + }, + { + "epoch": 1.0083160083160083, + "grad_norm": 0.01895778626203537, + "learning_rate": 0.299891637391159, + "loss": 0.3007, + "num_input_tokens_seen": 370456, + "step": 485 + }, + { + "epoch": 1.0187110187110187, + "grad_norm": 0.006097553763538599, + "learning_rate": 0.2998893872063612, + "loss": 0.3314, + "num_input_tokens_seen": 374296, + "step": 490 + }, + { + "epoch": 1.0291060291060292, + "grad_norm": 0.00845389161258936, + "learning_rate": 0.2998871139067363, + "loss": 0.3113, + "num_input_tokens_seen": 378168, + "step": 495 + }, + { + "epoch": 1.0395010395010396, + "grad_norm": 0.016880623996257782, + "learning_rate": 0.2998848174926348, + "loss": 0.3365, + "num_input_tokens_seen": 381944, + "step": 500 + }, + { + "epoch": 1.04989604989605, + "grad_norm": 0.013562766835093498, + "learning_rate": 0.2998824979644109, + "loss": 0.312, + "num_input_tokens_seen": 385720, + "step": 505 + }, + { + "epoch": 1.0602910602910602, + "grad_norm": 0.0033673536963760853, + "learning_rate": 0.29988015532242224, + "loss": 0.2701, + "num_input_tokens_seen": 389688, + "step": 510 + }, + { + "epoch": 1.0706860706860706, + "grad_norm": 0.0027311728335916996, + "learning_rate": 0.29987778956703015, + "loss": 0.3352, + "num_input_tokens_seen": 393496, + "step": 515 + }, + { + "epoch": 1.0810810810810811, + "grad_norm": 0.005902850069105625, + "learning_rate": 0.2998754006985994, + "loss": 0.3422, + "num_input_tokens_seen": 397464, + "step": 520 + }, + { + "epoch": 1.0914760914760915, + "grad_norm": 0.0026898488868027925, + "learning_rate": 0.29987298871749846, + "loss": 0.2888, + "num_input_tokens_seen": 401144, + "step": 525 + }, + { + "epoch": 1.1018711018711018, + "grad_norm": 0.0024025789462029934, + "learning_rate": 0.2998705536240992, + "loss": 0.2809, + "num_input_tokens_seen": 404984, + "step": 530 + }, + { + "epoch": 1.1122661122661124, + "grad_norm": 0.004416186828166246, + "learning_rate": 0.2998680954187772, + "loss": 0.2781, + "num_input_tokens_seen": 408856, + "step": 535 + }, + { + "epoch": 1.1226611226611227, + "grad_norm": 0.0009456760017201304, + "learning_rate": 0.2998656141019115, + "loss": 0.296, + "num_input_tokens_seen": 412696, + "step": 540 + }, + { + "epoch": 1.133056133056133, + "grad_norm": 0.019509410485625267, + "learning_rate": 0.2998631096738848, + "loss": 0.34, + "num_input_tokens_seen": 416440, + "step": 545 + }, + { + "epoch": 1.1434511434511434, + "grad_norm": 0.008890431374311447, + "learning_rate": 0.29986058213508326, + "loss": 0.2898, + "num_input_tokens_seen": 420312, + "step": 550 + }, + { + "epoch": 1.1538461538461537, + "grad_norm": 0.012911866419017315, + "learning_rate": 0.29985803148589674, + "loss": 0.2893, + "num_input_tokens_seen": 424152, + "step": 555 + }, + { + "epoch": 1.1642411642411643, + "grad_norm": 0.010390352457761765, + "learning_rate": 0.2998554577267185, + "loss": 0.2787, + "num_input_tokens_seen": 428152, + "step": 560 + }, + { + "epoch": 1.1746361746361746, + "grad_norm": 0.0026202909648418427, + "learning_rate": 0.2998528608579455, + "loss": 0.2808, + "num_input_tokens_seen": 431896, + "step": 565 + }, + { + "epoch": 1.185031185031185, + "grad_norm": 0.018476203083992004, + "learning_rate": 0.2998502408799781, + "loss": 0.3248, + "num_input_tokens_seen": 435704, + "step": 570 + }, + { + "epoch": 1.1954261954261955, + "grad_norm": 0.006836924236267805, + "learning_rate": 0.2998475977932205, + "loss": 0.3075, + "num_input_tokens_seen": 439544, + "step": 575 + }, + { + "epoch": 1.2058212058212059, + "grad_norm": 0.009780202060937881, + "learning_rate": 0.29984493159808023, + "loss": 0.2988, + "num_input_tokens_seen": 443480, + "step": 580 + }, + { + "epoch": 1.2162162162162162, + "grad_norm": 0.006580342072993517, + "learning_rate": 0.29984224229496836, + "loss": 0.2871, + "num_input_tokens_seen": 447320, + "step": 585 + }, + { + "epoch": 1.2266112266112266, + "grad_norm": 0.00594101520255208, + "learning_rate": 0.2998395298842998, + "loss": 0.4131, + "num_input_tokens_seen": 451096, + "step": 590 + }, + { + "epoch": 1.237006237006237, + "grad_norm": 0.012957035563886166, + "learning_rate": 0.29983679436649263, + "loss": 0.4489, + "num_input_tokens_seen": 454872, + "step": 595 + }, + { + "epoch": 1.2474012474012475, + "grad_norm": 0.010852313600480556, + "learning_rate": 0.2998340357419689, + "loss": 0.4244, + "num_input_tokens_seen": 458648, + "step": 600 + }, + { + "epoch": 1.2474012474012475, + "eval_loss": 0.2911844551563263, + "eval_runtime": 13.4237, + "eval_samples_per_second": 63.768, + "eval_steps_per_second": 15.942, + "num_input_tokens_seen": 458648, + "step": 600 + }, + { + "epoch": 1.2577962577962578, + "grad_norm": 0.00274831079877913, + "learning_rate": 0.29983125401115385, + "loss": 0.3973, + "num_input_tokens_seen": 462360, + "step": 605 + }, + { + "epoch": 1.2681912681912682, + "grad_norm": 0.010852557606995106, + "learning_rate": 0.29982844917447654, + "loss": 0.3923, + "num_input_tokens_seen": 466008, + "step": 610 + }, + { + "epoch": 1.2785862785862787, + "grad_norm": 0.005563750863075256, + "learning_rate": 0.2998256212323695, + "loss": 0.3145, + "num_input_tokens_seen": 469816, + "step": 615 + }, + { + "epoch": 1.288981288981289, + "grad_norm": 0.009022928774356842, + "learning_rate": 0.29982277018526887, + "loss": 0.3037, + "num_input_tokens_seen": 473592, + "step": 620 + }, + { + "epoch": 1.2993762993762994, + "grad_norm": 0.0011146427132189274, + "learning_rate": 0.2998198960336143, + "loss": 0.3358, + "num_input_tokens_seen": 477496, + "step": 625 + }, + { + "epoch": 1.3097713097713097, + "grad_norm": 0.008859907276928425, + "learning_rate": 0.299816998777849, + "loss": 0.3219, + "num_input_tokens_seen": 481304, + "step": 630 + }, + { + "epoch": 1.32016632016632, + "grad_norm": 0.01543111726641655, + "learning_rate": 0.2998140784184197, + "loss": 0.3113, + "num_input_tokens_seen": 485336, + "step": 635 + }, + { + "epoch": 1.3305613305613306, + "grad_norm": 0.004169647116214037, + "learning_rate": 0.2998111349557769, + "loss": 0.3054, + "num_input_tokens_seen": 489080, + "step": 640 + }, + { + "epoch": 1.340956340956341, + "grad_norm": 0.008798204362392426, + "learning_rate": 0.29980816839037444, + "loss": 0.3016, + "num_input_tokens_seen": 492760, + "step": 645 + }, + { + "epoch": 1.3513513513513513, + "grad_norm": 0.0036174955312162638, + "learning_rate": 0.2998051787226698, + "loss": 0.3035, + "num_input_tokens_seen": 496664, + "step": 650 + }, + { + "epoch": 1.3617463617463619, + "grad_norm": 0.0030274689197540283, + "learning_rate": 0.29980216595312403, + "loss": 0.3252, + "num_input_tokens_seen": 500376, + "step": 655 + }, + { + "epoch": 1.3721413721413722, + "grad_norm": 0.007786626927554607, + "learning_rate": 0.29979913008220177, + "loss": 0.3237, + "num_input_tokens_seen": 504056, + "step": 660 + }, + { + "epoch": 1.3825363825363826, + "grad_norm": 0.006295106373727322, + "learning_rate": 0.2997960711103711, + "loss": 0.2821, + "num_input_tokens_seen": 507864, + "step": 665 + }, + { + "epoch": 1.392931392931393, + "grad_norm": 0.0015057717682793736, + "learning_rate": 0.29979298903810386, + "loss": 0.2874, + "num_input_tokens_seen": 511608, + "step": 670 + }, + { + "epoch": 1.4033264033264032, + "grad_norm": 0.0022668475285172462, + "learning_rate": 0.29978988386587524, + "loss": 0.3142, + "num_input_tokens_seen": 515416, + "step": 675 + }, + { + "epoch": 1.4137214137214138, + "grad_norm": 0.015244388952851295, + "learning_rate": 0.2997867555941642, + "loss": 0.3406, + "num_input_tokens_seen": 519192, + "step": 680 + }, + { + "epoch": 1.4241164241164241, + "grad_norm": 0.00899590365588665, + "learning_rate": 0.299783604223453, + "loss": 0.3567, + "num_input_tokens_seen": 522968, + "step": 685 + }, + { + "epoch": 1.4345114345114345, + "grad_norm": 0.002670561196282506, + "learning_rate": 0.29978042975422786, + "loss": 0.2691, + "num_input_tokens_seen": 526712, + "step": 690 + }, + { + "epoch": 1.444906444906445, + "grad_norm": 0.015272680670022964, + "learning_rate": 0.29977723218697816, + "loss": 0.3362, + "num_input_tokens_seen": 530552, + "step": 695 + }, + { + "epoch": 1.4553014553014554, + "grad_norm": 0.004360658582299948, + "learning_rate": 0.299774011522197, + "loss": 0.2537, + "num_input_tokens_seen": 534360, + "step": 700 + }, + { + "epoch": 1.4656964656964657, + "grad_norm": 0.012599441222846508, + "learning_rate": 0.29977076776038114, + "loss": 0.3003, + "num_input_tokens_seen": 538232, + "step": 705 + }, + { + "epoch": 1.476091476091476, + "grad_norm": 0.0012430218048393726, + "learning_rate": 0.2997675009020307, + "loss": 0.2616, + "num_input_tokens_seen": 542008, + "step": 710 + }, + { + "epoch": 1.4864864864864864, + "grad_norm": 0.02020493522286415, + "learning_rate": 0.2997642109476496, + "loss": 0.3375, + "num_input_tokens_seen": 545848, + "step": 715 + }, + { + "epoch": 1.496881496881497, + "grad_norm": 0.005804737564176321, + "learning_rate": 0.299760897897745, + "loss": 0.3179, + "num_input_tokens_seen": 549688, + "step": 720 + }, + { + "epoch": 1.5072765072765073, + "grad_norm": 0.008141480386257172, + "learning_rate": 0.29975756175282803, + "loss": 0.3099, + "num_input_tokens_seen": 553400, + "step": 725 + }, + { + "epoch": 1.5176715176715176, + "grad_norm": 0.012977621518075466, + "learning_rate": 0.29975420251341306, + "loss": 0.3475, + "num_input_tokens_seen": 557368, + "step": 730 + }, + { + "epoch": 1.5280665280665282, + "grad_norm": 0.0030478150583803654, + "learning_rate": 0.29975082018001814, + "loss": 0.275, + "num_input_tokens_seen": 561176, + "step": 735 + }, + { + "epoch": 1.5384615384615383, + "grad_norm": 0.0014059582026675344, + "learning_rate": 0.2997474147531648, + "loss": 0.2837, + "num_input_tokens_seen": 565048, + "step": 740 + }, + { + "epoch": 1.5488565488565489, + "grad_norm": 0.004261959809809923, + "learning_rate": 0.29974398623337833, + "loss": 0.2866, + "num_input_tokens_seen": 568728, + "step": 745 + }, + { + "epoch": 1.5592515592515592, + "grad_norm": 0.004058978520333767, + "learning_rate": 0.2997405346211873, + "loss": 0.2968, + "num_input_tokens_seen": 572536, + "step": 750 + }, + { + "epoch": 1.5696465696465696, + "grad_norm": 0.0016424914356321096, + "learning_rate": 0.2997370599171241, + "loss": 0.2722, + "num_input_tokens_seen": 576408, + "step": 755 + }, + { + "epoch": 1.5800415800415801, + "grad_norm": 0.00536632165312767, + "learning_rate": 0.2997335621217246, + "loss": 0.2821, + "num_input_tokens_seen": 580216, + "step": 760 + }, + { + "epoch": 1.5904365904365905, + "grad_norm": 0.0010197683004662395, + "learning_rate": 0.29973004123552816, + "loss": 0.3525, + "num_input_tokens_seen": 584024, + "step": 765 + }, + { + "epoch": 1.6008316008316008, + "grad_norm": 0.004009200260043144, + "learning_rate": 0.2997264972590777, + "loss": 0.3089, + "num_input_tokens_seen": 587736, + "step": 770 + }, + { + "epoch": 1.6112266112266114, + "grad_norm": 0.007063076365739107, + "learning_rate": 0.29972293019291973, + "loss": 0.3178, + "num_input_tokens_seen": 591704, + "step": 775 + }, + { + "epoch": 1.6216216216216215, + "grad_norm": 0.01016835868358612, + "learning_rate": 0.2997193400376045, + "loss": 0.3011, + "num_input_tokens_seen": 595576, + "step": 780 + }, + { + "epoch": 1.632016632016632, + "grad_norm": 0.006793530657887459, + "learning_rate": 0.2997157267936854, + "loss": 0.3123, + "num_input_tokens_seen": 599384, + "step": 785 + }, + { + "epoch": 1.6424116424116424, + "grad_norm": 0.006273724604398012, + "learning_rate": 0.2997120904617199, + "loss": 0.2827, + "num_input_tokens_seen": 603032, + "step": 790 + }, + { + "epoch": 1.6528066528066527, + "grad_norm": 0.006716644391417503, + "learning_rate": 0.29970843104226863, + "loss": 0.2628, + "num_input_tokens_seen": 606904, + "step": 795 + }, + { + "epoch": 1.6632016632016633, + "grad_norm": 0.0029990433249622583, + "learning_rate": 0.2997047485358959, + "loss": 0.36, + "num_input_tokens_seen": 610680, + "step": 800 + }, + { + "epoch": 1.6632016632016633, + "eval_loss": 0.2836262285709381, + "eval_runtime": 13.3947, + "eval_samples_per_second": 63.906, + "eval_steps_per_second": 15.976, + "num_input_tokens_seen": 610680, + "step": 800 + }, + { + "epoch": 1.6735966735966736, + "grad_norm": 0.002816583961248398, + "learning_rate": 0.2997010429431697, + "loss": 0.3065, + "num_input_tokens_seen": 614552, + "step": 805 + }, + { + "epoch": 1.683991683991684, + "grad_norm": 0.006148617714643478, + "learning_rate": 0.29969731426466134, + "loss": 0.2967, + "num_input_tokens_seen": 618328, + "step": 810 + }, + { + "epoch": 1.6943866943866945, + "grad_norm": 0.004586759489029646, + "learning_rate": 0.299693562500946, + "loss": 0.319, + "num_input_tokens_seen": 622072, + "step": 815 + }, + { + "epoch": 1.7047817047817047, + "grad_norm": 0.001133651821874082, + "learning_rate": 0.29968978765260207, + "loss": 0.2803, + "num_input_tokens_seen": 625912, + "step": 820 + }, + { + "epoch": 1.7151767151767152, + "grad_norm": 0.005019965581595898, + "learning_rate": 0.2996859897202118, + "loss": 0.2548, + "num_input_tokens_seen": 629560, + "step": 825 + }, + { + "epoch": 1.7255717255717256, + "grad_norm": 0.0042550889775156975, + "learning_rate": 0.2996821687043609, + "loss": 0.3045, + "num_input_tokens_seen": 633304, + "step": 830 + }, + { + "epoch": 1.735966735966736, + "grad_norm": 0.00268346955999732, + "learning_rate": 0.2996783246056384, + "loss": 0.2763, + "num_input_tokens_seen": 637112, + "step": 835 + }, + { + "epoch": 1.7463617463617465, + "grad_norm": 0.0021750526502728462, + "learning_rate": 0.29967445742463744, + "loss": 0.2619, + "num_input_tokens_seen": 640888, + "step": 840 + }, + { + "epoch": 1.7567567567567568, + "grad_norm": 0.0048044403083622456, + "learning_rate": 0.29967056716195417, + "loss": 0.2555, + "num_input_tokens_seen": 644632, + "step": 845 + }, + { + "epoch": 1.7671517671517671, + "grad_norm": 0.004150399472564459, + "learning_rate": 0.2996666538181885, + "loss": 0.3068, + "num_input_tokens_seen": 648440, + "step": 850 + }, + { + "epoch": 1.7775467775467777, + "grad_norm": 0.008193307556211948, + "learning_rate": 0.29966271739394407, + "loss": 0.2639, + "num_input_tokens_seen": 652088, + "step": 855 + }, + { + "epoch": 1.7879417879417878, + "grad_norm": 0.009282145649194717, + "learning_rate": 0.29965875788982776, + "loss": 0.3415, + "num_input_tokens_seen": 655992, + "step": 860 + }, + { + "epoch": 1.7983367983367984, + "grad_norm": 0.007523451466113329, + "learning_rate": 0.2996547753064503, + "loss": 0.4057, + "num_input_tokens_seen": 659800, + "step": 865 + }, + { + "epoch": 1.8087318087318087, + "grad_norm": 0.010368818417191505, + "learning_rate": 0.29965076964442583, + "loss": 0.3145, + "num_input_tokens_seen": 663672, + "step": 870 + }, + { + "epoch": 1.819126819126819, + "grad_norm": 0.011550137773156166, + "learning_rate": 0.299646740904372, + "loss": 0.3179, + "num_input_tokens_seen": 667512, + "step": 875 + }, + { + "epoch": 1.8295218295218296, + "grad_norm": 0.008157351985573769, + "learning_rate": 0.29964268908691016, + "loss": 0.3393, + "num_input_tokens_seen": 671480, + "step": 880 + }, + { + "epoch": 1.83991683991684, + "grad_norm": 0.010317876003682613, + "learning_rate": 0.29963861419266513, + "loss": 0.3286, + "num_input_tokens_seen": 675416, + "step": 885 + }, + { + "epoch": 1.8503118503118503, + "grad_norm": 0.008963325060904026, + "learning_rate": 0.29963451622226533, + "loss": 0.3084, + "num_input_tokens_seen": 679096, + "step": 890 + }, + { + "epoch": 1.8607068607068609, + "grad_norm": 0.0024737264029681683, + "learning_rate": 0.29963039517634277, + "loss": 0.2849, + "num_input_tokens_seen": 683000, + "step": 895 + }, + { + "epoch": 1.871101871101871, + "grad_norm": 0.00416366895660758, + "learning_rate": 0.2996262510555328, + "loss": 0.2938, + "num_input_tokens_seen": 686904, + "step": 900 + }, + { + "epoch": 1.8814968814968815, + "grad_norm": 0.008106127381324768, + "learning_rate": 0.2996220838604746, + "loss": 0.2921, + "num_input_tokens_seen": 690552, + "step": 905 + }, + { + "epoch": 1.8918918918918919, + "grad_norm": 0.00660976255312562, + "learning_rate": 0.29961789359181085, + "loss": 0.2735, + "num_input_tokens_seen": 694456, + "step": 910 + }, + { + "epoch": 1.9022869022869022, + "grad_norm": 0.0022041278425604105, + "learning_rate": 0.29961368025018764, + "loss": 0.2903, + "num_input_tokens_seen": 698200, + "step": 915 + }, + { + "epoch": 1.9126819126819128, + "grad_norm": 0.003316605929285288, + "learning_rate": 0.2996094438362548, + "loss": 0.2906, + "num_input_tokens_seen": 702200, + "step": 920 + }, + { + "epoch": 1.9230769230769231, + "grad_norm": 0.0021033193916082382, + "learning_rate": 0.2996051843506657, + "loss": 0.2906, + "num_input_tokens_seen": 706072, + "step": 925 + }, + { + "epoch": 1.9334719334719335, + "grad_norm": 0.0011535674566403031, + "learning_rate": 0.299600901794077, + "loss": 0.3528, + "num_input_tokens_seen": 709816, + "step": 930 + }, + { + "epoch": 1.943866943866944, + "grad_norm": 0.003972208593040705, + "learning_rate": 0.29959659616714923, + "loss": 0.3141, + "num_input_tokens_seen": 713784, + "step": 935 + }, + { + "epoch": 1.9542619542619541, + "grad_norm": 0.0016242626588791609, + "learning_rate": 0.2995922674705464, + "loss": 0.2762, + "num_input_tokens_seen": 717592, + "step": 940 + }, + { + "epoch": 1.9646569646569647, + "grad_norm": 0.0049610259011387825, + "learning_rate": 0.2995879157049361, + "loss": 0.2725, + "num_input_tokens_seen": 721400, + "step": 945 + }, + { + "epoch": 1.975051975051975, + "grad_norm": 0.011431998573243618, + "learning_rate": 0.2995835408709893, + "loss": 0.2931, + "num_input_tokens_seen": 725240, + "step": 950 + }, + { + "epoch": 1.9854469854469854, + "grad_norm": 0.0026282817125320435, + "learning_rate": 0.29957914296938076, + "loss": 0.3078, + "num_input_tokens_seen": 729176, + "step": 955 + }, + { + "epoch": 1.995841995841996, + "grad_norm": 0.0013613426126539707, + "learning_rate": 0.2995747220007886, + "loss": 0.2721, + "num_input_tokens_seen": 733080, + "step": 960 + }, + { + "epoch": 2.006237006237006, + "grad_norm": 0.0009268623543903232, + "learning_rate": 0.2995702779658947, + "loss": 0.2691, + "num_input_tokens_seen": 736776, + "step": 965 + }, + { + "epoch": 2.0166320166320166, + "grad_norm": 0.0008056171354837716, + "learning_rate": 0.29956581086538425, + "loss": 0.2651, + "num_input_tokens_seen": 740584, + "step": 970 + }, + { + "epoch": 2.027027027027027, + "grad_norm": 0.0022148482967168093, + "learning_rate": 0.2995613206999462, + "loss": 0.2726, + "num_input_tokens_seen": 744392, + "step": 975 + }, + { + "epoch": 2.0374220374220373, + "grad_norm": 0.006720210425555706, + "learning_rate": 0.29955680747027297, + "loss": 0.2949, + "num_input_tokens_seen": 748264, + "step": 980 + }, + { + "epoch": 2.047817047817048, + "grad_norm": 0.005039123352617025, + "learning_rate": 0.2995522711770607, + "loss": 0.3024, + "num_input_tokens_seen": 752296, + "step": 985 + }, + { + "epoch": 2.0582120582120584, + "grad_norm": 0.0040133618749678135, + "learning_rate": 0.2995477118210087, + "loss": 0.2464, + "num_input_tokens_seen": 756296, + "step": 990 + }, + { + "epoch": 2.0686070686070686, + "grad_norm": 0.0010765615152195096, + "learning_rate": 0.29954312940282024, + "loss": 0.2711, + "num_input_tokens_seen": 760072, + "step": 995 + }, + { + "epoch": 2.079002079002079, + "grad_norm": 0.004728620406240225, + "learning_rate": 0.29953852392320196, + "loss": 0.3203, + "num_input_tokens_seen": 763880, + "step": 1000 + }, + { + "epoch": 2.079002079002079, + "eval_loss": 0.2618466913700104, + "eval_runtime": 13.4863, + "eval_samples_per_second": 63.472, + "eval_steps_per_second": 15.868, + "num_input_tokens_seen": 763880, + "step": 1000 + }, + { + "epoch": 2.0893970893970892, + "grad_norm": 0.0026711730752140284, + "learning_rate": 0.2995338953828641, + "loss": 0.2836, + "num_input_tokens_seen": 767656, + "step": 1005 + }, + { + "epoch": 2.0997920997921, + "grad_norm": 0.0010943195084109902, + "learning_rate": 0.2995292437825204, + "loss": 0.2887, + "num_input_tokens_seen": 771464, + "step": 1010 + }, + { + "epoch": 2.1101871101871104, + "grad_norm": 0.003534775460138917, + "learning_rate": 0.29952456912288816, + "loss": 0.274, + "num_input_tokens_seen": 775304, + "step": 1015 + }, + { + "epoch": 2.1205821205821205, + "grad_norm": 0.0011056277435272932, + "learning_rate": 0.2995198714046884, + "loss": 0.2687, + "num_input_tokens_seen": 779240, + "step": 1020 + }, + { + "epoch": 2.130977130977131, + "grad_norm": 0.005262910388410091, + "learning_rate": 0.2995151506286454, + "loss": 0.2717, + "num_input_tokens_seen": 783080, + "step": 1025 + }, + { + "epoch": 2.141372141372141, + "grad_norm": 0.002627658424898982, + "learning_rate": 0.2995104067954873, + "loss": 0.2692, + "num_input_tokens_seen": 786952, + "step": 1030 + }, + { + "epoch": 2.1517671517671517, + "grad_norm": 0.0009026611223816872, + "learning_rate": 0.2995056399059456, + "loss": 0.2799, + "num_input_tokens_seen": 790728, + "step": 1035 + }, + { + "epoch": 2.1621621621621623, + "grad_norm": 0.001486190245486796, + "learning_rate": 0.2995008499607554, + "loss": 0.278, + "num_input_tokens_seen": 794536, + "step": 1040 + }, + { + "epoch": 2.1725571725571724, + "grad_norm": 0.0032881435472518206, + "learning_rate": 0.2994960369606554, + "loss": 0.326, + "num_input_tokens_seen": 798472, + "step": 1045 + }, + { + "epoch": 2.182952182952183, + "grad_norm": 0.003928574733436108, + "learning_rate": 0.2994912009063878, + "loss": 0.2935, + "num_input_tokens_seen": 802440, + "step": 1050 + }, + { + "epoch": 2.1933471933471935, + "grad_norm": 0.0009759695967659354, + "learning_rate": 0.29948634179869843, + "loss": 0.2807, + "num_input_tokens_seen": 806376, + "step": 1055 + }, + { + "epoch": 2.2037422037422036, + "grad_norm": 0.0009490824886597693, + "learning_rate": 0.29948145963833656, + "loss": 0.2577, + "num_input_tokens_seen": 810184, + "step": 1060 + }, + { + "epoch": 2.214137214137214, + "grad_norm": 0.001630600425414741, + "learning_rate": 0.29947655442605514, + "loss": 0.2781, + "num_input_tokens_seen": 814056, + "step": 1065 + }, + { + "epoch": 2.2245322245322248, + "grad_norm": 0.0023589676711708307, + "learning_rate": 0.2994716261626106, + "loss": 0.2535, + "num_input_tokens_seen": 817864, + "step": 1070 + }, + { + "epoch": 2.234927234927235, + "grad_norm": 0.0018771227914839983, + "learning_rate": 0.2994666748487629, + "loss": 0.2867, + "num_input_tokens_seen": 821480, + "step": 1075 + }, + { + "epoch": 2.2453222453222454, + "grad_norm": 0.0007826806395314634, + "learning_rate": 0.2994617004852756, + "loss": 0.2733, + "num_input_tokens_seen": 825192, + "step": 1080 + }, + { + "epoch": 2.2557172557172556, + "grad_norm": 0.006928688380867243, + "learning_rate": 0.2994567030729159, + "loss": 0.2671, + "num_input_tokens_seen": 828936, + "step": 1085 + }, + { + "epoch": 2.266112266112266, + "grad_norm": 0.00867519062012434, + "learning_rate": 0.29945168261245436, + "loss": 0.3084, + "num_input_tokens_seen": 832744, + "step": 1090 + }, + { + "epoch": 2.2765072765072767, + "grad_norm": 0.001305434969253838, + "learning_rate": 0.29944663910466524, + "loss": 0.2839, + "num_input_tokens_seen": 836584, + "step": 1095 + }, + { + "epoch": 2.286902286902287, + "grad_norm": 0.0025245025753974915, + "learning_rate": 0.2994415725503263, + "loss": 0.2291, + "num_input_tokens_seen": 840456, + "step": 1100 + }, + { + "epoch": 2.2972972972972974, + "grad_norm": 0.002111408393830061, + "learning_rate": 0.29943648295021885, + "loss": 0.2861, + "num_input_tokens_seen": 844200, + "step": 1105 + }, + { + "epoch": 2.3076923076923075, + "grad_norm": 0.0007129978621378541, + "learning_rate": 0.2994313703051278, + "loss": 0.3016, + "num_input_tokens_seen": 847976, + "step": 1110 + }, + { + "epoch": 2.318087318087318, + "grad_norm": 0.0009313568007200956, + "learning_rate": 0.29942623461584156, + "loss": 0.2704, + "num_input_tokens_seen": 851944, + "step": 1115 + }, + { + "epoch": 2.3284823284823286, + "grad_norm": 0.0012056262930855155, + "learning_rate": 0.29942107588315214, + "loss": 0.266, + "num_input_tokens_seen": 855720, + "step": 1120 + }, + { + "epoch": 2.3388773388773387, + "grad_norm": 0.0022411260288208723, + "learning_rate": 0.29941589410785513, + "loss": 0.2979, + "num_input_tokens_seen": 859560, + "step": 1125 + }, + { + "epoch": 2.3492723492723493, + "grad_norm": 0.0026391763240098953, + "learning_rate": 0.29941068929074954, + "loss": 0.2738, + "num_input_tokens_seen": 863240, + "step": 1130 + }, + { + "epoch": 2.35966735966736, + "grad_norm": 0.0035182605497539043, + "learning_rate": 0.2994054614326381, + "loss": 0.2716, + "num_input_tokens_seen": 867048, + "step": 1135 + }, + { + "epoch": 2.37006237006237, + "grad_norm": 0.0032562746200710535, + "learning_rate": 0.29940021053432686, + "loss": 0.2895, + "num_input_tokens_seen": 870952, + "step": 1140 + }, + { + "epoch": 2.3804573804573805, + "grad_norm": 0.0007953010499477386, + "learning_rate": 0.29939493659662575, + "loss": 0.2655, + "num_input_tokens_seen": 874760, + "step": 1145 + }, + { + "epoch": 2.390852390852391, + "grad_norm": 0.0014599555870518088, + "learning_rate": 0.299389639620348, + "loss": 0.2714, + "num_input_tokens_seen": 878472, + "step": 1150 + }, + { + "epoch": 2.401247401247401, + "grad_norm": 0.005799624603241682, + "learning_rate": 0.29938431960631046, + "loss": 0.2764, + "num_input_tokens_seen": 882216, + "step": 1155 + }, + { + "epoch": 2.4116424116424118, + "grad_norm": 0.011294803582131863, + "learning_rate": 0.2993789765553335, + "loss": 0.2795, + "num_input_tokens_seen": 886152, + "step": 1160 + }, + { + "epoch": 2.422037422037422, + "grad_norm": 0.0027259537018835545, + "learning_rate": 0.2993736104682412, + "loss": 0.29, + "num_input_tokens_seen": 890024, + "step": 1165 + }, + { + "epoch": 2.4324324324324325, + "grad_norm": 0.013517460785806179, + "learning_rate": 0.299368221345861, + "loss": 0.3209, + "num_input_tokens_seen": 893896, + "step": 1170 + }, + { + "epoch": 2.442827442827443, + "grad_norm": 0.002101968042552471, + "learning_rate": 0.29936280918902397, + "loss": 0.232, + "num_input_tokens_seen": 897672, + "step": 1175 + }, + { + "epoch": 2.453222453222453, + "grad_norm": 0.001841456163674593, + "learning_rate": 0.2993573739985648, + "loss": 0.3339, + "num_input_tokens_seen": 901416, + "step": 1180 + }, + { + "epoch": 2.4636174636174637, + "grad_norm": 0.0020527432207018137, + "learning_rate": 0.2993519157753216, + "loss": 0.2993, + "num_input_tokens_seen": 905352, + "step": 1185 + }, + { + "epoch": 2.474012474012474, + "grad_norm": 0.0009343986166641116, + "learning_rate": 0.2993464345201361, + "loss": 0.284, + "num_input_tokens_seen": 909096, + "step": 1190 + }, + { + "epoch": 2.4844074844074844, + "grad_norm": 0.0026284432969987392, + "learning_rate": 0.2993409302338536, + "loss": 0.2879, + "num_input_tokens_seen": 913064, + "step": 1195 + }, + { + "epoch": 2.494802494802495, + "grad_norm": 0.0015324263367801905, + "learning_rate": 0.2993354029173229, + "loss": 0.2574, + "num_input_tokens_seen": 916648, + "step": 1200 + }, + { + "epoch": 2.494802494802495, + "eval_loss": 0.28690236806869507, + "eval_runtime": 13.4086, + "eval_samples_per_second": 63.84, + "eval_steps_per_second": 15.96, + "num_input_tokens_seen": 916648, + "step": 1200 + }, + { + "epoch": 2.505197505197505, + "grad_norm": 0.0026759032625705004, + "learning_rate": 0.2993298525713965, + "loss": 0.3513, + "num_input_tokens_seen": 920520, + "step": 1205 + }, + { + "epoch": 2.5155925155925156, + "grad_norm": 0.001370933372527361, + "learning_rate": 0.29932427919693017, + "loss": 0.3356, + "num_input_tokens_seen": 924456, + "step": 1210 + }, + { + "epoch": 2.525987525987526, + "grad_norm": 0.009586933068931103, + "learning_rate": 0.2993186827947834, + "loss": 0.4066, + "num_input_tokens_seen": 928168, + "step": 1215 + }, + { + "epoch": 2.5363825363825363, + "grad_norm": 0.0022334347013384104, + "learning_rate": 0.2993130633658194, + "loss": 0.2587, + "num_input_tokens_seen": 931944, + "step": 1220 + }, + { + "epoch": 2.546777546777547, + "grad_norm": 0.0017292560078203678, + "learning_rate": 0.29930742091090456, + "loss": 0.2827, + "num_input_tokens_seen": 935656, + "step": 1225 + }, + { + "epoch": 2.5571725571725574, + "grad_norm": 0.0005955182714387774, + "learning_rate": 0.29930175543090914, + "loss": 0.238, + "num_input_tokens_seen": 939528, + "step": 1230 + }, + { + "epoch": 2.5675675675675675, + "grad_norm": 0.00708902720361948, + "learning_rate": 0.2992960669267068, + "loss": 0.3014, + "num_input_tokens_seen": 943240, + "step": 1235 + }, + { + "epoch": 2.577962577962578, + "grad_norm": 0.0014608853962272406, + "learning_rate": 0.29929035539917476, + "loss": 0.2899, + "num_input_tokens_seen": 947176, + "step": 1240 + }, + { + "epoch": 2.5883575883575882, + "grad_norm": 0.0048211850225925446, + "learning_rate": 0.2992846208491938, + "loss": 0.2612, + "num_input_tokens_seen": 950984, + "step": 1245 + }, + { + "epoch": 2.598752598752599, + "grad_norm": 0.001732124132104218, + "learning_rate": 0.2992788632776483, + "loss": 0.2984, + "num_input_tokens_seen": 954792, + "step": 1250 + }, + { + "epoch": 2.609147609147609, + "grad_norm": 0.0011847848072648048, + "learning_rate": 0.29927308268542613, + "loss": 0.2928, + "num_input_tokens_seen": 958632, + "step": 1255 + }, + { + "epoch": 2.6195426195426195, + "grad_norm": 0.004880189895629883, + "learning_rate": 0.2992672790734187, + "loss": 0.2862, + "num_input_tokens_seen": 962472, + "step": 1260 + }, + { + "epoch": 2.62993762993763, + "grad_norm": 0.0031708574388176203, + "learning_rate": 0.299261452442521, + "loss": 0.2869, + "num_input_tokens_seen": 966248, + "step": 1265 + }, + { + "epoch": 2.64033264033264, + "grad_norm": 0.001979518448933959, + "learning_rate": 0.29925560279363167, + "loss": 0.2723, + "num_input_tokens_seen": 970152, + "step": 1270 + }, + { + "epoch": 2.6507276507276507, + "grad_norm": 0.006156404968351126, + "learning_rate": 0.29924973012765266, + "loss": 0.255, + "num_input_tokens_seen": 973992, + "step": 1275 + }, + { + "epoch": 2.6611226611226613, + "grad_norm": 0.003651246428489685, + "learning_rate": 0.29924383444548974, + "loss": 0.2463, + "num_input_tokens_seen": 977704, + "step": 1280 + }, + { + "epoch": 2.6715176715176714, + "grad_norm": 0.0013725819298997521, + "learning_rate": 0.299237915748052, + "loss": 0.2569, + "num_input_tokens_seen": 981384, + "step": 1285 + }, + { + "epoch": 2.681912681912682, + "grad_norm": 0.0019131888402625918, + "learning_rate": 0.2992319740362522, + "loss": 0.2642, + "num_input_tokens_seen": 985288, + "step": 1290 + }, + { + "epoch": 2.6923076923076925, + "grad_norm": 0.010615126229822636, + "learning_rate": 0.2992260093110066, + "loss": 0.2727, + "num_input_tokens_seen": 988968, + "step": 1295 + }, + { + "epoch": 2.7027027027027026, + "grad_norm": 0.0036567626520991325, + "learning_rate": 0.2992200215732352, + "loss": 0.2567, + "num_input_tokens_seen": 992648, + "step": 1300 + }, + { + "epoch": 2.713097713097713, + "grad_norm": 0.003921627998352051, + "learning_rate": 0.2992140108238611, + "loss": 0.2882, + "num_input_tokens_seen": 996456, + "step": 1305 + }, + { + "epoch": 2.7234927234927238, + "grad_norm": 0.008985362946987152, + "learning_rate": 0.2992079770638115, + "loss": 0.2483, + "num_input_tokens_seen": 1000200, + "step": 1310 + }, + { + "epoch": 2.733887733887734, + "grad_norm": 0.001910980325192213, + "learning_rate": 0.29920192029401677, + "loss": 0.2847, + "num_input_tokens_seen": 1004168, + "step": 1315 + }, + { + "epoch": 2.7442827442827444, + "grad_norm": 0.006979361176490784, + "learning_rate": 0.2991958405154109, + "loss": 0.3267, + "num_input_tokens_seen": 1007976, + "step": 1320 + }, + { + "epoch": 2.7546777546777546, + "grad_norm": 0.0023860628716647625, + "learning_rate": 0.29918973772893154, + "loss": 0.2462, + "num_input_tokens_seen": 1011784, + "step": 1325 + }, + { + "epoch": 2.765072765072765, + "grad_norm": 0.0018045055912807584, + "learning_rate": 0.29918361193551973, + "loss": 0.2794, + "num_input_tokens_seen": 1015560, + "step": 1330 + }, + { + "epoch": 2.7754677754677752, + "grad_norm": 0.004008368588984013, + "learning_rate": 0.29917746313612026, + "loss": 0.2971, + "num_input_tokens_seen": 1019176, + "step": 1335 + }, + { + "epoch": 2.785862785862786, + "grad_norm": 0.006423517595976591, + "learning_rate": 0.29917129133168124, + "loss": 0.3498, + "num_input_tokens_seen": 1023016, + "step": 1340 + }, + { + "epoch": 2.7962577962577964, + "grad_norm": 0.004084318410605192, + "learning_rate": 0.2991650965231546, + "loss": 0.3143, + "num_input_tokens_seen": 1026824, + "step": 1345 + }, + { + "epoch": 2.8066528066528065, + "grad_norm": 0.0011377106420695782, + "learning_rate": 0.29915887871149544, + "loss": 0.2945, + "num_input_tokens_seen": 1030440, + "step": 1350 + }, + { + "epoch": 2.817047817047817, + "grad_norm": 0.002099065575748682, + "learning_rate": 0.2991526378976628, + "loss": 0.2788, + "num_input_tokens_seen": 1034248, + "step": 1355 + }, + { + "epoch": 2.8274428274428276, + "grad_norm": 0.005464376416057348, + "learning_rate": 0.29914637408261896, + "loss": 0.2706, + "num_input_tokens_seen": 1038152, + "step": 1360 + }, + { + "epoch": 2.8378378378378377, + "grad_norm": 0.008425338193774223, + "learning_rate": 0.29914008726733, + "loss": 0.3003, + "num_input_tokens_seen": 1041928, + "step": 1365 + }, + { + "epoch": 2.8482328482328483, + "grad_norm": 0.0008336607133969665, + "learning_rate": 0.2991337774527653, + "loss": 0.2814, + "num_input_tokens_seen": 1045608, + "step": 1370 + }, + { + "epoch": 2.858627858627859, + "grad_norm": 0.001973898848518729, + "learning_rate": 0.2991274446398981, + "loss": 0.3095, + "num_input_tokens_seen": 1049544, + "step": 1375 + }, + { + "epoch": 2.869022869022869, + "grad_norm": 0.0015521857421845198, + "learning_rate": 0.29912108882970484, + "loss": 0.2808, + "num_input_tokens_seen": 1053416, + "step": 1380 + }, + { + "epoch": 2.8794178794178795, + "grad_norm": 0.0014273440465331078, + "learning_rate": 0.2991147100231657, + "loss": 0.276, + "num_input_tokens_seen": 1057160, + "step": 1385 + }, + { + "epoch": 2.88981288981289, + "grad_norm": 0.005998368840664625, + "learning_rate": 0.2991083082212644, + "loss": 0.2767, + "num_input_tokens_seen": 1060936, + "step": 1390 + }, + { + "epoch": 2.9002079002079, + "grad_norm": 0.0018427560571581125, + "learning_rate": 0.2991018834249881, + "loss": 0.3086, + "num_input_tokens_seen": 1064808, + "step": 1395 + }, + { + "epoch": 2.9106029106029108, + "grad_norm": 0.0029268977232277393, + "learning_rate": 0.29909543563532764, + "loss": 0.2787, + "num_input_tokens_seen": 1068552, + "step": 1400 + }, + { + "epoch": 2.9106029106029108, + "eval_loss": 0.2539924681186676, + "eval_runtime": 13.4137, + "eval_samples_per_second": 63.815, + "eval_steps_per_second": 15.954, + "num_input_tokens_seen": 1068552, + "step": 1400 + }, + { + "epoch": 2.920997920997921, + "grad_norm": 0.004619130399078131, + "learning_rate": 0.29908896485327746, + "loss": 0.2901, + "num_input_tokens_seen": 1072232, + "step": 1405 + }, + { + "epoch": 2.9313929313929314, + "grad_norm": 0.0013719034614041448, + "learning_rate": 0.29908247107983527, + "loss": 0.2775, + "num_input_tokens_seen": 1076072, + "step": 1410 + }, + { + "epoch": 2.9417879417879416, + "grad_norm": 0.008435007184743881, + "learning_rate": 0.29907595431600253, + "loss": 0.2615, + "num_input_tokens_seen": 1079880, + "step": 1415 + }, + { + "epoch": 2.952182952182952, + "grad_norm": 0.0015799006214365363, + "learning_rate": 0.29906941456278424, + "loss": 0.2552, + "num_input_tokens_seen": 1083752, + "step": 1420 + }, + { + "epoch": 2.9625779625779627, + "grad_norm": 0.0013041372876614332, + "learning_rate": 0.2990628518211889, + "loss": 0.2795, + "num_input_tokens_seen": 1087656, + "step": 1425 + }, + { + "epoch": 2.972972972972973, + "grad_norm": 0.009832659736275673, + "learning_rate": 0.2990562660922286, + "loss": 0.3622, + "num_input_tokens_seen": 1091560, + "step": 1430 + }, + { + "epoch": 2.9833679833679834, + "grad_norm": 0.010085932910442352, + "learning_rate": 0.2990496573769189, + "loss": 0.3336, + "num_input_tokens_seen": 1095336, + "step": 1435 + }, + { + "epoch": 2.993762993762994, + "grad_norm": 0.0038716529961675406, + "learning_rate": 0.29904302567627894, + "loss": 0.3155, + "num_input_tokens_seen": 1099016, + "step": 1440 + }, + { + "epoch": 3.004158004158004, + "grad_norm": 0.003527118358761072, + "learning_rate": 0.2990363709913314, + "loss": 0.2757, + "num_input_tokens_seen": 1102784, + "step": 1445 + }, + { + "epoch": 3.0145530145530146, + "grad_norm": 0.003817586461082101, + "learning_rate": 0.29902969332310264, + "loss": 0.2956, + "num_input_tokens_seen": 1106656, + "step": 1450 + }, + { + "epoch": 3.024948024948025, + "grad_norm": 0.004686552565544844, + "learning_rate": 0.2990229926726223, + "loss": 0.3025, + "num_input_tokens_seen": 1110464, + "step": 1455 + }, + { + "epoch": 3.0353430353430353, + "grad_norm": 0.006018358748406172, + "learning_rate": 0.29901626904092365, + "loss": 0.3118, + "num_input_tokens_seen": 1114400, + "step": 1460 + }, + { + "epoch": 3.045738045738046, + "grad_norm": 0.002419866854324937, + "learning_rate": 0.2990095224290438, + "loss": 0.3199, + "num_input_tokens_seen": 1118080, + "step": 1465 + }, + { + "epoch": 3.056133056133056, + "grad_norm": 0.0040710545144975185, + "learning_rate": 0.29900275283802297, + "loss": 0.3269, + "num_input_tokens_seen": 1121792, + "step": 1470 + }, + { + "epoch": 3.0665280665280665, + "grad_norm": 0.002144722267985344, + "learning_rate": 0.2989959602689051, + "loss": 0.3098, + "num_input_tokens_seen": 1125600, + "step": 1475 + }, + { + "epoch": 3.076923076923077, + "grad_norm": 0.002598235383629799, + "learning_rate": 0.2989891447227379, + "loss": 0.3198, + "num_input_tokens_seen": 1129472, + "step": 1480 + }, + { + "epoch": 3.087318087318087, + "grad_norm": 0.009648177772760391, + "learning_rate": 0.29898230620057215, + "loss": 0.293, + "num_input_tokens_seen": 1133408, + "step": 1485 + }, + { + "epoch": 3.0977130977130978, + "grad_norm": 0.005802724976092577, + "learning_rate": 0.2989754447034626, + "loss": 0.2905, + "num_input_tokens_seen": 1137280, + "step": 1490 + }, + { + "epoch": 3.108108108108108, + "grad_norm": 0.010249296203255653, + "learning_rate": 0.2989685602324673, + "loss": 0.2925, + "num_input_tokens_seen": 1141184, + "step": 1495 + }, + { + "epoch": 3.1185031185031185, + "grad_norm": 0.00676561426371336, + "learning_rate": 0.298961652788648, + "loss": 0.2932, + "num_input_tokens_seen": 1145056, + "step": 1500 + }, + { + "epoch": 3.128898128898129, + "grad_norm": 0.00040619613719172776, + "learning_rate": 0.29895472237306986, + "loss": 0.2533, + "num_input_tokens_seen": 1148832, + "step": 1505 + }, + { + "epoch": 3.139293139293139, + "grad_norm": 0.0019757540430873632, + "learning_rate": 0.29894776898680164, + "loss": 0.2765, + "num_input_tokens_seen": 1152640, + "step": 1510 + }, + { + "epoch": 3.1496881496881497, + "grad_norm": 0.0011986758327111602, + "learning_rate": 0.29894079263091566, + "loss": 0.3014, + "num_input_tokens_seen": 1156512, + "step": 1515 + }, + { + "epoch": 3.1600831600831603, + "grad_norm": 0.006642407737672329, + "learning_rate": 0.2989337933064877, + "loss": 0.2755, + "num_input_tokens_seen": 1160192, + "step": 1520 + }, + { + "epoch": 3.1704781704781704, + "grad_norm": 0.0009561611223034561, + "learning_rate": 0.29892677101459725, + "loss": 0.2558, + "num_input_tokens_seen": 1164128, + "step": 1525 + }, + { + "epoch": 3.180873180873181, + "grad_norm": 0.003825856139883399, + "learning_rate": 0.2989197257563272, + "loss": 0.2643, + "num_input_tokens_seen": 1167936, + "step": 1530 + }, + { + "epoch": 3.1912681912681915, + "grad_norm": 0.0021503549069166183, + "learning_rate": 0.2989126575327639, + "loss": 0.287, + "num_input_tokens_seen": 1171648, + "step": 1535 + }, + { + "epoch": 3.2016632016632016, + "grad_norm": 0.005350227002054453, + "learning_rate": 0.29890556634499754, + "loss": 0.282, + "num_input_tokens_seen": 1175456, + "step": 1540 + }, + { + "epoch": 3.212058212058212, + "grad_norm": 0.012126617133617401, + "learning_rate": 0.2988984521941216, + "loss": 0.3208, + "num_input_tokens_seen": 1179360, + "step": 1545 + }, + { + "epoch": 3.2224532224532223, + "grad_norm": 0.004827439319342375, + "learning_rate": 0.29889131508123307, + "loss": 0.3076, + "num_input_tokens_seen": 1183264, + "step": 1550 + }, + { + "epoch": 3.232848232848233, + "grad_norm": 0.005428845528513193, + "learning_rate": 0.2988841550074327, + "loss": 0.3239, + "num_input_tokens_seen": 1187040, + "step": 1555 + }, + { + "epoch": 3.2432432432432434, + "grad_norm": 0.0013735729735344648, + "learning_rate": 0.2988769719738246, + "loss": 0.2934, + "num_input_tokens_seen": 1190656, + "step": 1560 + }, + { + "epoch": 3.2536382536382535, + "grad_norm": 0.0024517911951988935, + "learning_rate": 0.29886976598151666, + "loss": 0.2814, + "num_input_tokens_seen": 1194400, + "step": 1565 + }, + { + "epoch": 3.264033264033264, + "grad_norm": 0.001760362065397203, + "learning_rate": 0.29886253703161986, + "loss": 0.2562, + "num_input_tokens_seen": 1198016, + "step": 1570 + }, + { + "epoch": 3.274428274428274, + "grad_norm": 0.002623186679556966, + "learning_rate": 0.29885528512524917, + "loss": 0.2422, + "num_input_tokens_seen": 1201792, + "step": 1575 + }, + { + "epoch": 3.284823284823285, + "grad_norm": 0.0008549263002350926, + "learning_rate": 0.29884801026352287, + "loss": 0.2888, + "num_input_tokens_seen": 1205728, + "step": 1580 + }, + { + "epoch": 3.2952182952182953, + "grad_norm": 0.007070057559758425, + "learning_rate": 0.2988407124475629, + "loss": 0.2951, + "num_input_tokens_seen": 1209664, + "step": 1585 + }, + { + "epoch": 3.3056133056133055, + "grad_norm": 0.0041161710396409035, + "learning_rate": 0.2988333916784945, + "loss": 0.2755, + "num_input_tokens_seen": 1213408, + "step": 1590 + }, + { + "epoch": 3.316008316008316, + "grad_norm": 0.001973289530724287, + "learning_rate": 0.2988260479574468, + "loss": 0.2776, + "num_input_tokens_seen": 1217152, + "step": 1595 + }, + { + "epoch": 3.3264033264033266, + "grad_norm": 0.0008285017102025449, + "learning_rate": 0.2988186812855523, + "loss": 0.2469, + "num_input_tokens_seen": 1220928, + "step": 1600 + }, + { + "epoch": 3.3264033264033266, + "eval_loss": 0.25966429710388184, + "eval_runtime": 13.4126, + "eval_samples_per_second": 63.821, + "eval_steps_per_second": 15.955, + "num_input_tokens_seen": 1220928, + "step": 1600 + }, + { + "epoch": 3.3367983367983367, + "grad_norm": 0.0073862276040017605, + "learning_rate": 0.29881129166394693, + "loss": 0.3005, + "num_input_tokens_seen": 1224736, + "step": 1605 + }, + { + "epoch": 3.3471933471933473, + "grad_norm": 0.001586039667017758, + "learning_rate": 0.29880387909377026, + "loss": 0.2683, + "num_input_tokens_seen": 1228512, + "step": 1610 + }, + { + "epoch": 3.357588357588358, + "grad_norm": 0.002718248637393117, + "learning_rate": 0.2987964435761655, + "loss": 0.3294, + "num_input_tokens_seen": 1232416, + "step": 1615 + }, + { + "epoch": 3.367983367983368, + "grad_norm": 0.0036075536627322435, + "learning_rate": 0.29878898511227925, + "loss": 0.2854, + "num_input_tokens_seen": 1236096, + "step": 1620 + }, + { + "epoch": 3.3783783783783785, + "grad_norm": 0.004828325472772121, + "learning_rate": 0.2987815037032617, + "loss": 0.2781, + "num_input_tokens_seen": 1239840, + "step": 1625 + }, + { + "epoch": 3.3887733887733886, + "grad_norm": 0.008463186211884022, + "learning_rate": 0.29877399935026655, + "loss": 0.2754, + "num_input_tokens_seen": 1243648, + "step": 1630 + }, + { + "epoch": 3.399168399168399, + "grad_norm": 0.002913767471909523, + "learning_rate": 0.2987664720544511, + "loss": 0.2612, + "num_input_tokens_seen": 1247584, + "step": 1635 + }, + { + "epoch": 3.4095634095634098, + "grad_norm": 0.004578461870551109, + "learning_rate": 0.2987589218169761, + "loss": 0.2937, + "num_input_tokens_seen": 1251488, + "step": 1640 + }, + { + "epoch": 3.41995841995842, + "grad_norm": 0.006026610732078552, + "learning_rate": 0.29875134863900604, + "loss": 0.2726, + "num_input_tokens_seen": 1255328, + "step": 1645 + }, + { + "epoch": 3.4303534303534304, + "grad_norm": 0.0063947043381631374, + "learning_rate": 0.29874375252170865, + "loss": 0.3089, + "num_input_tokens_seen": 1259104, + "step": 1650 + }, + { + "epoch": 3.4407484407484406, + "grad_norm": 0.006429970730096102, + "learning_rate": 0.2987361334662553, + "loss": 0.2674, + "num_input_tokens_seen": 1262848, + "step": 1655 + }, + { + "epoch": 3.451143451143451, + "grad_norm": 0.00415390869602561, + "learning_rate": 0.29872849147382113, + "loss": 0.2886, + "num_input_tokens_seen": 1266624, + "step": 1660 + }, + { + "epoch": 3.4615384615384617, + "grad_norm": 0.005241432227194309, + "learning_rate": 0.2987208265455845, + "loss": 0.2706, + "num_input_tokens_seen": 1270400, + "step": 1665 + }, + { + "epoch": 3.471933471933472, + "grad_norm": 0.0028793688397854567, + "learning_rate": 0.29871313868272753, + "loss": 0.269, + "num_input_tokens_seen": 1274176, + "step": 1670 + }, + { + "epoch": 3.4823284823284824, + "grad_norm": 0.0014138300903141499, + "learning_rate": 0.29870542788643567, + "loss": 0.2809, + "num_input_tokens_seen": 1277792, + "step": 1675 + }, + { + "epoch": 3.492723492723493, + "grad_norm": 0.008904961869120598, + "learning_rate": 0.2986976941578981, + "loss": 0.2915, + "num_input_tokens_seen": 1281600, + "step": 1680 + }, + { + "epoch": 3.503118503118503, + "grad_norm": 0.008952009491622448, + "learning_rate": 0.29868993749830747, + "loss": 0.2888, + "num_input_tokens_seen": 1285376, + "step": 1685 + }, + { + "epoch": 3.5135135135135136, + "grad_norm": 0.003298682626336813, + "learning_rate": 0.2986821579088598, + "loss": 0.2703, + "num_input_tokens_seen": 1289216, + "step": 1690 + }, + { + "epoch": 3.523908523908524, + "grad_norm": 0.0031915546860545874, + "learning_rate": 0.29867435539075504, + "loss": 0.2975, + "num_input_tokens_seen": 1293056, + "step": 1695 + }, + { + "epoch": 3.5343035343035343, + "grad_norm": 0.00769947050139308, + "learning_rate": 0.2986665299451963, + "loss": 0.284, + "num_input_tokens_seen": 1296992, + "step": 1700 + }, + { + "epoch": 3.544698544698545, + "grad_norm": 0.011347084306180477, + "learning_rate": 0.29865868157339037, + "loss": 0.3494, + "num_input_tokens_seen": 1300832, + "step": 1705 + }, + { + "epoch": 3.555093555093555, + "grad_norm": 0.009765577502548695, + "learning_rate": 0.2986508102765476, + "loss": 0.3548, + "num_input_tokens_seen": 1304704, + "step": 1710 + }, + { + "epoch": 3.5654885654885655, + "grad_norm": 0.0029405849054455757, + "learning_rate": 0.2986429160558818, + "loss": 0.294, + "num_input_tokens_seen": 1308512, + "step": 1715 + }, + { + "epoch": 3.5758835758835756, + "grad_norm": 0.001207957393489778, + "learning_rate": 0.2986349989126104, + "loss": 0.3166, + "num_input_tokens_seen": 1312320, + "step": 1720 + }, + { + "epoch": 3.586278586278586, + "grad_norm": 0.0012622280046343803, + "learning_rate": 0.29862705884795426, + "loss": 0.3154, + "num_input_tokens_seen": 1316320, + "step": 1725 + }, + { + "epoch": 3.5966735966735968, + "grad_norm": 0.0022347033955156803, + "learning_rate": 0.2986190958631379, + "loss": 0.2519, + "num_input_tokens_seen": 1320192, + "step": 1730 + }, + { + "epoch": 3.607068607068607, + "grad_norm": 0.0009959460003301501, + "learning_rate": 0.29861110995938933, + "loss": 0.2516, + "num_input_tokens_seen": 1323936, + "step": 1735 + }, + { + "epoch": 3.6174636174636174, + "grad_norm": 0.0010346631752327085, + "learning_rate": 0.29860310113794, + "loss": 0.2589, + "num_input_tokens_seen": 1327904, + "step": 1740 + }, + { + "epoch": 3.627858627858628, + "grad_norm": 0.0015725964913144708, + "learning_rate": 0.29859506940002506, + "loss": 0.2549, + "num_input_tokens_seen": 1331616, + "step": 1745 + }, + { + "epoch": 3.638253638253638, + "grad_norm": 0.0025410125963389874, + "learning_rate": 0.298587014746883, + "loss": 0.2893, + "num_input_tokens_seen": 1335488, + "step": 1750 + }, + { + "epoch": 3.6486486486486487, + "grad_norm": 0.0008813805179670453, + "learning_rate": 0.298578937179756, + "loss": 0.2843, + "num_input_tokens_seen": 1339296, + "step": 1755 + }, + { + "epoch": 3.6590436590436592, + "grad_norm": 0.0044625853188335896, + "learning_rate": 0.29857083669988976, + "loss": 0.2735, + "num_input_tokens_seen": 1343040, + "step": 1760 + }, + { + "epoch": 3.6694386694386694, + "grad_norm": 0.003523599123582244, + "learning_rate": 0.29856271330853346, + "loss": 0.3511, + "num_input_tokens_seen": 1346720, + "step": 1765 + }, + { + "epoch": 3.67983367983368, + "grad_norm": 0.004366528708487749, + "learning_rate": 0.2985545670069398, + "loss": 0.2915, + "num_input_tokens_seen": 1350720, + "step": 1770 + }, + { + "epoch": 3.6902286902286905, + "grad_norm": 0.005080975126475096, + "learning_rate": 0.29854639779636505, + "loss": 0.2513, + "num_input_tokens_seen": 1354688, + "step": 1775 + }, + { + "epoch": 3.7006237006237006, + "grad_norm": 0.002465195022523403, + "learning_rate": 0.298538205678069, + "loss": 0.2846, + "num_input_tokens_seen": 1358560, + "step": 1780 + }, + { + "epoch": 3.711018711018711, + "grad_norm": 0.005550271365791559, + "learning_rate": 0.298529990653315, + "loss": 0.2826, + "num_input_tokens_seen": 1362528, + "step": 1785 + }, + { + "epoch": 3.7214137214137213, + "grad_norm": 0.0009754779748618603, + "learning_rate": 0.29852175272336984, + "loss": 0.2833, + "num_input_tokens_seen": 1366240, + "step": 1790 + }, + { + "epoch": 3.731808731808732, + "grad_norm": 0.0019146411214023829, + "learning_rate": 0.29851349188950405, + "loss": 0.2937, + "num_input_tokens_seen": 1370080, + "step": 1795 + }, + { + "epoch": 3.742203742203742, + "grad_norm": 0.002858969150111079, + "learning_rate": 0.2985052081529914, + "loss": 0.2798, + "num_input_tokens_seen": 1373952, + "step": 1800 + }, + { + "epoch": 3.742203742203742, + "eval_loss": 0.2820360064506531, + "eval_runtime": 13.4935, + "eval_samples_per_second": 63.438, + "eval_steps_per_second": 15.859, + "num_input_tokens_seen": 1373952, + "step": 1800 + }, + { + "epoch": 3.7525987525987525, + "grad_norm": 0.0032404414378106594, + "learning_rate": 0.29849690151510944, + "loss": 0.2722, + "num_input_tokens_seen": 1377664, + "step": 1805 + }, + { + "epoch": 3.762993762993763, + "grad_norm": 0.00905431155115366, + "learning_rate": 0.2984885719771392, + "loss": 0.3257, + "num_input_tokens_seen": 1381376, + "step": 1810 + }, + { + "epoch": 3.773388773388773, + "grad_norm": 0.004402908496558666, + "learning_rate": 0.2984802195403651, + "loss": 0.2728, + "num_input_tokens_seen": 1385248, + "step": 1815 + }, + { + "epoch": 3.7837837837837838, + "grad_norm": 0.0018533398397266865, + "learning_rate": 0.2984718442060752, + "loss": 0.3062, + "num_input_tokens_seen": 1389056, + "step": 1820 + }, + { + "epoch": 3.7941787941787943, + "grad_norm": 0.0006382323917932808, + "learning_rate": 0.2984634459755611, + "loss": 0.2994, + "num_input_tokens_seen": 1392832, + "step": 1825 + }, + { + "epoch": 3.8045738045738045, + "grad_norm": 0.004639607388526201, + "learning_rate": 0.29845502485011793, + "loss": 0.3047, + "num_input_tokens_seen": 1396576, + "step": 1830 + }, + { + "epoch": 3.814968814968815, + "grad_norm": 0.007282274309545755, + "learning_rate": 0.2984465808310444, + "loss": 0.2881, + "num_input_tokens_seen": 1400384, + "step": 1835 + }, + { + "epoch": 3.8253638253638256, + "grad_norm": 0.003735373029485345, + "learning_rate": 0.29843811391964253, + "loss": 0.3207, + "num_input_tokens_seen": 1404256, + "step": 1840 + }, + { + "epoch": 3.8357588357588357, + "grad_norm": 0.00577420461922884, + "learning_rate": 0.2984296241172182, + "loss": 0.2973, + "num_input_tokens_seen": 1408064, + "step": 1845 + }, + { + "epoch": 3.8461538461538463, + "grad_norm": 0.0009359292453154922, + "learning_rate": 0.29842111142508043, + "loss": 0.2966, + "num_input_tokens_seen": 1411840, + "step": 1850 + }, + { + "epoch": 3.856548856548857, + "grad_norm": 0.0036411641631275415, + "learning_rate": 0.29841257584454217, + "loss": 0.2721, + "num_input_tokens_seen": 1415648, + "step": 1855 + }, + { + "epoch": 3.866943866943867, + "grad_norm": 0.006938981357961893, + "learning_rate": 0.29840401737691963, + "loss": 0.3088, + "num_input_tokens_seen": 1419328, + "step": 1860 + }, + { + "epoch": 3.8773388773388775, + "grad_norm": 0.0011995135573670268, + "learning_rate": 0.29839543602353263, + "loss": 0.2921, + "num_input_tokens_seen": 1423200, + "step": 1865 + }, + { + "epoch": 3.8877338877338876, + "grad_norm": 0.0016656472580507398, + "learning_rate": 0.2983868317857046, + "loss": 0.2475, + "num_input_tokens_seen": 1427072, + "step": 1870 + }, + { + "epoch": 3.898128898128898, + "grad_norm": 0.002118919976055622, + "learning_rate": 0.2983782046647623, + "loss": 0.3024, + "num_input_tokens_seen": 1430976, + "step": 1875 + }, + { + "epoch": 3.9085239085239083, + "grad_norm": 0.005580019671469927, + "learning_rate": 0.2983695546620362, + "loss": 0.2788, + "num_input_tokens_seen": 1434720, + "step": 1880 + }, + { + "epoch": 3.918918918918919, + "grad_norm": 0.0051811556331813335, + "learning_rate": 0.2983608817788603, + "loss": 0.3212, + "num_input_tokens_seen": 1438432, + "step": 1885 + }, + { + "epoch": 3.9293139293139294, + "grad_norm": 0.0028633195906877518, + "learning_rate": 0.29835218601657193, + "loss": 0.2815, + "num_input_tokens_seen": 1442272, + "step": 1890 + }, + { + "epoch": 3.9397089397089395, + "grad_norm": 0.0019312766380608082, + "learning_rate": 0.29834346737651224, + "loss": 0.2888, + "num_input_tokens_seen": 1446016, + "step": 1895 + }, + { + "epoch": 3.95010395010395, + "grad_norm": 0.0031070923432707787, + "learning_rate": 0.29833472586002563, + "loss": 0.2875, + "num_input_tokens_seen": 1449920, + "step": 1900 + }, + { + "epoch": 3.9604989604989607, + "grad_norm": 0.00412020506337285, + "learning_rate": 0.29832596146846024, + "loss": 0.2841, + "num_input_tokens_seen": 1453792, + "step": 1905 + }, + { + "epoch": 3.970893970893971, + "grad_norm": 0.0054773990996181965, + "learning_rate": 0.2983171742031676, + "loss": 0.2991, + "num_input_tokens_seen": 1457536, + "step": 1910 + }, + { + "epoch": 3.9812889812889813, + "grad_norm": 0.002289661904796958, + "learning_rate": 0.2983083640655028, + "loss": 0.2544, + "num_input_tokens_seen": 1461408, + "step": 1915 + }, + { + "epoch": 3.991683991683992, + "grad_norm": 0.005125665571540594, + "learning_rate": 0.29829953105682455, + "loss": 0.2994, + "num_input_tokens_seen": 1465376, + "step": 1920 + }, + { + "epoch": 4.002079002079002, + "grad_norm": 0.002368838293477893, + "learning_rate": 0.29829067517849495, + "loss": 0.2736, + "num_input_tokens_seen": 1469320, + "step": 1925 + }, + { + "epoch": 4.012474012474012, + "grad_norm": 0.002574007725343108, + "learning_rate": 0.2982817964318797, + "loss": 0.2693, + "num_input_tokens_seen": 1473192, + "step": 1930 + }, + { + "epoch": 4.022869022869023, + "grad_norm": 0.0022572060115635395, + "learning_rate": 0.298272894818348, + "loss": 0.2654, + "num_input_tokens_seen": 1476840, + "step": 1935 + }, + { + "epoch": 4.033264033264033, + "grad_norm": 0.00462284404784441, + "learning_rate": 0.2982639703392726, + "loss": 0.3062, + "num_input_tokens_seen": 1480584, + "step": 1940 + }, + { + "epoch": 4.043659043659043, + "grad_norm": 0.000538617605343461, + "learning_rate": 0.29825502299602974, + "loss": 0.2708, + "num_input_tokens_seen": 1484328, + "step": 1945 + }, + { + "epoch": 4.054054054054054, + "grad_norm": 0.006655883975327015, + "learning_rate": 0.2982460527899993, + "loss": 0.2983, + "num_input_tokens_seen": 1488328, + "step": 1950 + }, + { + "epoch": 4.0644490644490645, + "grad_norm": 0.0034301530104130507, + "learning_rate": 0.29823705972256453, + "loss": 0.2851, + "num_input_tokens_seen": 1492008, + "step": 1955 + }, + { + "epoch": 4.074844074844075, + "grad_norm": 0.003470534924417734, + "learning_rate": 0.2982280437951123, + "loss": 0.2993, + "num_input_tokens_seen": 1495880, + "step": 1960 + }, + { + "epoch": 4.085239085239086, + "grad_norm": 0.0010923275258392096, + "learning_rate": 0.298219005009033, + "loss": 0.2746, + "num_input_tokens_seen": 1499784, + "step": 1965 + }, + { + "epoch": 4.095634095634096, + "grad_norm": 0.0030100836884230375, + "learning_rate": 0.29820994336572043, + "loss": 0.2968, + "num_input_tokens_seen": 1503496, + "step": 1970 + }, + { + "epoch": 4.106029106029106, + "grad_norm": 0.0008962377905845642, + "learning_rate": 0.2982008588665721, + "loss": 0.2863, + "num_input_tokens_seen": 1507240, + "step": 1975 + }, + { + "epoch": 4.116424116424117, + "grad_norm": 0.0038430250715464354, + "learning_rate": 0.2981917515129889, + "loss": 0.3135, + "num_input_tokens_seen": 1511016, + "step": 1980 + }, + { + "epoch": 4.126819126819127, + "grad_norm": 0.00257163611240685, + "learning_rate": 0.2981826213063753, + "loss": 0.2918, + "num_input_tokens_seen": 1514792, + "step": 1985 + }, + { + "epoch": 4.137214137214137, + "grad_norm": 0.0024485746398568153, + "learning_rate": 0.2981734682481394, + "loss": 0.2808, + "num_input_tokens_seen": 1518600, + "step": 1990 + }, + { + "epoch": 4.147609147609147, + "grad_norm": 0.0027945362962782383, + "learning_rate": 0.29816429233969255, + "loss": 0.2654, + "num_input_tokens_seen": 1522408, + "step": 1995 + }, + { + "epoch": 4.158004158004158, + "grad_norm": 0.0020251369569450617, + "learning_rate": 0.2981550935824499, + "loss": 0.2612, + "num_input_tokens_seen": 1526312, + "step": 2000 + }, + { + "epoch": 4.158004158004158, + "eval_loss": 0.2517845630645752, + "eval_runtime": 13.4608, + "eval_samples_per_second": 63.592, + "eval_steps_per_second": 15.898, + "num_input_tokens_seen": 1526312, + "step": 2000 + }, + { + "epoch": 4.168399168399168, + "grad_norm": 0.006499717012047768, + "learning_rate": 0.29814587197783, + "loss": 0.2769, + "num_input_tokens_seen": 1530184, + "step": 2005 + }, + { + "epoch": 4.1787941787941785, + "grad_norm": 0.005391437094658613, + "learning_rate": 0.29813662752725495, + "loss": 0.2777, + "num_input_tokens_seen": 1534024, + "step": 2010 + }, + { + "epoch": 4.1891891891891895, + "grad_norm": 0.009302061051130295, + "learning_rate": 0.29812736023215025, + "loss": 0.3033, + "num_input_tokens_seen": 1537672, + "step": 2015 + }, + { + "epoch": 4.1995841995842, + "grad_norm": 0.007286230102181435, + "learning_rate": 0.29811807009394514, + "loss": 0.3335, + "num_input_tokens_seen": 1541672, + "step": 2020 + }, + { + "epoch": 4.20997920997921, + "grad_norm": 0.0037706149742007256, + "learning_rate": 0.2981087571140723, + "loss": 0.2943, + "num_input_tokens_seen": 1545416, + "step": 2025 + }, + { + "epoch": 4.220374220374221, + "grad_norm": 0.0032468908466398716, + "learning_rate": 0.2980994212939678, + "loss": 0.2753, + "num_input_tokens_seen": 1549256, + "step": 2030 + }, + { + "epoch": 4.230769230769231, + "grad_norm": 0.004540004767477512, + "learning_rate": 0.2980900626350715, + "loss": 0.2288, + "num_input_tokens_seen": 1552968, + "step": 2035 + }, + { + "epoch": 4.241164241164241, + "grad_norm": 0.0013811299577355385, + "learning_rate": 0.29808068113882646, + "loss": 0.2762, + "num_input_tokens_seen": 1556808, + "step": 2040 + }, + { + "epoch": 4.251559251559252, + "grad_norm": 0.0029561335686594248, + "learning_rate": 0.2980712768066795, + "loss": 0.3181, + "num_input_tokens_seen": 1560552, + "step": 2045 + }, + { + "epoch": 4.261954261954262, + "grad_norm": 0.0013340101577341557, + "learning_rate": 0.2980618496400809, + "loss": 0.2793, + "num_input_tokens_seen": 1564392, + "step": 2050 + }, + { + "epoch": 4.272349272349272, + "grad_norm": 0.003951576072722673, + "learning_rate": 0.2980523996404844, + "loss": 0.3006, + "num_input_tokens_seen": 1568104, + "step": 2055 + }, + { + "epoch": 4.282744282744282, + "grad_norm": 0.005777365528047085, + "learning_rate": 0.2980429268093473, + "loss": 0.2779, + "num_input_tokens_seen": 1571880, + "step": 2060 + }, + { + "epoch": 4.293139293139293, + "grad_norm": 0.0010954671306535602, + "learning_rate": 0.29803343114813047, + "loss": 0.2699, + "num_input_tokens_seen": 1575624, + "step": 2065 + }, + { + "epoch": 4.303534303534303, + "grad_norm": 0.003955200780183077, + "learning_rate": 0.2980239126582983, + "loss": 0.3168, + "num_input_tokens_seen": 1579368, + "step": 2070 + }, + { + "epoch": 4.313929313929314, + "grad_norm": 0.002212780062109232, + "learning_rate": 0.2980143713413186, + "loss": 0.3223, + "num_input_tokens_seen": 1583176, + "step": 2075 + }, + { + "epoch": 4.324324324324325, + "grad_norm": 0.004549290519207716, + "learning_rate": 0.29800480719866274, + "loss": 0.3243, + "num_input_tokens_seen": 1586952, + "step": 2080 + }, + { + "epoch": 4.334719334719335, + "grad_norm": 0.0031297740060836077, + "learning_rate": 0.2979952202318057, + "loss": 0.2759, + "num_input_tokens_seen": 1590792, + "step": 2085 + }, + { + "epoch": 4.345114345114345, + "grad_norm": 0.002063532592728734, + "learning_rate": 0.2979856104422259, + "loss": 0.2945, + "num_input_tokens_seen": 1594504, + "step": 2090 + }, + { + "epoch": 4.355509355509356, + "grad_norm": 0.004682120401412249, + "learning_rate": 0.2979759778314052, + "loss": 0.2836, + "num_input_tokens_seen": 1598280, + "step": 2095 + }, + { + "epoch": 4.365904365904366, + "grad_norm": 0.0016346347983926535, + "learning_rate": 0.2979663224008292, + "loss": 0.2672, + "num_input_tokens_seen": 1601928, + "step": 2100 + }, + { + "epoch": 4.376299376299376, + "grad_norm": 0.0013662363635376096, + "learning_rate": 0.2979566441519868, + "loss": 0.2608, + "num_input_tokens_seen": 1605832, + "step": 2105 + }, + { + "epoch": 4.386694386694387, + "grad_norm": 0.0035485406406223774, + "learning_rate": 0.29794694308637054, + "loss": 0.2626, + "num_input_tokens_seen": 1609736, + "step": 2110 + }, + { + "epoch": 4.397089397089397, + "grad_norm": 0.004933365620672703, + "learning_rate": 0.2979372192054764, + "loss": 0.2703, + "num_input_tokens_seen": 1613512, + "step": 2115 + }, + { + "epoch": 4.407484407484407, + "grad_norm": 0.0011550626950338483, + "learning_rate": 0.297927472510804, + "loss": 0.2841, + "num_input_tokens_seen": 1617320, + "step": 2120 + }, + { + "epoch": 4.417879417879418, + "grad_norm": 0.002945635002106428, + "learning_rate": 0.29791770300385634, + "loss": 0.2596, + "num_input_tokens_seen": 1621096, + "step": 2125 + }, + { + "epoch": 4.428274428274428, + "grad_norm": 0.006348751485347748, + "learning_rate": 0.29790791068614003, + "loss": 0.276, + "num_input_tokens_seen": 1624936, + "step": 2130 + }, + { + "epoch": 4.4386694386694385, + "grad_norm": 0.004368406254798174, + "learning_rate": 0.2978980955591652, + "loss": 0.2913, + "num_input_tokens_seen": 1628904, + "step": 2135 + }, + { + "epoch": 4.4490644490644495, + "grad_norm": 0.002178372349590063, + "learning_rate": 0.2978882576244454, + "loss": 0.2806, + "num_input_tokens_seen": 1632712, + "step": 2140 + }, + { + "epoch": 4.45945945945946, + "grad_norm": 0.0010187494335696101, + "learning_rate": 0.2978783968834978, + "loss": 0.2463, + "num_input_tokens_seen": 1636712, + "step": 2145 + }, + { + "epoch": 4.46985446985447, + "grad_norm": 0.003212394891306758, + "learning_rate": 0.29786851333784303, + "loss": 0.2748, + "num_input_tokens_seen": 1640552, + "step": 2150 + }, + { + "epoch": 4.48024948024948, + "grad_norm": 0.004137763287872076, + "learning_rate": 0.2978586069890053, + "loss": 0.3186, + "num_input_tokens_seen": 1644328, + "step": 2155 + }, + { + "epoch": 4.490644490644491, + "grad_norm": 0.0007519324426539242, + "learning_rate": 0.29784867783851227, + "loss": 0.2862, + "num_input_tokens_seen": 1648104, + "step": 2160 + }, + { + "epoch": 4.501039501039501, + "grad_norm": 0.004415489267557859, + "learning_rate": 0.2978387258878951, + "loss": 0.2679, + "num_input_tokens_seen": 1651944, + "step": 2165 + }, + { + "epoch": 4.511434511434511, + "grad_norm": 0.0046625323593616486, + "learning_rate": 0.29782875113868856, + "loss": 0.3172, + "num_input_tokens_seen": 1655752, + "step": 2170 + }, + { + "epoch": 4.521829521829522, + "grad_norm": 0.005025898106396198, + "learning_rate": 0.2978187535924309, + "loss": 0.316, + "num_input_tokens_seen": 1659528, + "step": 2175 + }, + { + "epoch": 4.532224532224532, + "grad_norm": 0.006388816982507706, + "learning_rate": 0.29780873325066376, + "loss": 0.2865, + "num_input_tokens_seen": 1663336, + "step": 2180 + }, + { + "epoch": 4.542619542619542, + "grad_norm": 0.004953952040523291, + "learning_rate": 0.2977986901149325, + "loss": 0.2819, + "num_input_tokens_seen": 1667048, + "step": 2185 + }, + { + "epoch": 4.553014553014553, + "grad_norm": 0.007029210217297077, + "learning_rate": 0.29778862418678587, + "loss": 0.2956, + "num_input_tokens_seen": 1670728, + "step": 2190 + }, + { + "epoch": 4.5634095634095635, + "grad_norm": 0.0015059924917295575, + "learning_rate": 0.29777853546777616, + "loss": 0.2669, + "num_input_tokens_seen": 1674408, + "step": 2195 + }, + { + "epoch": 4.573804573804574, + "grad_norm": 0.0033020901028066874, + "learning_rate": 0.2977684239594592, + "loss": 0.2863, + "num_input_tokens_seen": 1678248, + "step": 2200 + }, + { + "epoch": 4.573804573804574, + "eval_loss": 0.25246745347976685, + "eval_runtime": 13.4113, + "eval_samples_per_second": 63.827, + "eval_steps_per_second": 15.957, + "num_input_tokens_seen": 1678248, + "step": 2200 + }, + { + "epoch": 4.584199584199585, + "grad_norm": 0.0023702087346464396, + "learning_rate": 0.29775828966339424, + "loss": 0.2615, + "num_input_tokens_seen": 1682152, + "step": 2205 + }, + { + "epoch": 4.594594594594595, + "grad_norm": 0.0032688830979168415, + "learning_rate": 0.29774813258114424, + "loss": 0.2823, + "num_input_tokens_seen": 1685992, + "step": 2210 + }, + { + "epoch": 4.604989604989605, + "grad_norm": 0.003169362898916006, + "learning_rate": 0.29773795271427544, + "loss": 0.2983, + "num_input_tokens_seen": 1689672, + "step": 2215 + }, + { + "epoch": 4.615384615384615, + "grad_norm": 0.0028951934073120356, + "learning_rate": 0.2977277500643577, + "loss": 0.2678, + "num_input_tokens_seen": 1693352, + "step": 2220 + }, + { + "epoch": 4.625779625779626, + "grad_norm": 0.003787314984947443, + "learning_rate": 0.29771752463296447, + "loss": 0.3028, + "num_input_tokens_seen": 1697064, + "step": 2225 + }, + { + "epoch": 4.636174636174636, + "grad_norm": 0.002762764459475875, + "learning_rate": 0.29770727642167266, + "loss": 0.276, + "num_input_tokens_seen": 1700712, + "step": 2230 + }, + { + "epoch": 4.646569646569646, + "grad_norm": 0.005911846645176411, + "learning_rate": 0.29769700543206257, + "loss": 0.2744, + "num_input_tokens_seen": 1704648, + "step": 2235 + }, + { + "epoch": 4.656964656964657, + "grad_norm": 0.006923735607415438, + "learning_rate": 0.2976867116657182, + "loss": 0.2973, + "num_input_tokens_seen": 1708296, + "step": 2240 + }, + { + "epoch": 4.667359667359667, + "grad_norm": 0.0024055743124336004, + "learning_rate": 0.2976763951242269, + "loss": 0.293, + "num_input_tokens_seen": 1712040, + "step": 2245 + }, + { + "epoch": 4.6777546777546775, + "grad_norm": 0.0024562478065490723, + "learning_rate": 0.29766605580917965, + "loss": 0.2825, + "num_input_tokens_seen": 1715912, + "step": 2250 + }, + { + "epoch": 4.6881496881496885, + "grad_norm": 0.00522797042503953, + "learning_rate": 0.29765569372217093, + "loss": 0.2786, + "num_input_tokens_seen": 1719880, + "step": 2255 + }, + { + "epoch": 4.698544698544699, + "grad_norm": 0.008458050899207592, + "learning_rate": 0.2976453088647987, + "loss": 0.3103, + "num_input_tokens_seen": 1723720, + "step": 2260 + }, + { + "epoch": 4.708939708939709, + "grad_norm": 0.0029204199090600014, + "learning_rate": 0.2976349012386644, + "loss": 0.2633, + "num_input_tokens_seen": 1727688, + "step": 2265 + }, + { + "epoch": 4.71933471933472, + "grad_norm": 0.0006647854461334646, + "learning_rate": 0.29762447084537297, + "loss": 0.2928, + "num_input_tokens_seen": 1731656, + "step": 2270 + }, + { + "epoch": 4.72972972972973, + "grad_norm": 0.000478226545965299, + "learning_rate": 0.29761401768653306, + "loss": 0.282, + "num_input_tokens_seen": 1735464, + "step": 2275 + }, + { + "epoch": 4.74012474012474, + "grad_norm": 0.0017165506724268198, + "learning_rate": 0.29760354176375653, + "loss": 0.2863, + "num_input_tokens_seen": 1739432, + "step": 2280 + }, + { + "epoch": 4.75051975051975, + "grad_norm": 0.0007923941011540592, + "learning_rate": 0.29759304307865897, + "loss": 0.2904, + "num_input_tokens_seen": 1743272, + "step": 2285 + }, + { + "epoch": 4.760914760914761, + "grad_norm": 0.0012859933776780963, + "learning_rate": 0.2975825216328594, + "loss": 0.297, + "num_input_tokens_seen": 1747048, + "step": 2290 + }, + { + "epoch": 4.771309771309771, + "grad_norm": 0.005715373437851667, + "learning_rate": 0.2975719774279804, + "loss": 0.3093, + "num_input_tokens_seen": 1750984, + "step": 2295 + }, + { + "epoch": 4.781704781704782, + "grad_norm": 0.002364710671827197, + "learning_rate": 0.29756141046564794, + "loss": 0.2623, + "num_input_tokens_seen": 1754824, + "step": 2300 + }, + { + "epoch": 4.792099792099792, + "grad_norm": 0.0005401739617809653, + "learning_rate": 0.2975508207474916, + "loss": 0.2799, + "num_input_tokens_seen": 1758888, + "step": 2305 + }, + { + "epoch": 4.802494802494802, + "grad_norm": 0.003149572992697358, + "learning_rate": 0.2975402082751445, + "loss": 0.2817, + "num_input_tokens_seen": 1762728, + "step": 2310 + }, + { + "epoch": 4.8128898128898125, + "grad_norm": 0.0023899332154542208, + "learning_rate": 0.29752957305024313, + "loss": 0.2836, + "num_input_tokens_seen": 1766536, + "step": 2315 + }, + { + "epoch": 4.8232848232848236, + "grad_norm": 0.0008628643117845058, + "learning_rate": 0.2975189150744277, + "loss": 0.2728, + "num_input_tokens_seen": 1770440, + "step": 2320 + }, + { + "epoch": 4.833679833679834, + "grad_norm": 0.001297208247706294, + "learning_rate": 0.29750823434934165, + "loss": 0.3134, + "num_input_tokens_seen": 1774216, + "step": 2325 + }, + { + "epoch": 4.844074844074844, + "grad_norm": 0.002594448858872056, + "learning_rate": 0.29749753087663217, + "loss": 0.2829, + "num_input_tokens_seen": 1777960, + "step": 2330 + }, + { + "epoch": 4.854469854469855, + "grad_norm": 0.00308404047973454, + "learning_rate": 0.29748680465794985, + "loss": 0.3007, + "num_input_tokens_seen": 1781768, + "step": 2335 + }, + { + "epoch": 4.864864864864865, + "grad_norm": 0.0015814410289749503, + "learning_rate": 0.29747605569494884, + "loss": 0.2741, + "num_input_tokens_seen": 1785480, + "step": 2340 + }, + { + "epoch": 4.875259875259875, + "grad_norm": 0.0005505598383024335, + "learning_rate": 0.29746528398928673, + "loss": 0.2768, + "num_input_tokens_seen": 1789320, + "step": 2345 + }, + { + "epoch": 4.885654885654886, + "grad_norm": 0.0033011562190949917, + "learning_rate": 0.2974544895426247, + "loss": 0.2726, + "num_input_tokens_seen": 1793032, + "step": 2350 + }, + { + "epoch": 4.896049896049896, + "grad_norm": 0.007996849715709686, + "learning_rate": 0.29744367235662733, + "loss": 0.2753, + "num_input_tokens_seen": 1796872, + "step": 2355 + }, + { + "epoch": 4.906444906444906, + "grad_norm": 0.0008668152149766684, + "learning_rate": 0.29743283243296276, + "loss": 0.2592, + "num_input_tokens_seen": 1800680, + "step": 2360 + }, + { + "epoch": 4.916839916839917, + "grad_norm": 0.000578886887524277, + "learning_rate": 0.29742196977330276, + "loss": 0.2792, + "num_input_tokens_seen": 1804552, + "step": 2365 + }, + { + "epoch": 4.927234927234927, + "grad_norm": 0.0006882133893668652, + "learning_rate": 0.2974110843793223, + "loss": 0.2769, + "num_input_tokens_seen": 1808392, + "step": 2370 + }, + { + "epoch": 4.9376299376299375, + "grad_norm": 0.002639227779582143, + "learning_rate": 0.2974001762527002, + "loss": 0.2504, + "num_input_tokens_seen": 1812232, + "step": 2375 + }, + { + "epoch": 4.948024948024948, + "grad_norm": 0.0014049126766622066, + "learning_rate": 0.2973892453951186, + "loss": 0.2921, + "num_input_tokens_seen": 1815944, + "step": 2380 + }, + { + "epoch": 4.958419958419959, + "grad_norm": 0.0032679408323019743, + "learning_rate": 0.2973782918082631, + "loss": 0.2953, + "num_input_tokens_seen": 1819688, + "step": 2385 + }, + { + "epoch": 4.968814968814969, + "grad_norm": 0.0038558205123990774, + "learning_rate": 0.29736731549382295, + "loss": 0.3072, + "num_input_tokens_seen": 1823432, + "step": 2390 + }, + { + "epoch": 4.979209979209979, + "grad_norm": 0.007292178925126791, + "learning_rate": 0.2973563164534908, + "loss": 0.322, + "num_input_tokens_seen": 1827240, + "step": 2395 + }, + { + "epoch": 4.98960498960499, + "grad_norm": 0.003858346026390791, + "learning_rate": 0.29734529468896287, + "loss": 0.3352, + "num_input_tokens_seen": 1831112, + "step": 2400 + }, + { + "epoch": 4.98960498960499, + "eval_loss": 0.2800733745098114, + "eval_runtime": 13.4053, + "eval_samples_per_second": 63.856, + "eval_steps_per_second": 15.964, + "num_input_tokens_seen": 1831112, + "step": 2400 + }, + { + "epoch": 5.0, + "grad_norm": 0.0040953257121145725, + "learning_rate": 0.2973342502019388, + "loss": 0.3131, + "num_input_tokens_seen": 1834816, + "step": 2405 + }, + { + "epoch": 5.01039501039501, + "grad_norm": 0.0005722068017348647, + "learning_rate": 0.2973231829941219, + "loss": 0.3025, + "num_input_tokens_seen": 1838592, + "step": 2410 + }, + { + "epoch": 5.020790020790021, + "grad_norm": 0.006669714115560055, + "learning_rate": 0.2973120930672188, + "loss": 0.3195, + "num_input_tokens_seen": 1842368, + "step": 2415 + }, + { + "epoch": 5.031185031185031, + "grad_norm": 0.0056261541321873665, + "learning_rate": 0.2973009804229397, + "loss": 0.3021, + "num_input_tokens_seen": 1846208, + "step": 2420 + }, + { + "epoch": 5.041580041580041, + "grad_norm": 0.001198282465338707, + "learning_rate": 0.29728984506299827, + "loss": 0.2817, + "num_input_tokens_seen": 1850144, + "step": 2425 + }, + { + "epoch": 5.051975051975052, + "grad_norm": 0.0013224403373897076, + "learning_rate": 0.2972786869891118, + "loss": 0.3173, + "num_input_tokens_seen": 1854176, + "step": 2430 + }, + { + "epoch": 5.0623700623700625, + "grad_norm": 0.0025285715237259865, + "learning_rate": 0.29726750620300096, + "loss": 0.2896, + "num_input_tokens_seen": 1858048, + "step": 2435 + }, + { + "epoch": 5.072765072765073, + "grad_norm": 0.004811470862478018, + "learning_rate": 0.29725630270639003, + "loss": 0.2944, + "num_input_tokens_seen": 1861888, + "step": 2440 + }, + { + "epoch": 5.083160083160083, + "grad_norm": 0.0016600388335064054, + "learning_rate": 0.2972450765010067, + "loss": 0.273, + "num_input_tokens_seen": 1865760, + "step": 2445 + }, + { + "epoch": 5.093555093555094, + "grad_norm": 0.003098409855738282, + "learning_rate": 0.29723382758858213, + "loss": 0.2806, + "num_input_tokens_seen": 1869472, + "step": 2450 + }, + { + "epoch": 5.103950103950104, + "grad_norm": 0.0032900734804570675, + "learning_rate": 0.29722255597085107, + "loss": 0.2622, + "num_input_tokens_seen": 1873184, + "step": 2455 + }, + { + "epoch": 5.114345114345114, + "grad_norm": 0.002768657635897398, + "learning_rate": 0.2972112616495518, + "loss": 0.2683, + "num_input_tokens_seen": 1877088, + "step": 2460 + }, + { + "epoch": 5.124740124740125, + "grad_norm": 0.004758150782436132, + "learning_rate": 0.297199944626426, + "loss": 0.2793, + "num_input_tokens_seen": 1881056, + "step": 2465 + }, + { + "epoch": 5.135135135135135, + "grad_norm": 0.002628802787512541, + "learning_rate": 0.2971886049032189, + "loss": 0.2707, + "num_input_tokens_seen": 1884864, + "step": 2470 + }, + { + "epoch": 5.145530145530145, + "grad_norm": 0.0006630115676671267, + "learning_rate": 0.29717724248167926, + "loss": 0.2628, + "num_input_tokens_seen": 1888704, + "step": 2475 + }, + { + "epoch": 5.155925155925156, + "grad_norm": 0.001930988859385252, + "learning_rate": 0.29716585736355927, + "loss": 0.2806, + "num_input_tokens_seen": 1892608, + "step": 2480 + }, + { + "epoch": 5.166320166320166, + "grad_norm": 0.0005219251615926623, + "learning_rate": 0.2971544495506147, + "loss": 0.2778, + "num_input_tokens_seen": 1896448, + "step": 2485 + }, + { + "epoch": 5.1767151767151764, + "grad_norm": 0.0009427057229913771, + "learning_rate": 0.2971430190446048, + "loss": 0.3103, + "num_input_tokens_seen": 1900416, + "step": 2490 + }, + { + "epoch": 5.1871101871101875, + "grad_norm": 0.0029217449482530355, + "learning_rate": 0.2971315658472921, + "loss": 0.3026, + "num_input_tokens_seen": 1904192, + "step": 2495 + }, + { + "epoch": 5.197505197505198, + "grad_norm": 0.0024619351606816053, + "learning_rate": 0.2971200899604431, + "loss": 0.285, + "num_input_tokens_seen": 1907808, + "step": 2500 + }, + { + "epoch": 5.207900207900208, + "grad_norm": 0.0010934073943644762, + "learning_rate": 0.29710859138582735, + "loss": 0.2577, + "num_input_tokens_seen": 1911584, + "step": 2505 + }, + { + "epoch": 5.218295218295219, + "grad_norm": 0.003578747157007456, + "learning_rate": 0.29709707012521813, + "loss": 0.2985, + "num_input_tokens_seen": 1915264, + "step": 2510 + }, + { + "epoch": 5.228690228690229, + "grad_norm": 0.0009640550124458969, + "learning_rate": 0.29708552618039213, + "loss": 0.3255, + "num_input_tokens_seen": 1919072, + "step": 2515 + }, + { + "epoch": 5.239085239085239, + "grad_norm": 0.0012125809444114566, + "learning_rate": 0.2970739595531296, + "loss": 0.2725, + "num_input_tokens_seen": 1922720, + "step": 2520 + }, + { + "epoch": 5.24948024948025, + "grad_norm": 0.0006683564861305058, + "learning_rate": 0.2970623702452143, + "loss": 0.2718, + "num_input_tokens_seen": 1926496, + "step": 2525 + }, + { + "epoch": 5.25987525987526, + "grad_norm": 0.0027965367771685123, + "learning_rate": 0.2970507582584334, + "loss": 0.2324, + "num_input_tokens_seen": 1930304, + "step": 2530 + }, + { + "epoch": 5.27027027027027, + "grad_norm": 0.001029761740937829, + "learning_rate": 0.2970391235945776, + "loss": 0.2776, + "num_input_tokens_seen": 1934080, + "step": 2535 + }, + { + "epoch": 5.28066528066528, + "grad_norm": 0.0016207931330427527, + "learning_rate": 0.2970274662554412, + "loss": 0.2822, + "num_input_tokens_seen": 1937920, + "step": 2540 + }, + { + "epoch": 5.291060291060291, + "grad_norm": 0.0030946426559239626, + "learning_rate": 0.2970157862428218, + "loss": 0.2942, + "num_input_tokens_seen": 1941728, + "step": 2545 + }, + { + "epoch": 5.301455301455301, + "grad_norm": 0.00411140127107501, + "learning_rate": 0.2970040835585206, + "loss": 0.2876, + "num_input_tokens_seen": 1945696, + "step": 2550 + }, + { + "epoch": 5.3118503118503115, + "grad_norm": 0.0006096501019783318, + "learning_rate": 0.2969923582043424, + "loss": 0.2862, + "num_input_tokens_seen": 1949696, + "step": 2555 + }, + { + "epoch": 5.3222453222453225, + "grad_norm": 0.0034708271268755198, + "learning_rate": 0.2969806101820953, + "loss": 0.2511, + "num_input_tokens_seen": 1953376, + "step": 2560 + }, + { + "epoch": 5.332640332640333, + "grad_norm": 0.009027624502778053, + "learning_rate": 0.2969688394935911, + "loss": 0.3366, + "num_input_tokens_seen": 1957120, + "step": 2565 + }, + { + "epoch": 5.343035343035343, + "grad_norm": 0.0005166474147699773, + "learning_rate": 0.2969570461406449, + "loss": 0.2772, + "num_input_tokens_seen": 1960992, + "step": 2570 + }, + { + "epoch": 5.353430353430354, + "grad_norm": 0.0013280416605994105, + "learning_rate": 0.29694523012507534, + "loss": 0.2569, + "num_input_tokens_seen": 1964576, + "step": 2575 + }, + { + "epoch": 5.363825363825364, + "grad_norm": 0.002501196227967739, + "learning_rate": 0.2969333914487048, + "loss": 0.2734, + "num_input_tokens_seen": 1968224, + "step": 2580 + }, + { + "epoch": 5.374220374220374, + "grad_norm": 0.0019239579560235143, + "learning_rate": 0.2969215301133587, + "loss": 0.2572, + "num_input_tokens_seen": 1972096, + "step": 2585 + }, + { + "epoch": 5.384615384615385, + "grad_norm": 0.0003502235922496766, + "learning_rate": 0.29690964612086634, + "loss": 0.2348, + "num_input_tokens_seen": 1975872, + "step": 2590 + }, + { + "epoch": 5.395010395010395, + "grad_norm": 0.00165514275431633, + "learning_rate": 0.2968977394730604, + "loss": 0.2919, + "num_input_tokens_seen": 1979584, + "step": 2595 + }, + { + "epoch": 5.405405405405405, + "grad_norm": 0.0012301926035434008, + "learning_rate": 0.296885810171777, + "loss": 0.2685, + "num_input_tokens_seen": 1983296, + "step": 2600 + }, + { + "epoch": 5.405405405405405, + "eval_loss": 0.26522520184516907, + "eval_runtime": 13.3936, + "eval_samples_per_second": 63.911, + "eval_steps_per_second": 15.978, + "num_input_tokens_seen": 1983296, + "step": 2600 + }, + { + "epoch": 5.415800415800415, + "grad_norm": 0.004574560094624758, + "learning_rate": 0.2968738582188558, + "loss": 0.3089, + "num_input_tokens_seen": 1986944, + "step": 2605 + }, + { + "epoch": 5.426195426195426, + "grad_norm": 0.004007366485893726, + "learning_rate": 0.2968618836161399, + "loss": 0.2758, + "num_input_tokens_seen": 1990816, + "step": 2610 + }, + { + "epoch": 5.4365904365904365, + "grad_norm": 0.0022180613595992327, + "learning_rate": 0.296849886365476, + "loss": 0.2945, + "num_input_tokens_seen": 1994688, + "step": 2615 + }, + { + "epoch": 5.446985446985447, + "grad_norm": 0.0002435646892990917, + "learning_rate": 0.2968378664687142, + "loss": 0.2629, + "num_input_tokens_seen": 1998560, + "step": 2620 + }, + { + "epoch": 5.457380457380458, + "grad_norm": 0.0024614459834992886, + "learning_rate": 0.296825823927708, + "loss": 0.2667, + "num_input_tokens_seen": 2002368, + "step": 2625 + }, + { + "epoch": 5.467775467775468, + "grad_norm": 0.0045992545783519745, + "learning_rate": 0.29681375874431476, + "loss": 0.2709, + "num_input_tokens_seen": 2006176, + "step": 2630 + }, + { + "epoch": 5.478170478170478, + "grad_norm": 0.0017498929519206285, + "learning_rate": 0.29680167092039483, + "loss": 0.2509, + "num_input_tokens_seen": 2010016, + "step": 2635 + }, + { + "epoch": 5.488565488565489, + "grad_norm": 0.002731350716203451, + "learning_rate": 0.2967895604578125, + "loss": 0.2681, + "num_input_tokens_seen": 2013760, + "step": 2640 + }, + { + "epoch": 5.498960498960499, + "grad_norm": 0.0024128558579832315, + "learning_rate": 0.2967774273584352, + "loss": 0.2642, + "num_input_tokens_seen": 2017632, + "step": 2645 + }, + { + "epoch": 5.509355509355509, + "grad_norm": 0.0014635191764682531, + "learning_rate": 0.2967652716241342, + "loss": 0.2815, + "num_input_tokens_seen": 2021600, + "step": 2650 + }, + { + "epoch": 5.51975051975052, + "grad_norm": 0.00026850702124647796, + "learning_rate": 0.29675309325678384, + "loss": 0.2845, + "num_input_tokens_seen": 2025408, + "step": 2655 + }, + { + "epoch": 5.53014553014553, + "grad_norm": 0.0028420162852853537, + "learning_rate": 0.29674089225826233, + "loss": 0.2824, + "num_input_tokens_seen": 2029280, + "step": 2660 + }, + { + "epoch": 5.54054054054054, + "grad_norm": 0.0022541400976479053, + "learning_rate": 0.29672866863045116, + "loss": 0.2433, + "num_input_tokens_seen": 2033120, + "step": 2665 + }, + { + "epoch": 5.5509355509355505, + "grad_norm": 0.0012572930427268147, + "learning_rate": 0.2967164223752354, + "loss": 0.2603, + "num_input_tokens_seen": 2036864, + "step": 2670 + }, + { + "epoch": 5.5613305613305615, + "grad_norm": 0.0013288218760862947, + "learning_rate": 0.2967041534945035, + "loss": 0.2599, + "num_input_tokens_seen": 2040864, + "step": 2675 + }, + { + "epoch": 5.571725571725572, + "grad_norm": 0.002293024677783251, + "learning_rate": 0.2966918619901476, + "loss": 0.2546, + "num_input_tokens_seen": 2044768, + "step": 2680 + }, + { + "epoch": 5.582120582120583, + "grad_norm": 0.000882921798620373, + "learning_rate": 0.2966795478640631, + "loss": 0.2704, + "num_input_tokens_seen": 2048544, + "step": 2685 + }, + { + "epoch": 5.592515592515593, + "grad_norm": 0.004462075419723988, + "learning_rate": 0.29666721111814903, + "loss": 0.2929, + "num_input_tokens_seen": 2052416, + "step": 2690 + }, + { + "epoch": 5.602910602910603, + "grad_norm": 0.0016282545402646065, + "learning_rate": 0.2966548517543079, + "loss": 0.2663, + "num_input_tokens_seen": 2056128, + "step": 2695 + }, + { + "epoch": 5.613305613305613, + "grad_norm": 0.001500854385085404, + "learning_rate": 0.29664246977444564, + "loss": 0.2896, + "num_input_tokens_seen": 2059968, + "step": 2700 + }, + { + "epoch": 5.623700623700624, + "grad_norm": 0.0007602103287354112, + "learning_rate": 0.2966300651804717, + "loss": 0.2673, + "num_input_tokens_seen": 2063744, + "step": 2705 + }, + { + "epoch": 5.634095634095634, + "grad_norm": 0.0056661940179765224, + "learning_rate": 0.296617637974299, + "loss": 0.3036, + "num_input_tokens_seen": 2067328, + "step": 2710 + }, + { + "epoch": 5.644490644490644, + "grad_norm": 0.0006392626091837883, + "learning_rate": 0.2966051881578441, + "loss": 0.2784, + "num_input_tokens_seen": 2071200, + "step": 2715 + }, + { + "epoch": 5.654885654885655, + "grad_norm": 0.0009071764652617276, + "learning_rate": 0.29659271573302676, + "loss": 0.3175, + "num_input_tokens_seen": 2075040, + "step": 2720 + }, + { + "epoch": 5.665280665280665, + "grad_norm": 0.0009053181856870651, + "learning_rate": 0.2965802207017705, + "loss": 0.2811, + "num_input_tokens_seen": 2078944, + "step": 2725 + }, + { + "epoch": 5.675675675675675, + "grad_norm": 0.0002588829374872148, + "learning_rate": 0.2965677030660021, + "loss": 0.2849, + "num_input_tokens_seen": 2082784, + "step": 2730 + }, + { + "epoch": 5.686070686070686, + "grad_norm": 0.00215737521648407, + "learning_rate": 0.2965551628276521, + "loss": 0.2698, + "num_input_tokens_seen": 2086560, + "step": 2735 + }, + { + "epoch": 5.696465696465697, + "grad_norm": 0.0003844249586109072, + "learning_rate": 0.29654259998865423, + "loss": 0.273, + "num_input_tokens_seen": 2090464, + "step": 2740 + }, + { + "epoch": 5.706860706860707, + "grad_norm": 0.003719937987625599, + "learning_rate": 0.2965300145509458, + "loss": 0.2704, + "num_input_tokens_seen": 2094240, + "step": 2745 + }, + { + "epoch": 5.717255717255718, + "grad_norm": 0.0026224376633763313, + "learning_rate": 0.2965174065164678, + "loss": 0.2646, + "num_input_tokens_seen": 2098080, + "step": 2750 + }, + { + "epoch": 5.727650727650728, + "grad_norm": 0.0019469807157292962, + "learning_rate": 0.2965047758871644, + "loss": 0.2812, + "num_input_tokens_seen": 2101856, + "step": 2755 + }, + { + "epoch": 5.738045738045738, + "grad_norm": 0.0030355053022503853, + "learning_rate": 0.2964921226649835, + "loss": 0.2613, + "num_input_tokens_seen": 2105568, + "step": 2760 + }, + { + "epoch": 5.748440748440748, + "grad_norm": 0.0016812717076390982, + "learning_rate": 0.2964794468518763, + "loss": 0.2641, + "num_input_tokens_seen": 2109376, + "step": 2765 + }, + { + "epoch": 5.758835758835759, + "grad_norm": 0.0022579412907361984, + "learning_rate": 0.2964667484497977, + "loss": 0.2337, + "num_input_tokens_seen": 2113056, + "step": 2770 + }, + { + "epoch": 5.769230769230769, + "grad_norm": 0.001475276192650199, + "learning_rate": 0.29645402746070587, + "loss": 0.2851, + "num_input_tokens_seen": 2116864, + "step": 2775 + }, + { + "epoch": 5.779625779625779, + "grad_norm": 0.00034311559284105897, + "learning_rate": 0.2964412838865625, + "loss": 0.2585, + "num_input_tokens_seen": 2120672, + "step": 2780 + }, + { + "epoch": 5.79002079002079, + "grad_norm": 0.000345252308761701, + "learning_rate": 0.29642851772933293, + "loss": 0.2467, + "num_input_tokens_seen": 2124448, + "step": 2785 + }, + { + "epoch": 5.8004158004158, + "grad_norm": 0.004911020863801241, + "learning_rate": 0.29641572899098567, + "loss": 0.2839, + "num_input_tokens_seen": 2128288, + "step": 2790 + }, + { + "epoch": 5.8108108108108105, + "grad_norm": 0.0007516889600083232, + "learning_rate": 0.29640291767349314, + "loss": 0.2351, + "num_input_tokens_seen": 2132192, + "step": 2795 + }, + { + "epoch": 5.8212058212058215, + "grad_norm": 0.0025285689625889063, + "learning_rate": 0.2963900837788308, + "loss": 0.3158, + "num_input_tokens_seen": 2135968, + "step": 2800 + }, + { + "epoch": 5.8212058212058215, + "eval_loss": 0.2622532546520233, + "eval_runtime": 13.3972, + "eval_samples_per_second": 63.894, + "eval_steps_per_second": 15.974, + "num_input_tokens_seen": 2135968, + "step": 2800 + }, + { + "epoch": 5.831600831600832, + "grad_norm": 0.003932430408895016, + "learning_rate": 0.2963772273089779, + "loss": 0.2827, + "num_input_tokens_seen": 2139712, + "step": 2805 + }, + { + "epoch": 5.841995841995842, + "grad_norm": 0.00220858515240252, + "learning_rate": 0.2963643482659171, + "loss": 0.2926, + "num_input_tokens_seen": 2143552, + "step": 2810 + }, + { + "epoch": 5.852390852390853, + "grad_norm": 0.0028687971644103527, + "learning_rate": 0.2963514466516345, + "loss": 0.2778, + "num_input_tokens_seen": 2147264, + "step": 2815 + }, + { + "epoch": 5.862785862785863, + "grad_norm": 0.004031349439173937, + "learning_rate": 0.2963385224681196, + "loss": 0.2788, + "num_input_tokens_seen": 2151040, + "step": 2820 + }, + { + "epoch": 5.873180873180873, + "grad_norm": 0.005414790008217096, + "learning_rate": 0.29632557571736556, + "loss": 0.2851, + "num_input_tokens_seen": 2154784, + "step": 2825 + }, + { + "epoch": 5.883575883575883, + "grad_norm": 0.001311891246587038, + "learning_rate": 0.2963126064013689, + "loss": 0.275, + "num_input_tokens_seen": 2158464, + "step": 2830 + }, + { + "epoch": 5.893970893970894, + "grad_norm": 0.005444258917123079, + "learning_rate": 0.29629961452212966, + "loss": 0.2729, + "num_input_tokens_seen": 2162336, + "step": 2835 + }, + { + "epoch": 5.904365904365904, + "grad_norm": 0.0008525390876457095, + "learning_rate": 0.2962866000816513, + "loss": 0.2796, + "num_input_tokens_seen": 2166336, + "step": 2840 + }, + { + "epoch": 5.914760914760915, + "grad_norm": 0.0009947591461241245, + "learning_rate": 0.2962735630819409, + "loss": 0.2658, + "num_input_tokens_seen": 2170048, + "step": 2845 + }, + { + "epoch": 5.925155925155925, + "grad_norm": 0.005027471575886011, + "learning_rate": 0.2962605035250089, + "loss": 0.2902, + "num_input_tokens_seen": 2173952, + "step": 2850 + }, + { + "epoch": 5.9355509355509355, + "grad_norm": 0.0036935494281351566, + "learning_rate": 0.29624742141286914, + "loss": 0.284, + "num_input_tokens_seen": 2177728, + "step": 2855 + }, + { + "epoch": 5.945945945945946, + "grad_norm": 0.0033489062916487455, + "learning_rate": 0.29623431674753925, + "loss": 0.2897, + "num_input_tokens_seen": 2181632, + "step": 2860 + }, + { + "epoch": 5.956340956340957, + "grad_norm": 0.0029152543283998966, + "learning_rate": 0.29622118953103993, + "loss": 0.2894, + "num_input_tokens_seen": 2185568, + "step": 2865 + }, + { + "epoch": 5.966735966735967, + "grad_norm": 0.003034088760614395, + "learning_rate": 0.2962080397653957, + "loss": 0.292, + "num_input_tokens_seen": 2189376, + "step": 2870 + }, + { + "epoch": 5.977130977130977, + "grad_norm": 0.0007119469228200614, + "learning_rate": 0.29619486745263435, + "loss": 0.29, + "num_input_tokens_seen": 2193312, + "step": 2875 + }, + { + "epoch": 5.987525987525988, + "grad_norm": 0.0004276215913705528, + "learning_rate": 0.2961816725947873, + "loss": 0.2983, + "num_input_tokens_seen": 2197120, + "step": 2880 + }, + { + "epoch": 5.997920997920998, + "grad_norm": 0.0034098317846655846, + "learning_rate": 0.29616845519388924, + "loss": 0.2898, + "num_input_tokens_seen": 2200832, + "step": 2885 + }, + { + "epoch": 6.008316008316008, + "grad_norm": 0.0021364279091358185, + "learning_rate": 0.2961552152519785, + "loss": 0.2149, + "num_input_tokens_seen": 2204592, + "step": 2890 + }, + { + "epoch": 6.018711018711019, + "grad_norm": 0.0005487263551913202, + "learning_rate": 0.29614195277109695, + "loss": 0.2731, + "num_input_tokens_seen": 2208400, + "step": 2895 + }, + { + "epoch": 6.029106029106029, + "grad_norm": 0.002146113896742463, + "learning_rate": 0.2961286677532897, + "loss": 0.2719, + "num_input_tokens_seen": 2212208, + "step": 2900 + }, + { + "epoch": 6.039501039501039, + "grad_norm": 0.002831290476024151, + "learning_rate": 0.2961153602006055, + "loss": 0.278, + "num_input_tokens_seen": 2216016, + "step": 2905 + }, + { + "epoch": 6.04989604989605, + "grad_norm": 0.0004945373511873186, + "learning_rate": 0.29610203011509656, + "loss": 0.2545, + "num_input_tokens_seen": 2219760, + "step": 2910 + }, + { + "epoch": 6.0602910602910605, + "grad_norm": 0.0011749834520742297, + "learning_rate": 0.29608867749881856, + "loss": 0.2568, + "num_input_tokens_seen": 2223760, + "step": 2915 + }, + { + "epoch": 6.070686070686071, + "grad_norm": 0.00045744189992547035, + "learning_rate": 0.29607530235383067, + "loss": 0.2561, + "num_input_tokens_seen": 2227472, + "step": 2920 + }, + { + "epoch": 6.081081081081081, + "grad_norm": 0.0016379895387217402, + "learning_rate": 0.2960619046821954, + "loss": 0.2692, + "num_input_tokens_seen": 2231152, + "step": 2925 + }, + { + "epoch": 6.091476091476092, + "grad_norm": 0.0038132204208523035, + "learning_rate": 0.2960484844859789, + "loss": 0.2721, + "num_input_tokens_seen": 2234960, + "step": 2930 + }, + { + "epoch": 6.101871101871102, + "grad_norm": 0.0005531749338842928, + "learning_rate": 0.29603504176725076, + "loss": 0.2831, + "num_input_tokens_seen": 2238704, + "step": 2935 + }, + { + "epoch": 6.112266112266112, + "grad_norm": 0.001137790153734386, + "learning_rate": 0.296021576528084, + "loss": 0.289, + "num_input_tokens_seen": 2242768, + "step": 2940 + }, + { + "epoch": 6.122661122661123, + "grad_norm": 0.0008443903061561286, + "learning_rate": 0.29600808877055507, + "loss": 0.272, + "num_input_tokens_seen": 2246800, + "step": 2945 + }, + { + "epoch": 6.133056133056133, + "grad_norm": 0.002689443761482835, + "learning_rate": 0.29599457849674404, + "loss": 0.2428, + "num_input_tokens_seen": 2250608, + "step": 2950 + }, + { + "epoch": 6.143451143451143, + "grad_norm": 0.002616341458633542, + "learning_rate": 0.2959810457087343, + "loss": 0.2855, + "num_input_tokens_seen": 2254480, + "step": 2955 + }, + { + "epoch": 6.153846153846154, + "grad_norm": 0.0014948769239708781, + "learning_rate": 0.2959674904086128, + "loss": 0.2705, + "num_input_tokens_seen": 2258192, + "step": 2960 + }, + { + "epoch": 6.164241164241164, + "grad_norm": 0.0017134882509708405, + "learning_rate": 0.2959539125984699, + "loss": 0.3003, + "num_input_tokens_seen": 2262096, + "step": 2965 + }, + { + "epoch": 6.174636174636174, + "grad_norm": 0.003938702866435051, + "learning_rate": 0.2959403122803996, + "loss": 0.2826, + "num_input_tokens_seen": 2265968, + "step": 2970 + }, + { + "epoch": 6.185031185031185, + "grad_norm": 0.0038301756139844656, + "learning_rate": 0.2959266894564991, + "loss": 0.3034, + "num_input_tokens_seen": 2269712, + "step": 2975 + }, + { + "epoch": 6.1954261954261955, + "grad_norm": 0.002198318485170603, + "learning_rate": 0.2959130441288692, + "loss": 0.3166, + "num_input_tokens_seen": 2273584, + "step": 2980 + }, + { + "epoch": 6.205821205821206, + "grad_norm": 0.0024367321748286486, + "learning_rate": 0.2958993762996143, + "loss": 0.2881, + "num_input_tokens_seen": 2277488, + "step": 2985 + }, + { + "epoch": 6.216216216216216, + "grad_norm": 0.0013135808985680342, + "learning_rate": 0.2958856859708421, + "loss": 0.2756, + "num_input_tokens_seen": 2281584, + "step": 2990 + }, + { + "epoch": 6.226611226611227, + "grad_norm": 0.0008482606499455869, + "learning_rate": 0.2958719731446638, + "loss": 0.2673, + "num_input_tokens_seen": 2285360, + "step": 2995 + }, + { + "epoch": 6.237006237006237, + "grad_norm": 0.00019554028403945267, + "learning_rate": 0.29585823782319404, + "loss": 0.2789, + "num_input_tokens_seen": 2289200, + "step": 3000 + }, + { + "epoch": 6.237006237006237, + "eval_loss": 0.24987664818763733, + "eval_runtime": 13.4423, + "eval_samples_per_second": 63.679, + "eval_steps_per_second": 15.92, + "num_input_tokens_seen": 2289200, + "step": 3000 + }, + { + "epoch": 6.247401247401247, + "grad_norm": 0.0013581948587670922, + "learning_rate": 0.2958444800085511, + "loss": 0.2615, + "num_input_tokens_seen": 2293104, + "step": 3005 + }, + { + "epoch": 6.257796257796258, + "grad_norm": 0.0062060169875621796, + "learning_rate": 0.2958306997028565, + "loss": 0.3159, + "num_input_tokens_seen": 2296912, + "step": 3010 + }, + { + "epoch": 6.268191268191268, + "grad_norm": 0.0007732873782515526, + "learning_rate": 0.2958168969082354, + "loss": 0.2535, + "num_input_tokens_seen": 2300720, + "step": 3015 + }, + { + "epoch": 6.278586278586278, + "grad_norm": 0.0018295745830982924, + "learning_rate": 0.2958030716268164, + "loss": 0.3605, + "num_input_tokens_seen": 2304592, + "step": 3020 + }, + { + "epoch": 6.288981288981289, + "grad_norm": 0.003372908802703023, + "learning_rate": 0.2957892238607314, + "loss": 0.2894, + "num_input_tokens_seen": 2308368, + "step": 3025 + }, + { + "epoch": 6.299376299376299, + "grad_norm": 0.001269126427359879, + "learning_rate": 0.2957753536121161, + "loss": 0.2808, + "num_input_tokens_seen": 2312080, + "step": 3030 + }, + { + "epoch": 6.3097713097713095, + "grad_norm": 0.0008721768972463906, + "learning_rate": 0.29576146088310923, + "loss": 0.2833, + "num_input_tokens_seen": 2315824, + "step": 3035 + }, + { + "epoch": 6.3201663201663205, + "grad_norm": 0.004103085491806269, + "learning_rate": 0.2957475456758533, + "loss": 0.2859, + "num_input_tokens_seen": 2319536, + "step": 3040 + }, + { + "epoch": 6.330561330561331, + "grad_norm": 0.0017040999373421073, + "learning_rate": 0.2957336079924944, + "loss": 0.2791, + "num_input_tokens_seen": 2323472, + "step": 3045 + }, + { + "epoch": 6.340956340956341, + "grad_norm": 0.001673279912211001, + "learning_rate": 0.2957196478351816, + "loss": 0.3094, + "num_input_tokens_seen": 2327248, + "step": 3050 + }, + { + "epoch": 6.351351351351352, + "grad_norm": 0.0047913952730596066, + "learning_rate": 0.295705665206068, + "loss": 0.2627, + "num_input_tokens_seen": 2330992, + "step": 3055 + }, + { + "epoch": 6.361746361746362, + "grad_norm": 0.00355904595926404, + "learning_rate": 0.2956916601073097, + "loss": 0.2803, + "num_input_tokens_seen": 2334736, + "step": 3060 + }, + { + "epoch": 6.372141372141372, + "grad_norm": 0.0006098496378399432, + "learning_rate": 0.29567763254106655, + "loss": 0.2729, + "num_input_tokens_seen": 2338608, + "step": 3065 + }, + { + "epoch": 6.382536382536383, + "grad_norm": 0.000949639012105763, + "learning_rate": 0.29566358250950175, + "loss": 0.2689, + "num_input_tokens_seen": 2342416, + "step": 3070 + }, + { + "epoch": 6.392931392931393, + "grad_norm": 0.0008701894548721611, + "learning_rate": 0.295649510014782, + "loss": 0.279, + "num_input_tokens_seen": 2346256, + "step": 3075 + }, + { + "epoch": 6.403326403326403, + "grad_norm": 0.0036095604300498962, + "learning_rate": 0.2956354150590775, + "loss": 0.2884, + "num_input_tokens_seen": 2350096, + "step": 3080 + }, + { + "epoch": 6.413721413721413, + "grad_norm": 0.003506356617435813, + "learning_rate": 0.2956212976445618, + "loss": 0.2603, + "num_input_tokens_seen": 2353904, + "step": 3085 + }, + { + "epoch": 6.424116424116424, + "grad_norm": 0.0013488584663718939, + "learning_rate": 0.295607157773412, + "loss": 0.2847, + "num_input_tokens_seen": 2357648, + "step": 3090 + }, + { + "epoch": 6.4345114345114345, + "grad_norm": 0.001138151972554624, + "learning_rate": 0.2955929954478087, + "loss": 0.2736, + "num_input_tokens_seen": 2361296, + "step": 3095 + }, + { + "epoch": 6.444906444906445, + "grad_norm": 0.002459130482748151, + "learning_rate": 0.29557881066993585, + "loss": 0.2574, + "num_input_tokens_seen": 2365136, + "step": 3100 + }, + { + "epoch": 6.455301455301456, + "grad_norm": 0.0009467218187637627, + "learning_rate": 0.29556460344198093, + "loss": 0.2858, + "num_input_tokens_seen": 2368880, + "step": 3105 + }, + { + "epoch": 6.465696465696466, + "grad_norm": 0.0010017661843448877, + "learning_rate": 0.29555037376613486, + "loss": 0.3172, + "num_input_tokens_seen": 2372624, + "step": 3110 + }, + { + "epoch": 6.476091476091476, + "grad_norm": 0.00029432657174766064, + "learning_rate": 0.29553612164459203, + "loss": 0.3104, + "num_input_tokens_seen": 2376560, + "step": 3115 + }, + { + "epoch": 6.486486486486487, + "grad_norm": 0.0008937453385442495, + "learning_rate": 0.29552184707955037, + "loss": 0.2895, + "num_input_tokens_seen": 2380496, + "step": 3120 + }, + { + "epoch": 6.496881496881497, + "grad_norm": 0.0012816459638997912, + "learning_rate": 0.29550755007321117, + "loss": 0.2721, + "num_input_tokens_seen": 2384080, + "step": 3125 + }, + { + "epoch": 6.507276507276507, + "grad_norm": 0.002156519563868642, + "learning_rate": 0.29549323062777916, + "loss": 0.2664, + "num_input_tokens_seen": 2387952, + "step": 3130 + }, + { + "epoch": 6.517671517671518, + "grad_norm": 0.0029360484331846237, + "learning_rate": 0.29547888874546263, + "loss": 0.2703, + "num_input_tokens_seen": 2391856, + "step": 3135 + }, + { + "epoch": 6.528066528066528, + "grad_norm": 0.002531585283577442, + "learning_rate": 0.2954645244284732, + "loss": 0.2289, + "num_input_tokens_seen": 2395760, + "step": 3140 + }, + { + "epoch": 6.538461538461538, + "grad_norm": 0.0006937634316273034, + "learning_rate": 0.2954501376790261, + "loss": 0.2559, + "num_input_tokens_seen": 2399600, + "step": 3145 + }, + { + "epoch": 6.548856548856548, + "grad_norm": 0.0014957934617996216, + "learning_rate": 0.29543572849933997, + "loss": 0.2536, + "num_input_tokens_seen": 2403312, + "step": 3150 + }, + { + "epoch": 6.5592515592515594, + "grad_norm": 0.0010256493696942925, + "learning_rate": 0.2954212968916368, + "loss": 0.2818, + "num_input_tokens_seen": 2407088, + "step": 3155 + }, + { + "epoch": 6.56964656964657, + "grad_norm": 0.004824563395231962, + "learning_rate": 0.29540684285814217, + "loss": 0.2902, + "num_input_tokens_seen": 2410960, + "step": 3160 + }, + { + "epoch": 6.58004158004158, + "grad_norm": 0.0028179758228361607, + "learning_rate": 0.2953923664010851, + "loss": 0.2556, + "num_input_tokens_seen": 2414864, + "step": 3165 + }, + { + "epoch": 6.590436590436591, + "grad_norm": 0.0012605141382664442, + "learning_rate": 0.295377867522698, + "loss": 0.2871, + "num_input_tokens_seen": 2418672, + "step": 3170 + }, + { + "epoch": 6.600831600831601, + "grad_norm": 0.0008157733827829361, + "learning_rate": 0.2953633462252168, + "loss": 0.2458, + "num_input_tokens_seen": 2422480, + "step": 3175 + }, + { + "epoch": 6.611226611226611, + "grad_norm": 0.003589311381801963, + "learning_rate": 0.2953488025108809, + "loss": 0.3349, + "num_input_tokens_seen": 2426192, + "step": 3180 + }, + { + "epoch": 6.621621621621622, + "grad_norm": 0.004647602327167988, + "learning_rate": 0.295334236381933, + "loss": 0.3446, + "num_input_tokens_seen": 2430064, + "step": 3185 + }, + { + "epoch": 6.632016632016632, + "grad_norm": 0.0009125987417064607, + "learning_rate": 0.29531964784061954, + "loss": 0.2961, + "num_input_tokens_seen": 2433936, + "step": 3190 + }, + { + "epoch": 6.642411642411642, + "grad_norm": 0.0011381659423932433, + "learning_rate": 0.2953050368891902, + "loss": 0.2817, + "num_input_tokens_seen": 2437744, + "step": 3195 + }, + { + "epoch": 6.652806652806653, + "grad_norm": 0.0020581961143761873, + "learning_rate": 0.29529040352989805, + "loss": 0.2968, + "num_input_tokens_seen": 2441648, + "step": 3200 + }, + { + "epoch": 6.652806652806653, + "eval_loss": 0.2987210750579834, + "eval_runtime": 13.441, + "eval_samples_per_second": 63.686, + "eval_steps_per_second": 15.921, + "num_input_tokens_seen": 2441648, + "step": 3200 + }, + { + "epoch": 6.663201663201663, + "grad_norm": 0.002160570817068219, + "learning_rate": 0.29527574776499993, + "loss": 0.2882, + "num_input_tokens_seen": 2445296, + "step": 3205 + }, + { + "epoch": 6.673596673596673, + "grad_norm": 0.0015864783199504018, + "learning_rate": 0.2952610695967558, + "loss": 0.2745, + "num_input_tokens_seen": 2449072, + "step": 3210 + }, + { + "epoch": 6.6839916839916835, + "grad_norm": 0.0049042971804738045, + "learning_rate": 0.29524636902742935, + "loss": 0.2593, + "num_input_tokens_seen": 2452944, + "step": 3215 + }, + { + "epoch": 6.6943866943866945, + "grad_norm": 0.004881016444414854, + "learning_rate": 0.2952316460592875, + "loss": 0.2545, + "num_input_tokens_seen": 2456784, + "step": 3220 + }, + { + "epoch": 6.704781704781705, + "grad_norm": 0.002456564689055085, + "learning_rate": 0.29521690069460066, + "loss": 0.3016, + "num_input_tokens_seen": 2460656, + "step": 3225 + }, + { + "epoch": 6.715176715176716, + "grad_norm": 0.0019492740975692868, + "learning_rate": 0.29520213293564285, + "loss": 0.3149, + "num_input_tokens_seen": 2464464, + "step": 3230 + }, + { + "epoch": 6.725571725571726, + "grad_norm": 0.002612174255773425, + "learning_rate": 0.29518734278469144, + "loss": 0.365, + "num_input_tokens_seen": 2468304, + "step": 3235 + }, + { + "epoch": 6.735966735966736, + "grad_norm": 0.001191766932606697, + "learning_rate": 0.29517253024402723, + "loss": 0.2907, + "num_input_tokens_seen": 2472240, + "step": 3240 + }, + { + "epoch": 6.746361746361746, + "grad_norm": 0.002067373599857092, + "learning_rate": 0.2951576953159345, + "loss": 0.2997, + "num_input_tokens_seen": 2476144, + "step": 3245 + }, + { + "epoch": 6.756756756756757, + "grad_norm": 0.003907051403075457, + "learning_rate": 0.29514283800270097, + "loss": 0.2957, + "num_input_tokens_seen": 2479824, + "step": 3250 + }, + { + "epoch": 6.767151767151767, + "grad_norm": 0.006945164408534765, + "learning_rate": 0.2951279583066179, + "loss": 0.2891, + "num_input_tokens_seen": 2483344, + "step": 3255 + }, + { + "epoch": 6.777546777546777, + "grad_norm": 0.00037364737363532186, + "learning_rate": 0.2951130562299798, + "loss": 0.2368, + "num_input_tokens_seen": 2487120, + "step": 3260 + }, + { + "epoch": 6.787941787941788, + "grad_norm": 0.0015776692889630795, + "learning_rate": 0.29509813177508487, + "loss": 0.249, + "num_input_tokens_seen": 2490832, + "step": 3265 + }, + { + "epoch": 6.798336798336798, + "grad_norm": 0.0016386404167860746, + "learning_rate": 0.2950831849442346, + "loss": 0.3019, + "num_input_tokens_seen": 2494544, + "step": 3270 + }, + { + "epoch": 6.8087318087318085, + "grad_norm": 0.0033174215350300074, + "learning_rate": 0.2950682157397339, + "loss": 0.296, + "num_input_tokens_seen": 2498352, + "step": 3275 + }, + { + "epoch": 6.8191268191268195, + "grad_norm": 0.0018637643661350012, + "learning_rate": 0.2950532241638914, + "loss": 0.2782, + "num_input_tokens_seen": 2502128, + "step": 3280 + }, + { + "epoch": 6.82952182952183, + "grad_norm": 0.0008951465133577585, + "learning_rate": 0.2950382102190188, + "loss": 0.255, + "num_input_tokens_seen": 2505936, + "step": 3285 + }, + { + "epoch": 6.83991683991684, + "grad_norm": 0.0025906923692673445, + "learning_rate": 0.2950231739074316, + "loss": 0.2779, + "num_input_tokens_seen": 2509680, + "step": 3290 + }, + { + "epoch": 6.850311850311851, + "grad_norm": 0.001334195607341826, + "learning_rate": 0.29500811523144843, + "loss": 0.2717, + "num_input_tokens_seen": 2513744, + "step": 3295 + }, + { + "epoch": 6.860706860706861, + "grad_norm": 0.0022794196847826242, + "learning_rate": 0.2949930341933917, + "loss": 0.3051, + "num_input_tokens_seen": 2517488, + "step": 3300 + }, + { + "epoch": 6.871101871101871, + "grad_norm": 0.001611954066902399, + "learning_rate": 0.29497793079558693, + "loss": 0.2277, + "num_input_tokens_seen": 2521360, + "step": 3305 + }, + { + "epoch": 6.881496881496881, + "grad_norm": 0.0005894139758311212, + "learning_rate": 0.2949628050403633, + "loss": 0.2974, + "num_input_tokens_seen": 2525040, + "step": 3310 + }, + { + "epoch": 6.891891891891892, + "grad_norm": 0.004006496164947748, + "learning_rate": 0.2949476569300535, + "loss": 0.2432, + "num_input_tokens_seen": 2528816, + "step": 3315 + }, + { + "epoch": 6.902286902286902, + "grad_norm": 0.0034496912267059088, + "learning_rate": 0.29493248646699344, + "loss": 0.2927, + "num_input_tokens_seen": 2532560, + "step": 3320 + }, + { + "epoch": 6.912681912681912, + "grad_norm": 0.0014164462918415666, + "learning_rate": 0.29491729365352265, + "loss": 0.2795, + "num_input_tokens_seen": 2536432, + "step": 3325 + }, + { + "epoch": 6.923076923076923, + "grad_norm": 0.0016756703844293952, + "learning_rate": 0.29490207849198397, + "loss": 0.2917, + "num_input_tokens_seen": 2540272, + "step": 3330 + }, + { + "epoch": 6.9334719334719335, + "grad_norm": 0.003685267176479101, + "learning_rate": 0.29488684098472384, + "loss": 0.3848, + "num_input_tokens_seen": 2544112, + "step": 3335 + }, + { + "epoch": 6.943866943866944, + "grad_norm": 0.00037594392779283226, + "learning_rate": 0.2948715811340921, + "loss": 0.2555, + "num_input_tokens_seen": 2547984, + "step": 3340 + }, + { + "epoch": 6.954261954261955, + "grad_norm": 0.0014869710430502892, + "learning_rate": 0.294856298942442, + "loss": 0.2918, + "num_input_tokens_seen": 2551760, + "step": 3345 + }, + { + "epoch": 6.964656964656965, + "grad_norm": 0.001702604815363884, + "learning_rate": 0.2948409944121302, + "loss": 0.2872, + "num_input_tokens_seen": 2555536, + "step": 3350 + }, + { + "epoch": 6.975051975051975, + "grad_norm": 0.0037030810490250587, + "learning_rate": 0.29482566754551687, + "loss": 0.2821, + "num_input_tokens_seen": 2559312, + "step": 3355 + }, + { + "epoch": 6.985446985446986, + "grad_norm": 0.0009733354090712965, + "learning_rate": 0.2948103183449656, + "loss": 0.261, + "num_input_tokens_seen": 2563056, + "step": 3360 + }, + { + "epoch": 6.995841995841996, + "grad_norm": 0.0014379547210410237, + "learning_rate": 0.2947949468128435, + "loss": 0.2887, + "num_input_tokens_seen": 2566832, + "step": 3365 + }, + { + "epoch": 7.006237006237006, + "grad_norm": 0.0029334097635000944, + "learning_rate": 0.2947795529515209, + "loss": 0.2605, + "num_input_tokens_seen": 2570592, + "step": 3370 + }, + { + "epoch": 7.016632016632016, + "grad_norm": 0.0014982214197516441, + "learning_rate": 0.29476413676337193, + "loss": 0.2564, + "num_input_tokens_seen": 2574336, + "step": 3375 + }, + { + "epoch": 7.027027027027027, + "grad_norm": 0.005589826963841915, + "learning_rate": 0.2947486982507738, + "loss": 0.2951, + "num_input_tokens_seen": 2578080, + "step": 3380 + }, + { + "epoch": 7.037422037422037, + "grad_norm": 0.0009938370203599334, + "learning_rate": 0.29473323741610735, + "loss": 0.2814, + "num_input_tokens_seen": 2581920, + "step": 3385 + }, + { + "epoch": 7.047817047817047, + "grad_norm": 0.001087399898096919, + "learning_rate": 0.2947177542617569, + "loss": 0.256, + "num_input_tokens_seen": 2585760, + "step": 3390 + }, + { + "epoch": 7.058212058212058, + "grad_norm": 0.0014066743897274137, + "learning_rate": 0.2947022487901101, + "loss": 0.2837, + "num_input_tokens_seen": 2589568, + "step": 3395 + }, + { + "epoch": 7.0686070686070686, + "grad_norm": 0.0015377472154796124, + "learning_rate": 0.2946867210035581, + "loss": 0.2783, + "num_input_tokens_seen": 2593344, + "step": 3400 + }, + { + "epoch": 7.0686070686070686, + "eval_loss": 0.2496897280216217, + "eval_runtime": 13.4173, + "eval_samples_per_second": 63.798, + "eval_steps_per_second": 15.95, + "num_input_tokens_seen": 2593344, + "step": 3400 + }, + { + "epoch": 7.079002079002079, + "grad_norm": 0.0012426524190232158, + "learning_rate": 0.2946711709044954, + "loss": 0.2362, + "num_input_tokens_seen": 2597056, + "step": 3405 + }, + { + "epoch": 7.08939708939709, + "grad_norm": 0.0032184498850256205, + "learning_rate": 0.2946555984953202, + "loss": 0.3063, + "num_input_tokens_seen": 2600896, + "step": 3410 + }, + { + "epoch": 7.0997920997921, + "grad_norm": 0.003795068711042404, + "learning_rate": 0.2946400037784338, + "loss": 0.2877, + "num_input_tokens_seen": 2604512, + "step": 3415 + }, + { + "epoch": 7.11018711018711, + "grad_norm": 0.0013317748671397567, + "learning_rate": 0.29462438675624114, + "loss": 0.2838, + "num_input_tokens_seen": 2608448, + "step": 3420 + }, + { + "epoch": 7.120582120582121, + "grad_norm": 0.001264024875126779, + "learning_rate": 0.2946087474311506, + "loss": 0.2967, + "num_input_tokens_seen": 2612128, + "step": 3425 + }, + { + "epoch": 7.130977130977131, + "grad_norm": 0.0018578324234113097, + "learning_rate": 0.294593085805574, + "loss": 0.2696, + "num_input_tokens_seen": 2615968, + "step": 3430 + }, + { + "epoch": 7.141372141372141, + "grad_norm": 0.0028327356558293104, + "learning_rate": 0.2945774018819264, + "loss": 0.2713, + "num_input_tokens_seen": 2619776, + "step": 3435 + }, + { + "epoch": 7.151767151767152, + "grad_norm": 0.0003140619373880327, + "learning_rate": 0.2945616956626266, + "loss": 0.2403, + "num_input_tokens_seen": 2623648, + "step": 3440 + }, + { + "epoch": 7.162162162162162, + "grad_norm": 0.0021300730295479298, + "learning_rate": 0.2945459671500966, + "loss": 0.2527, + "num_input_tokens_seen": 2627520, + "step": 3445 + }, + { + "epoch": 7.172557172557172, + "grad_norm": 0.0008381623192690313, + "learning_rate": 0.2945302163467621, + "loss": 0.3084, + "num_input_tokens_seen": 2631232, + "step": 3450 + }, + { + "epoch": 7.182952182952183, + "grad_norm": 0.00034687481820583344, + "learning_rate": 0.2945144432550519, + "loss": 0.335, + "num_input_tokens_seen": 2635008, + "step": 3455 + }, + { + "epoch": 7.1933471933471935, + "grad_norm": 0.000970735854934901, + "learning_rate": 0.29449864787739843, + "loss": 0.3605, + "num_input_tokens_seen": 2638656, + "step": 3460 + }, + { + "epoch": 7.203742203742204, + "grad_norm": 0.0009498162544332445, + "learning_rate": 0.2944828302162376, + "loss": 0.2899, + "num_input_tokens_seen": 2642304, + "step": 3465 + }, + { + "epoch": 7.214137214137214, + "grad_norm": 0.000339907273883, + "learning_rate": 0.2944669902740087, + "loss": 0.2785, + "num_input_tokens_seen": 2646144, + "step": 3470 + }, + { + "epoch": 7.224532224532225, + "grad_norm": 0.0004050501447636634, + "learning_rate": 0.2944511280531544, + "loss": 0.2714, + "num_input_tokens_seen": 2649856, + "step": 3475 + }, + { + "epoch": 7.234927234927235, + "grad_norm": 0.0009633766603656113, + "learning_rate": 0.29443524355612083, + "loss": 0.2849, + "num_input_tokens_seen": 2653568, + "step": 3480 + }, + { + "epoch": 7.245322245322245, + "grad_norm": 0.0017693584086373448, + "learning_rate": 0.29441933678535764, + "loss": 0.2542, + "num_input_tokens_seen": 2657440, + "step": 3485 + }, + { + "epoch": 7.255717255717256, + "grad_norm": 0.0033374472986906767, + "learning_rate": 0.29440340774331786, + "loss": 0.2527, + "num_input_tokens_seen": 2661248, + "step": 3490 + }, + { + "epoch": 7.266112266112266, + "grad_norm": 0.0020857155323028564, + "learning_rate": 0.2943874564324579, + "loss": 0.3139, + "num_input_tokens_seen": 2665088, + "step": 3495 + }, + { + "epoch": 7.276507276507276, + "grad_norm": 0.002290264004841447, + "learning_rate": 0.2943714828552376, + "loss": 0.2983, + "num_input_tokens_seen": 2668832, + "step": 3500 + }, + { + "epoch": 7.286902286902287, + "grad_norm": 0.00248618028126657, + "learning_rate": 0.29435548701412045, + "loss": 0.2686, + "num_input_tokens_seen": 2672608, + "step": 3505 + }, + { + "epoch": 7.297297297297297, + "grad_norm": 0.0010215550428256392, + "learning_rate": 0.2943394689115731, + "loss": 0.3741, + "num_input_tokens_seen": 2676288, + "step": 3510 + }, + { + "epoch": 7.3076923076923075, + "grad_norm": 0.004506244324147701, + "learning_rate": 0.29432342855006577, + "loss": 0.3209, + "num_input_tokens_seen": 2680032, + "step": 3515 + }, + { + "epoch": 7.3180873180873185, + "grad_norm": 0.00153873220551759, + "learning_rate": 0.294307365932072, + "loss": 0.2855, + "num_input_tokens_seen": 2683936, + "step": 3520 + }, + { + "epoch": 7.328482328482329, + "grad_norm": 0.0015606615925207734, + "learning_rate": 0.294291281060069, + "loss": 0.2514, + "num_input_tokens_seen": 2687776, + "step": 3525 + }, + { + "epoch": 7.338877338877339, + "grad_norm": 0.0003289362066425383, + "learning_rate": 0.29427517393653724, + "loss": 0.2474, + "num_input_tokens_seen": 2691584, + "step": 3530 + }, + { + "epoch": 7.349272349272349, + "grad_norm": 0.0014647473581135273, + "learning_rate": 0.29425904456396046, + "loss": 0.2737, + "num_input_tokens_seen": 2695520, + "step": 3535 + }, + { + "epoch": 7.35966735966736, + "grad_norm": 0.0013536863261833787, + "learning_rate": 0.2942428929448262, + "loss": 0.2728, + "num_input_tokens_seen": 2699392, + "step": 3540 + }, + { + "epoch": 7.37006237006237, + "grad_norm": 0.001209497102536261, + "learning_rate": 0.2942267190816252, + "loss": 0.2741, + "num_input_tokens_seen": 2703360, + "step": 3545 + }, + { + "epoch": 7.38045738045738, + "grad_norm": 0.00017752284475136548, + "learning_rate": 0.2942105229768516, + "loss": 0.2493, + "num_input_tokens_seen": 2707168, + "step": 3550 + }, + { + "epoch": 7.390852390852391, + "grad_norm": 0.001956959953531623, + "learning_rate": 0.29419430463300306, + "loss": 0.3017, + "num_input_tokens_seen": 2711040, + "step": 3555 + }, + { + "epoch": 7.401247401247401, + "grad_norm": 0.00018391320190858096, + "learning_rate": 0.2941780640525808, + "loss": 0.294, + "num_input_tokens_seen": 2714816, + "step": 3560 + }, + { + "epoch": 7.411642411642411, + "grad_norm": 0.0013431685511022806, + "learning_rate": 0.2941618012380891, + "loss": 0.2956, + "num_input_tokens_seen": 2718720, + "step": 3565 + }, + { + "epoch": 7.422037422037422, + "grad_norm": 0.0035378821194171906, + "learning_rate": 0.29414551619203605, + "loss": 0.3033, + "num_input_tokens_seen": 2722720, + "step": 3570 + }, + { + "epoch": 7.4324324324324325, + "grad_norm": 0.0012397931423038244, + "learning_rate": 0.29412920891693295, + "loss": 0.2678, + "num_input_tokens_seen": 2726560, + "step": 3575 + }, + { + "epoch": 7.442827442827443, + "grad_norm": 0.0019612584728747606, + "learning_rate": 0.2941128794152946, + "loss": 0.2785, + "num_input_tokens_seen": 2730368, + "step": 3580 + }, + { + "epoch": 7.453222453222454, + "grad_norm": 0.0014950396725907922, + "learning_rate": 0.2940965276896392, + "loss": 0.2603, + "num_input_tokens_seen": 2734048, + "step": 3585 + }, + { + "epoch": 7.463617463617464, + "grad_norm": 0.0005553779774345458, + "learning_rate": 0.2940801537424884, + "loss": 0.2645, + "num_input_tokens_seen": 2737888, + "step": 3590 + }, + { + "epoch": 7.474012474012474, + "grad_norm": 0.0005526858731172979, + "learning_rate": 0.2940637575763673, + "loss": 0.2479, + "num_input_tokens_seen": 2741728, + "step": 3595 + }, + { + "epoch": 7.484407484407485, + "grad_norm": 0.002084500854834914, + "learning_rate": 0.2940473391938043, + "loss": 0.3116, + "num_input_tokens_seen": 2745792, + "step": 3600 + }, + { + "epoch": 7.484407484407485, + "eval_loss": 0.30892422795295715, + "eval_runtime": 13.4337, + "eval_samples_per_second": 63.72, + "eval_steps_per_second": 15.93, + "num_input_tokens_seen": 2745792, + "step": 3600 + }, + { + "epoch": 7.494802494802495, + "grad_norm": 0.0005621399031952024, + "learning_rate": 0.29403089859733145, + "loss": 0.295, + "num_input_tokens_seen": 2749760, + "step": 3605 + }, + { + "epoch": 7.505197505197505, + "grad_norm": 0.002951402449980378, + "learning_rate": 0.294014435789484, + "loss": 0.2926, + "num_input_tokens_seen": 2753440, + "step": 3610 + }, + { + "epoch": 7.515592515592516, + "grad_norm": 0.0028559663332998753, + "learning_rate": 0.2939979507728007, + "loss": 0.264, + "num_input_tokens_seen": 2757280, + "step": 3615 + }, + { + "epoch": 7.525987525987526, + "grad_norm": 0.0005421764799393713, + "learning_rate": 0.2939814435498239, + "loss": 0.2761, + "num_input_tokens_seen": 2761248, + "step": 3620 + }, + { + "epoch": 7.536382536382536, + "grad_norm": 0.00045091958600096405, + "learning_rate": 0.29396491412309905, + "loss": 0.273, + "num_input_tokens_seen": 2764992, + "step": 3625 + }, + { + "epoch": 7.546777546777546, + "grad_norm": 0.0005407624994404614, + "learning_rate": 0.2939483624951753, + "loss": 0.2534, + "num_input_tokens_seen": 2768896, + "step": 3630 + }, + { + "epoch": 7.557172557172557, + "grad_norm": 0.0006500000017695129, + "learning_rate": 0.2939317886686051, + "loss": 0.2463, + "num_input_tokens_seen": 2772768, + "step": 3635 + }, + { + "epoch": 7.5675675675675675, + "grad_norm": 0.0007543907267972827, + "learning_rate": 0.2939151926459443, + "loss": 0.2548, + "num_input_tokens_seen": 2776544, + "step": 3640 + }, + { + "epoch": 7.577962577962578, + "grad_norm": 0.0010387434158474207, + "learning_rate": 0.2938985744297522, + "loss": 0.2815, + "num_input_tokens_seen": 2780416, + "step": 3645 + }, + { + "epoch": 7.588357588357589, + "grad_norm": 0.0008292027050629258, + "learning_rate": 0.29388193402259166, + "loss": 0.2733, + "num_input_tokens_seen": 2784256, + "step": 3650 + }, + { + "epoch": 7.598752598752599, + "grad_norm": 0.00036755899782292545, + "learning_rate": 0.29386527142702873, + "loss": 0.2312, + "num_input_tokens_seen": 2788064, + "step": 3655 + }, + { + "epoch": 7.609147609147609, + "grad_norm": 0.0016179765807464719, + "learning_rate": 0.293848586645633, + "loss": 0.2233, + "num_input_tokens_seen": 2791712, + "step": 3660 + }, + { + "epoch": 7.61954261954262, + "grad_norm": 0.00041799162863753736, + "learning_rate": 0.2938318796809775, + "loss": 0.2966, + "num_input_tokens_seen": 2795456, + "step": 3665 + }, + { + "epoch": 7.62993762993763, + "grad_norm": 0.0023691749665886164, + "learning_rate": 0.29381515053563867, + "loss": 0.2724, + "num_input_tokens_seen": 2799168, + "step": 3670 + }, + { + "epoch": 7.64033264033264, + "grad_norm": 0.002142353681847453, + "learning_rate": 0.29379839921219636, + "loss": 0.2782, + "num_input_tokens_seen": 2802944, + "step": 3675 + }, + { + "epoch": 7.650727650727651, + "grad_norm": 0.0021882164292037487, + "learning_rate": 0.2937816257132338, + "loss": 0.2738, + "num_input_tokens_seen": 2806880, + "step": 3680 + }, + { + "epoch": 7.661122661122661, + "grad_norm": 0.0012558167800307274, + "learning_rate": 0.2937648300413376, + "loss": 0.2815, + "num_input_tokens_seen": 2810848, + "step": 3685 + }, + { + "epoch": 7.671517671517671, + "grad_norm": 0.0031289570033550262, + "learning_rate": 0.293748012199098, + "loss": 0.2927, + "num_input_tokens_seen": 2814592, + "step": 3690 + }, + { + "epoch": 7.6819126819126815, + "grad_norm": 0.0004973011091351509, + "learning_rate": 0.29373117218910844, + "loss": 0.2799, + "num_input_tokens_seen": 2818464, + "step": 3695 + }, + { + "epoch": 7.6923076923076925, + "grad_norm": 0.0031013141851872206, + "learning_rate": 0.2937143100139659, + "loss": 0.2829, + "num_input_tokens_seen": 2822368, + "step": 3700 + }, + { + "epoch": 7.702702702702703, + "grad_norm": 0.0011804451933130622, + "learning_rate": 0.29369742567627083, + "loss": 0.2649, + "num_input_tokens_seen": 2826304, + "step": 3705 + }, + { + "epoch": 7.713097713097713, + "grad_norm": 0.005373354535549879, + "learning_rate": 0.29368051917862675, + "loss": 0.2741, + "num_input_tokens_seen": 2830176, + "step": 3710 + }, + { + "epoch": 7.723492723492724, + "grad_norm": 0.0027358466759324074, + "learning_rate": 0.2936635905236411, + "loss": 0.2761, + "num_input_tokens_seen": 2833824, + "step": 3715 + }, + { + "epoch": 7.733887733887734, + "grad_norm": 0.0011990158818662167, + "learning_rate": 0.2936466397139244, + "loss": 0.2605, + "num_input_tokens_seen": 2837568, + "step": 3720 + }, + { + "epoch": 7.744282744282744, + "grad_norm": 0.0020165364257991314, + "learning_rate": 0.2936296667520907, + "loss": 0.2798, + "num_input_tokens_seen": 2841344, + "step": 3725 + }, + { + "epoch": 7.754677754677755, + "grad_norm": 0.0003979332104790956, + "learning_rate": 0.2936126716407574, + "loss": 0.2664, + "num_input_tokens_seen": 2845120, + "step": 3730 + }, + { + "epoch": 7.765072765072765, + "grad_norm": 0.0034349027555435896, + "learning_rate": 0.29359565438254537, + "loss": 0.2747, + "num_input_tokens_seen": 2848896, + "step": 3735 + }, + { + "epoch": 7.775467775467775, + "grad_norm": 0.0005567230982705951, + "learning_rate": 0.29357861498007887, + "loss": 0.2665, + "num_input_tokens_seen": 2852736, + "step": 3740 + }, + { + "epoch": 7.785862785862786, + "grad_norm": 0.0022427982185035944, + "learning_rate": 0.29356155343598567, + "loss": 0.2629, + "num_input_tokens_seen": 2856608, + "step": 3745 + }, + { + "epoch": 7.796257796257796, + "grad_norm": 0.002074519172310829, + "learning_rate": 0.2935444697528968, + "loss": 0.2976, + "num_input_tokens_seen": 2860448, + "step": 3750 + }, + { + "epoch": 7.8066528066528065, + "grad_norm": 0.0010802532779052854, + "learning_rate": 0.2935273639334468, + "loss": 0.2817, + "num_input_tokens_seen": 2864256, + "step": 3755 + }, + { + "epoch": 7.817047817047817, + "grad_norm": 0.0012731625465676188, + "learning_rate": 0.29351023598027365, + "loss": 0.2544, + "num_input_tokens_seen": 2867936, + "step": 3760 + }, + { + "epoch": 7.827442827442828, + "grad_norm": 0.0017853660974651575, + "learning_rate": 0.2934930858960186, + "loss": 0.2812, + "num_input_tokens_seen": 2871904, + "step": 3765 + }, + { + "epoch": 7.837837837837838, + "grad_norm": 0.0004487301339395344, + "learning_rate": 0.29347591368332643, + "loss": 0.2822, + "num_input_tokens_seen": 2875840, + "step": 3770 + }, + { + "epoch": 7.848232848232849, + "grad_norm": 0.0007442218484356999, + "learning_rate": 0.2934587193448454, + "loss": 0.2929, + "num_input_tokens_seen": 2879648, + "step": 3775 + }, + { + "epoch": 7.858627858627859, + "grad_norm": 0.0013928577536717057, + "learning_rate": 0.29344150288322696, + "loss": 0.2816, + "num_input_tokens_seen": 2883456, + "step": 3780 + }, + { + "epoch": 7.869022869022869, + "grad_norm": 0.0022304090671241283, + "learning_rate": 0.2934242643011263, + "loss": 0.2832, + "num_input_tokens_seen": 2887456, + "step": 3785 + }, + { + "epoch": 7.879417879417879, + "grad_norm": 0.0011471620528027415, + "learning_rate": 0.2934070036012016, + "loss": 0.2489, + "num_input_tokens_seen": 2891296, + "step": 3790 + }, + { + "epoch": 7.88981288981289, + "grad_norm": 0.0011504165595397353, + "learning_rate": 0.29338972078611475, + "loss": 0.2446, + "num_input_tokens_seen": 2895200, + "step": 3795 + }, + { + "epoch": 7.9002079002079, + "grad_norm": 0.0012059705331921577, + "learning_rate": 0.2933724158585311, + "loss": 0.2829, + "num_input_tokens_seen": 2898816, + "step": 3800 + }, + { + "epoch": 7.9002079002079, + "eval_loss": 0.2539701759815216, + "eval_runtime": 13.4268, + "eval_samples_per_second": 63.753, + "eval_steps_per_second": 15.938, + "num_input_tokens_seen": 2898816, + "step": 3800 + }, + { + "epoch": 7.91060291060291, + "grad_norm": 0.0031327817123383284, + "learning_rate": 0.29335508882111916, + "loss": 0.2795, + "num_input_tokens_seen": 2902592, + "step": 3805 + }, + { + "epoch": 7.920997920997921, + "grad_norm": 0.0013943853555247188, + "learning_rate": 0.29333773967655097, + "loss": 0.2698, + "num_input_tokens_seen": 2906432, + "step": 3810 + }, + { + "epoch": 7.9313929313929314, + "grad_norm": 0.001740603824146092, + "learning_rate": 0.2933203684275021, + "loss": 0.2827, + "num_input_tokens_seen": 2910240, + "step": 3815 + }, + { + "epoch": 7.941787941787942, + "grad_norm": 0.002541606780141592, + "learning_rate": 0.2933029750766513, + "loss": 0.2706, + "num_input_tokens_seen": 2913952, + "step": 3820 + }, + { + "epoch": 7.952182952182953, + "grad_norm": 0.002188492566347122, + "learning_rate": 0.2932855596266809, + "loss": 0.2795, + "num_input_tokens_seen": 2917568, + "step": 3825 + }, + { + "epoch": 7.962577962577963, + "grad_norm": 0.0017209012294188142, + "learning_rate": 0.2932681220802765, + "loss": 0.25, + "num_input_tokens_seen": 2921344, + "step": 3830 + }, + { + "epoch": 7.972972972972973, + "grad_norm": 0.0017609309870749712, + "learning_rate": 0.2932506624401274, + "loss": 0.3102, + "num_input_tokens_seen": 2925184, + "step": 3835 + }, + { + "epoch": 7.983367983367984, + "grad_norm": 0.004350841045379639, + "learning_rate": 0.29323318070892584, + "loss": 0.3085, + "num_input_tokens_seen": 2928992, + "step": 3840 + }, + { + "epoch": 7.993762993762994, + "grad_norm": 0.00346102355979383, + "learning_rate": 0.29321567688936784, + "loss": 0.2773, + "num_input_tokens_seen": 2932896, + "step": 3845 + }, + { + "epoch": 8.004158004158004, + "grad_norm": 0.00240493705496192, + "learning_rate": 0.29319815098415275, + "loss": 0.287, + "num_input_tokens_seen": 2936528, + "step": 3850 + }, + { + "epoch": 8.014553014553014, + "grad_norm": 0.00035926824784837663, + "learning_rate": 0.2931806029959832, + "loss": 0.2727, + "num_input_tokens_seen": 2940336, + "step": 3855 + }, + { + "epoch": 8.024948024948024, + "grad_norm": 0.0017733067506924272, + "learning_rate": 0.29316303292756535, + "loss": 0.269, + "num_input_tokens_seen": 2944144, + "step": 3860 + }, + { + "epoch": 8.035343035343036, + "grad_norm": 0.0013503055088222027, + "learning_rate": 0.29314544078160876, + "loss": 0.255, + "num_input_tokens_seen": 2947888, + "step": 3865 + }, + { + "epoch": 8.045738045738046, + "grad_norm": 0.0005201065796427429, + "learning_rate": 0.2931278265608263, + "loss": 0.2952, + "num_input_tokens_seen": 2951664, + "step": 3870 + }, + { + "epoch": 8.056133056133056, + "grad_norm": 0.0031352173537015915, + "learning_rate": 0.29311019026793433, + "loss": 0.2741, + "num_input_tokens_seen": 2955632, + "step": 3875 + }, + { + "epoch": 8.066528066528067, + "grad_norm": 0.002389635192230344, + "learning_rate": 0.29309253190565254, + "loss": 0.3125, + "num_input_tokens_seen": 2959280, + "step": 3880 + }, + { + "epoch": 8.076923076923077, + "grad_norm": 0.00141516060102731, + "learning_rate": 0.2930748514767042, + "loss": 0.276, + "num_input_tokens_seen": 2962928, + "step": 3885 + }, + { + "epoch": 8.087318087318087, + "grad_norm": 0.0027491434011608362, + "learning_rate": 0.29305714898381574, + "loss": 0.2798, + "num_input_tokens_seen": 2966640, + "step": 3890 + }, + { + "epoch": 8.097713097713097, + "grad_norm": 0.0011779993074014783, + "learning_rate": 0.29303942442971714, + "loss": 0.2699, + "num_input_tokens_seen": 2970544, + "step": 3895 + }, + { + "epoch": 8.108108108108109, + "grad_norm": 0.00158534268848598, + "learning_rate": 0.2930216778171417, + "loss": 0.2629, + "num_input_tokens_seen": 2974288, + "step": 3900 + }, + { + "epoch": 8.118503118503119, + "grad_norm": 0.00036598744918592274, + "learning_rate": 0.2930039091488263, + "loss": 0.28, + "num_input_tokens_seen": 2978256, + "step": 3905 + }, + { + "epoch": 8.128898128898129, + "grad_norm": 0.001971879508346319, + "learning_rate": 0.29298611842751093, + "loss": 0.2559, + "num_input_tokens_seen": 2982064, + "step": 3910 + }, + { + "epoch": 8.13929313929314, + "grad_norm": 0.002802992006763816, + "learning_rate": 0.29296830565593923, + "loss": 0.2706, + "num_input_tokens_seen": 2985936, + "step": 3915 + }, + { + "epoch": 8.14968814968815, + "grad_norm": 0.00185109069570899, + "learning_rate": 0.2929504708368582, + "loss": 0.2977, + "num_input_tokens_seen": 2989872, + "step": 3920 + }, + { + "epoch": 8.16008316008316, + "grad_norm": 0.000660832563880831, + "learning_rate": 0.29293261397301806, + "loss": 0.299, + "num_input_tokens_seen": 2993648, + "step": 3925 + }, + { + "epoch": 8.170478170478171, + "grad_norm": 0.000330803421093151, + "learning_rate": 0.29291473506717275, + "loss": 0.2802, + "num_input_tokens_seen": 2997456, + "step": 3930 + }, + { + "epoch": 8.180873180873181, + "grad_norm": 0.001893747365102172, + "learning_rate": 0.29289683412207923, + "loss": 0.2714, + "num_input_tokens_seen": 3001296, + "step": 3935 + }, + { + "epoch": 8.191268191268192, + "grad_norm": 0.0011223821202293038, + "learning_rate": 0.29287891114049813, + "loss": 0.2556, + "num_input_tokens_seen": 3005168, + "step": 3940 + }, + { + "epoch": 8.201663201663202, + "grad_norm": 0.0025791542138904333, + "learning_rate": 0.29286096612519347, + "loss": 0.2947, + "num_input_tokens_seen": 3008912, + "step": 3945 + }, + { + "epoch": 8.212058212058212, + "grad_norm": 0.0005244700005277991, + "learning_rate": 0.2928429990789325, + "loss": 0.2845, + "num_input_tokens_seen": 3012688, + "step": 3950 + }, + { + "epoch": 8.222453222453222, + "grad_norm": 0.0007702698931097984, + "learning_rate": 0.29282501000448596, + "loss": 0.2641, + "num_input_tokens_seen": 3016496, + "step": 3955 + }, + { + "epoch": 8.232848232848234, + "grad_norm": 0.0014458136865869164, + "learning_rate": 0.2928069989046281, + "loss": 0.2372, + "num_input_tokens_seen": 3020400, + "step": 3960 + }, + { + "epoch": 8.243243243243244, + "grad_norm": 0.0004499837523326278, + "learning_rate": 0.2927889657821363, + "loss": 0.2583, + "num_input_tokens_seen": 3024144, + "step": 3965 + }, + { + "epoch": 8.253638253638254, + "grad_norm": 0.0013547156704589725, + "learning_rate": 0.2927709106397916, + "loss": 0.2741, + "num_input_tokens_seen": 3027824, + "step": 3970 + }, + { + "epoch": 8.264033264033264, + "grad_norm": 0.0007258598925545812, + "learning_rate": 0.29275283348037834, + "loss": 0.2432, + "num_input_tokens_seen": 3031536, + "step": 3975 + }, + { + "epoch": 8.274428274428274, + "grad_norm": 0.00018851507047656924, + "learning_rate": 0.29273473430668423, + "loss": 0.3405, + "num_input_tokens_seen": 3035248, + "step": 3980 + }, + { + "epoch": 8.284823284823284, + "grad_norm": 0.0009012404480017722, + "learning_rate": 0.2927166131215003, + "loss": 0.298, + "num_input_tokens_seen": 3038928, + "step": 3985 + }, + { + "epoch": 8.295218295218294, + "grad_norm": 0.002488626865670085, + "learning_rate": 0.2926984699276212, + "loss": 0.2816, + "num_input_tokens_seen": 3042832, + "step": 3990 + }, + { + "epoch": 8.305613305613306, + "grad_norm": 0.0018329822923988104, + "learning_rate": 0.29268030472784473, + "loss": 0.2446, + "num_input_tokens_seen": 3046544, + "step": 3995 + }, + { + "epoch": 8.316008316008316, + "grad_norm": 0.002680974779650569, + "learning_rate": 0.2926621175249723, + "loss": 0.2801, + "num_input_tokens_seen": 3050480, + "step": 4000 + }, + { + "epoch": 8.316008316008316, + "eval_loss": 0.2654391825199127, + "eval_runtime": 13.4682, + "eval_samples_per_second": 63.557, + "eval_steps_per_second": 15.889, + "num_input_tokens_seen": 3050480, + "step": 4000 + }, + { + "epoch": 8.326403326403327, + "grad_norm": 0.0003244733961764723, + "learning_rate": 0.29264390832180853, + "loss": 0.2722, + "num_input_tokens_seen": 3054352, + "step": 4005 + }, + { + "epoch": 8.336798336798337, + "grad_norm": 0.0008291470585390925, + "learning_rate": 0.29262567712116144, + "loss": 0.2867, + "num_input_tokens_seen": 3058096, + "step": 4010 + }, + { + "epoch": 8.347193347193347, + "grad_norm": 0.001917881309054792, + "learning_rate": 0.29260742392584266, + "loss": 0.2743, + "num_input_tokens_seen": 3061776, + "step": 4015 + }, + { + "epoch": 8.357588357588357, + "grad_norm": 0.00015456200344488025, + "learning_rate": 0.292589148738667, + "loss": 0.2811, + "num_input_tokens_seen": 3065648, + "step": 4020 + }, + { + "epoch": 8.367983367983369, + "grad_norm": 0.0018043442396447062, + "learning_rate": 0.2925708515624527, + "loss": 0.262, + "num_input_tokens_seen": 3069456, + "step": 4025 + }, + { + "epoch": 8.378378378378379, + "grad_norm": 0.001464469125494361, + "learning_rate": 0.29255253240002144, + "loss": 0.2627, + "num_input_tokens_seen": 3073232, + "step": 4030 + }, + { + "epoch": 8.388773388773389, + "grad_norm": 0.00012295882334001362, + "learning_rate": 0.2925341912541983, + "loss": 0.2699, + "num_input_tokens_seen": 3077008, + "step": 4035 + }, + { + "epoch": 8.3991683991684, + "grad_norm": 0.003647369332611561, + "learning_rate": 0.2925158281278116, + "loss": 0.283, + "num_input_tokens_seen": 3080848, + "step": 4040 + }, + { + "epoch": 8.40956340956341, + "grad_norm": 0.0002791236329358071, + "learning_rate": 0.29249744302369324, + "loss": 0.2689, + "num_input_tokens_seen": 3084528, + "step": 4045 + }, + { + "epoch": 8.41995841995842, + "grad_norm": 0.0015925226034596562, + "learning_rate": 0.29247903594467844, + "loss": 0.2485, + "num_input_tokens_seen": 3088432, + "step": 4050 + }, + { + "epoch": 8.43035343035343, + "grad_norm": 0.0017383011290803552, + "learning_rate": 0.2924606068936058, + "loss": 0.2554, + "num_input_tokens_seen": 3092464, + "step": 4055 + }, + { + "epoch": 8.440748440748441, + "grad_norm": 0.0008245429489761591, + "learning_rate": 0.2924421558733173, + "loss": 0.2538, + "num_input_tokens_seen": 3096336, + "step": 4060 + }, + { + "epoch": 8.451143451143452, + "grad_norm": 0.0013962909579277039, + "learning_rate": 0.2924236828866583, + "loss": 0.2372, + "num_input_tokens_seen": 3100176, + "step": 4065 + }, + { + "epoch": 8.461538461538462, + "grad_norm": 0.0009744802955538034, + "learning_rate": 0.29240518793647763, + "loss": 0.2861, + "num_input_tokens_seen": 3103856, + "step": 4070 + }, + { + "epoch": 8.471933471933472, + "grad_norm": 0.0014368684496730566, + "learning_rate": 0.29238667102562743, + "loss": 0.2711, + "num_input_tokens_seen": 3107760, + "step": 4075 + }, + { + "epoch": 8.482328482328482, + "grad_norm": 0.003547325963154435, + "learning_rate": 0.29236813215696317, + "loss": 0.2776, + "num_input_tokens_seen": 3111504, + "step": 4080 + }, + { + "epoch": 8.492723492723492, + "grad_norm": 0.0006229123682714999, + "learning_rate": 0.2923495713333439, + "loss": 0.2548, + "num_input_tokens_seen": 3115344, + "step": 4085 + }, + { + "epoch": 8.503118503118504, + "grad_norm": 0.0014495555078610778, + "learning_rate": 0.29233098855763173, + "loss": 0.2835, + "num_input_tokens_seen": 3119152, + "step": 4090 + }, + { + "epoch": 8.513513513513514, + "grad_norm": 0.0003765155270230025, + "learning_rate": 0.29231238383269254, + "loss": 0.2834, + "num_input_tokens_seen": 3122992, + "step": 4095 + }, + { + "epoch": 8.523908523908524, + "grad_norm": 0.001166557427495718, + "learning_rate": 0.2922937571613954, + "loss": 0.2828, + "num_input_tokens_seen": 3126896, + "step": 4100 + }, + { + "epoch": 8.534303534303534, + "grad_norm": 0.0004082966479472816, + "learning_rate": 0.29227510854661265, + "loss": 0.2615, + "num_input_tokens_seen": 3130768, + "step": 4105 + }, + { + "epoch": 8.544698544698544, + "grad_norm": 0.0010949320858344436, + "learning_rate": 0.29225643799122025, + "loss": 0.2904, + "num_input_tokens_seen": 3134640, + "step": 4110 + }, + { + "epoch": 8.555093555093555, + "grad_norm": 0.0008516350644640625, + "learning_rate": 0.2922377454980974, + "loss": 0.2736, + "num_input_tokens_seen": 3138448, + "step": 4115 + }, + { + "epoch": 8.565488565488565, + "grad_norm": 0.0012211723951622844, + "learning_rate": 0.29221903107012676, + "loss": 0.2595, + "num_input_tokens_seen": 3142128, + "step": 4120 + }, + { + "epoch": 8.575883575883577, + "grad_norm": 0.0018546601058915257, + "learning_rate": 0.29220029471019426, + "loss": 0.2411, + "num_input_tokens_seen": 3145776, + "step": 4125 + }, + { + "epoch": 8.586278586278587, + "grad_norm": 0.0007068297709338367, + "learning_rate": 0.2921815364211893, + "loss": 0.2713, + "num_input_tokens_seen": 3149744, + "step": 4130 + }, + { + "epoch": 8.596673596673597, + "grad_norm": 0.0033452603965997696, + "learning_rate": 0.29216275620600474, + "loss": 0.2792, + "num_input_tokens_seen": 3153584, + "step": 4135 + }, + { + "epoch": 8.607068607068607, + "grad_norm": 0.00016673705249559134, + "learning_rate": 0.29214395406753657, + "loss": 0.2522, + "num_input_tokens_seen": 3157488, + "step": 4140 + }, + { + "epoch": 8.617463617463617, + "grad_norm": 0.002090153517201543, + "learning_rate": 0.2921251300086844, + "loss": 0.2922, + "num_input_tokens_seen": 3161232, + "step": 4145 + }, + { + "epoch": 8.627858627858627, + "grad_norm": 0.0019381395541131496, + "learning_rate": 0.2921062840323511, + "loss": 0.2791, + "num_input_tokens_seen": 3165008, + "step": 4150 + }, + { + "epoch": 8.638253638253639, + "grad_norm": 0.0018437692197039723, + "learning_rate": 0.29208741614144307, + "loss": 0.2783, + "num_input_tokens_seen": 3168784, + "step": 4155 + }, + { + "epoch": 8.64864864864865, + "grad_norm": 0.00018089324294123799, + "learning_rate": 0.2920685263388698, + "loss": 0.2609, + "num_input_tokens_seen": 3172496, + "step": 4160 + }, + { + "epoch": 8.65904365904366, + "grad_norm": 0.00019504009105730802, + "learning_rate": 0.2920496146275445, + "loss": 0.2674, + "num_input_tokens_seen": 3176432, + "step": 4165 + }, + { + "epoch": 8.66943866943867, + "grad_norm": 0.0022795891854912043, + "learning_rate": 0.29203068101038343, + "loss": 0.2775, + "num_input_tokens_seen": 3180240, + "step": 4170 + }, + { + "epoch": 8.67983367983368, + "grad_norm": 0.0019669150933623314, + "learning_rate": 0.2920117254903065, + "loss": 0.2729, + "num_input_tokens_seen": 3183984, + "step": 4175 + }, + { + "epoch": 8.69022869022869, + "grad_norm": 0.0022099860943853855, + "learning_rate": 0.29199274807023695, + "loss": 0.2844, + "num_input_tokens_seen": 3187792, + "step": 4180 + }, + { + "epoch": 8.700623700623701, + "grad_norm": 0.0015788617311045527, + "learning_rate": 0.29197374875310117, + "loss": 0.2856, + "num_input_tokens_seen": 3191664, + "step": 4185 + }, + { + "epoch": 8.711018711018712, + "grad_norm": 0.0013265166198834777, + "learning_rate": 0.2919547275418292, + "loss": 0.283, + "num_input_tokens_seen": 3195472, + "step": 4190 + }, + { + "epoch": 8.721413721413722, + "grad_norm": 0.0020737501326948404, + "learning_rate": 0.29193568443935436, + "loss": 0.265, + "num_input_tokens_seen": 3199120, + "step": 4195 + }, + { + "epoch": 8.731808731808732, + "grad_norm": 0.0028038134332746267, + "learning_rate": 0.2919166194486133, + "loss": 0.2669, + "num_input_tokens_seen": 3202864, + "step": 4200 + }, + { + "epoch": 8.731808731808732, + "eval_loss": 0.25126686692237854, + "eval_runtime": 13.4091, + "eval_samples_per_second": 63.837, + "eval_steps_per_second": 15.959, + "num_input_tokens_seen": 3202864, + "step": 4200 + }, + { + "epoch": 8.742203742203742, + "grad_norm": 0.0002503364812582731, + "learning_rate": 0.2918975325725461, + "loss": 0.2679, + "num_input_tokens_seen": 3206640, + "step": 4205 + }, + { + "epoch": 8.752598752598752, + "grad_norm": 0.0005465574213303626, + "learning_rate": 0.29187842381409607, + "loss": 0.2869, + "num_input_tokens_seen": 3210512, + "step": 4210 + }, + { + "epoch": 8.762993762993762, + "grad_norm": 0.0005704878713004291, + "learning_rate": 0.29185929317621023, + "loss": 0.2823, + "num_input_tokens_seen": 3214256, + "step": 4215 + }, + { + "epoch": 8.773388773388774, + "grad_norm": 0.0007609032327309251, + "learning_rate": 0.29184014066183867, + "loss": 0.2863, + "num_input_tokens_seen": 3218128, + "step": 4220 + }, + { + "epoch": 8.783783783783784, + "grad_norm": 0.0003097786975558847, + "learning_rate": 0.2918209662739349, + "loss": 0.2701, + "num_input_tokens_seen": 3222000, + "step": 4225 + }, + { + "epoch": 8.794178794178794, + "grad_norm": 0.0021032600197941065, + "learning_rate": 0.29180177001545593, + "loss": 0.2868, + "num_input_tokens_seen": 3225808, + "step": 4230 + }, + { + "epoch": 8.804573804573804, + "grad_norm": 0.0005216242861934006, + "learning_rate": 0.29178255188936203, + "loss": 0.2809, + "num_input_tokens_seen": 3229840, + "step": 4235 + }, + { + "epoch": 8.814968814968815, + "grad_norm": 0.0001477514742873609, + "learning_rate": 0.2917633118986169, + "loss": 0.2641, + "num_input_tokens_seen": 3233648, + "step": 4240 + }, + { + "epoch": 8.825363825363825, + "grad_norm": 0.00027837895322591066, + "learning_rate": 0.2917440500461875, + "loss": 0.2835, + "num_input_tokens_seen": 3237680, + "step": 4245 + }, + { + "epoch": 8.835758835758837, + "grad_norm": 0.0004899093182757497, + "learning_rate": 0.29172476633504435, + "loss": 0.2742, + "num_input_tokens_seen": 3241648, + "step": 4250 + }, + { + "epoch": 8.846153846153847, + "grad_norm": 0.00013976104673929513, + "learning_rate": 0.2917054607681612, + "loss": 0.2796, + "num_input_tokens_seen": 3245392, + "step": 4255 + }, + { + "epoch": 8.856548856548857, + "grad_norm": 0.002230289625003934, + "learning_rate": 0.29168613334851523, + "loss": 0.2758, + "num_input_tokens_seen": 3249072, + "step": 4260 + }, + { + "epoch": 8.866943866943867, + "grad_norm": 0.0009966047946363688, + "learning_rate": 0.2916667840790869, + "loss": 0.2501, + "num_input_tokens_seen": 3253040, + "step": 4265 + }, + { + "epoch": 8.877338877338877, + "grad_norm": 0.0005958973779343069, + "learning_rate": 0.2916474129628603, + "loss": 0.2715, + "num_input_tokens_seen": 3256848, + "step": 4270 + }, + { + "epoch": 8.887733887733887, + "grad_norm": 0.00016244281141553074, + "learning_rate": 0.29162802000282245, + "loss": 0.2485, + "num_input_tokens_seen": 3260720, + "step": 4275 + }, + { + "epoch": 8.898128898128899, + "grad_norm": 0.0021869537886232138, + "learning_rate": 0.2916086052019642, + "loss": 0.2599, + "num_input_tokens_seen": 3264560, + "step": 4280 + }, + { + "epoch": 8.90852390852391, + "grad_norm": 0.002946468535810709, + "learning_rate": 0.2915891685632794, + "loss": 0.2726, + "num_input_tokens_seen": 3268368, + "step": 4285 + }, + { + "epoch": 8.91891891891892, + "grad_norm": 0.0028296001255512238, + "learning_rate": 0.29156971008976545, + "loss": 0.2895, + "num_input_tokens_seen": 3272112, + "step": 4290 + }, + { + "epoch": 8.92931392931393, + "grad_norm": 0.0012245416874065995, + "learning_rate": 0.2915502297844232, + "loss": 0.2724, + "num_input_tokens_seen": 3275952, + "step": 4295 + }, + { + "epoch": 8.93970893970894, + "grad_norm": 0.0005565624451264739, + "learning_rate": 0.2915307276502566, + "loss": 0.2998, + "num_input_tokens_seen": 3279792, + "step": 4300 + }, + { + "epoch": 8.95010395010395, + "grad_norm": 0.0018246360123157501, + "learning_rate": 0.29151120369027334, + "loss": 0.2822, + "num_input_tokens_seen": 3283632, + "step": 4305 + }, + { + "epoch": 8.96049896049896, + "grad_norm": 0.002096001524478197, + "learning_rate": 0.29149165790748405, + "loss": 0.2792, + "num_input_tokens_seen": 3287440, + "step": 4310 + }, + { + "epoch": 8.970893970893972, + "grad_norm": 0.0006191023858264089, + "learning_rate": 0.291472090304903, + "loss": 0.2917, + "num_input_tokens_seen": 3291120, + "step": 4315 + }, + { + "epoch": 8.981288981288982, + "grad_norm": 0.0015177656896412373, + "learning_rate": 0.2914525008855478, + "loss": 0.2863, + "num_input_tokens_seen": 3294960, + "step": 4320 + }, + { + "epoch": 8.991683991683992, + "grad_norm": 0.0011767708929255605, + "learning_rate": 0.2914328896524394, + "loss": 0.2591, + "num_input_tokens_seen": 3298768, + "step": 4325 + }, + { + "epoch": 9.002079002079002, + "grad_norm": 0.0001497742923675105, + "learning_rate": 0.291413256608602, + "loss": 0.2477, + "num_input_tokens_seen": 3302528, + "step": 4330 + }, + { + "epoch": 9.012474012474012, + "grad_norm": 0.0015019926941022277, + "learning_rate": 0.29139360175706336, + "loss": 0.2695, + "num_input_tokens_seen": 3306464, + "step": 4335 + }, + { + "epoch": 9.022869022869022, + "grad_norm": 0.0022210744209587574, + "learning_rate": 0.2913739251008544, + "loss": 0.2802, + "num_input_tokens_seen": 3310208, + "step": 4340 + }, + { + "epoch": 9.033264033264032, + "grad_norm": 0.0024578555021435022, + "learning_rate": 0.29135422664300964, + "loss": 0.2803, + "num_input_tokens_seen": 3314080, + "step": 4345 + }, + { + "epoch": 9.043659043659044, + "grad_norm": 0.00010010333062382415, + "learning_rate": 0.29133450638656677, + "loss": 0.307, + "num_input_tokens_seen": 3317952, + "step": 4350 + }, + { + "epoch": 9.054054054054054, + "grad_norm": 0.0009152779821306467, + "learning_rate": 0.2913147643345669, + "loss": 0.2943, + "num_input_tokens_seen": 3321728, + "step": 4355 + }, + { + "epoch": 9.064449064449065, + "grad_norm": 0.0010701373685151339, + "learning_rate": 0.29129500049005447, + "loss": 0.2658, + "num_input_tokens_seen": 3325600, + "step": 4360 + }, + { + "epoch": 9.074844074844075, + "grad_norm": 0.0029930565506219864, + "learning_rate": 0.2912752148560773, + "loss": 0.2969, + "num_input_tokens_seen": 3329248, + "step": 4365 + }, + { + "epoch": 9.085239085239085, + "grad_norm": 0.0003962019691243768, + "learning_rate": 0.2912554074356866, + "loss": 0.2617, + "num_input_tokens_seen": 3333024, + "step": 4370 + }, + { + "epoch": 9.095634095634095, + "grad_norm": 0.00014814978931099176, + "learning_rate": 0.2912355782319371, + "loss": 0.2787, + "num_input_tokens_seen": 3336768, + "step": 4375 + }, + { + "epoch": 9.106029106029107, + "grad_norm": 0.00015518043073825538, + "learning_rate": 0.2912157272478864, + "loss": 0.2823, + "num_input_tokens_seen": 3340768, + "step": 4380 + }, + { + "epoch": 9.116424116424117, + "grad_norm": 0.0012748459121212363, + "learning_rate": 0.291195854486596, + "loss": 0.2712, + "num_input_tokens_seen": 3344576, + "step": 4385 + }, + { + "epoch": 9.126819126819127, + "grad_norm": 0.00031417200807482004, + "learning_rate": 0.2911759599511305, + "loss": 0.3012, + "num_input_tokens_seen": 3348352, + "step": 4390 + }, + { + "epoch": 9.137214137214137, + "grad_norm": 0.0007369376835413277, + "learning_rate": 0.29115604364455777, + "loss": 0.2809, + "num_input_tokens_seen": 3352064, + "step": 4395 + }, + { + "epoch": 9.147609147609147, + "grad_norm": 0.00012129736569477245, + "learning_rate": 0.2911361055699493, + "loss": 0.2617, + "num_input_tokens_seen": 3355680, + "step": 4400 + }, + { + "epoch": 9.147609147609147, + "eval_loss": 0.25105273723602295, + "eval_runtime": 13.4218, + "eval_samples_per_second": 63.777, + "eval_steps_per_second": 15.944, + "num_input_tokens_seen": 3355680, + "step": 4400 + }, + { + "epoch": 9.158004158004157, + "grad_norm": 0.000679651799146086, + "learning_rate": 0.2911161457303797, + "loss": 0.2642, + "num_input_tokens_seen": 3359360, + "step": 4405 + }, + { + "epoch": 9.16839916839917, + "grad_norm": 0.0017391651635989547, + "learning_rate": 0.291096164128927, + "loss": 0.2793, + "num_input_tokens_seen": 3363104, + "step": 4410 + }, + { + "epoch": 9.17879417879418, + "grad_norm": 0.0006626248941756785, + "learning_rate": 0.2910761607686727, + "loss": 0.2821, + "num_input_tokens_seen": 3366848, + "step": 4415 + }, + { + "epoch": 9.18918918918919, + "grad_norm": 0.0018741786479949951, + "learning_rate": 0.2910561356527016, + "loss": 0.2822, + "num_input_tokens_seen": 3370528, + "step": 4420 + }, + { + "epoch": 9.1995841995842, + "grad_norm": 0.0007630917243659496, + "learning_rate": 0.2910360887841017, + "loss": 0.2904, + "num_input_tokens_seen": 3374304, + "step": 4425 + }, + { + "epoch": 9.20997920997921, + "grad_norm": 0.0006876765401102602, + "learning_rate": 0.2910160201659645, + "loss": 0.2793, + "num_input_tokens_seen": 3378112, + "step": 4430 + }, + { + "epoch": 9.22037422037422, + "grad_norm": 0.0033967948984354734, + "learning_rate": 0.29099592980138494, + "loss": 0.3019, + "num_input_tokens_seen": 3381984, + "step": 4435 + }, + { + "epoch": 9.23076923076923, + "grad_norm": 0.0001733355165924877, + "learning_rate": 0.29097581769346115, + "loss": 0.2511, + "num_input_tokens_seen": 3385856, + "step": 4440 + }, + { + "epoch": 9.241164241164242, + "grad_norm": 0.0008578650304116309, + "learning_rate": 0.29095568384529463, + "loss": 0.2849, + "num_input_tokens_seen": 3389856, + "step": 4445 + }, + { + "epoch": 9.251559251559252, + "grad_norm": 0.002634449163451791, + "learning_rate": 0.2909355282599903, + "loss": 0.2698, + "num_input_tokens_seen": 3393728, + "step": 4450 + }, + { + "epoch": 9.261954261954262, + "grad_norm": 0.0010786305647343397, + "learning_rate": 0.29091535094065635, + "loss": 0.2778, + "num_input_tokens_seen": 3397600, + "step": 4455 + }, + { + "epoch": 9.272349272349272, + "grad_norm": 0.0017043001716956496, + "learning_rate": 0.2908951518904045, + "loss": 0.2514, + "num_input_tokens_seen": 3401504, + "step": 4460 + }, + { + "epoch": 9.282744282744282, + "grad_norm": 0.00040898696170188487, + "learning_rate": 0.29087493111234963, + "loss": 0.2622, + "num_input_tokens_seen": 3405376, + "step": 4465 + }, + { + "epoch": 9.293139293139292, + "grad_norm": 0.0008852222817949951, + "learning_rate": 0.29085468860961, + "loss": 0.2916, + "num_input_tokens_seen": 3409120, + "step": 4470 + }, + { + "epoch": 9.303534303534304, + "grad_norm": 0.000428498926339671, + "learning_rate": 0.2908344243853073, + "loss": 0.2795, + "num_input_tokens_seen": 3412992, + "step": 4475 + }, + { + "epoch": 9.313929313929314, + "grad_norm": 0.00045505608431994915, + "learning_rate": 0.2908141384425666, + "loss": 0.2748, + "num_input_tokens_seen": 3416736, + "step": 4480 + }, + { + "epoch": 9.324324324324325, + "grad_norm": 0.0015887983608990908, + "learning_rate": 0.2907938307845161, + "loss": 0.2784, + "num_input_tokens_seen": 3420320, + "step": 4485 + }, + { + "epoch": 9.334719334719335, + "grad_norm": 0.0005984515300951898, + "learning_rate": 0.2907735014142876, + "loss": 0.2617, + "num_input_tokens_seen": 3424160, + "step": 4490 + }, + { + "epoch": 9.345114345114345, + "grad_norm": 0.0005569813656620681, + "learning_rate": 0.2907531503350161, + "loss": 0.2816, + "num_input_tokens_seen": 3428192, + "step": 4495 + }, + { + "epoch": 9.355509355509355, + "grad_norm": 0.00018043076852336526, + "learning_rate": 0.29073277754983995, + "loss": 0.2804, + "num_input_tokens_seen": 3431904, + "step": 4500 + }, + { + "epoch": 9.365904365904367, + "grad_norm": 0.0013101734220981598, + "learning_rate": 0.290712383061901, + "loss": 0.2865, + "num_input_tokens_seen": 3435776, + "step": 4505 + }, + { + "epoch": 9.376299376299377, + "grad_norm": 0.002785986987873912, + "learning_rate": 0.2906919668743443, + "loss": 0.2891, + "num_input_tokens_seen": 3439616, + "step": 4510 + }, + { + "epoch": 9.386694386694387, + "grad_norm": 0.0003400310524739325, + "learning_rate": 0.29067152899031823, + "loss": 0.2636, + "num_input_tokens_seen": 3443360, + "step": 4515 + }, + { + "epoch": 9.397089397089397, + "grad_norm": 0.0018372413469478488, + "learning_rate": 0.2906510694129746, + "loss": 0.2682, + "num_input_tokens_seen": 3447168, + "step": 4520 + }, + { + "epoch": 9.407484407484407, + "grad_norm": 0.0008135041571222246, + "learning_rate": 0.2906305881454685, + "loss": 0.261, + "num_input_tokens_seen": 3450848, + "step": 4525 + }, + { + "epoch": 9.417879417879417, + "grad_norm": 0.0006873209495097399, + "learning_rate": 0.2906100851909585, + "loss": 0.2778, + "num_input_tokens_seen": 3454688, + "step": 4530 + }, + { + "epoch": 9.428274428274428, + "grad_norm": 0.00018847943283617496, + "learning_rate": 0.29058956055260626, + "loss": 0.2477, + "num_input_tokens_seen": 3458720, + "step": 4535 + }, + { + "epoch": 9.43866943866944, + "grad_norm": 0.0006840723799541593, + "learning_rate": 0.2905690142335771, + "loss": 0.2659, + "num_input_tokens_seen": 3462464, + "step": 4540 + }, + { + "epoch": 9.44906444906445, + "grad_norm": 0.001151498407125473, + "learning_rate": 0.29054844623703946, + "loss": 0.2526, + "num_input_tokens_seen": 3466112, + "step": 4545 + }, + { + "epoch": 9.45945945945946, + "grad_norm": 0.0005568534252233803, + "learning_rate": 0.2905278565661651, + "loss": 0.2565, + "num_input_tokens_seen": 3470016, + "step": 4550 + }, + { + "epoch": 9.46985446985447, + "grad_norm": 0.00042003337875939906, + "learning_rate": 0.2905072452241293, + "loss": 0.2567, + "num_input_tokens_seen": 3473952, + "step": 4555 + }, + { + "epoch": 9.48024948024948, + "grad_norm": 0.0012943726032972336, + "learning_rate": 0.2904866122141106, + "loss": 0.2749, + "num_input_tokens_seen": 3477920, + "step": 4560 + }, + { + "epoch": 9.49064449064449, + "grad_norm": 0.0007788283401168883, + "learning_rate": 0.2904659575392908, + "loss": 0.2454, + "num_input_tokens_seen": 3482016, + "step": 4565 + }, + { + "epoch": 9.5010395010395, + "grad_norm": 0.0014733957359567285, + "learning_rate": 0.2904452812028551, + "loss": 0.2804, + "num_input_tokens_seen": 3485728, + "step": 4570 + }, + { + "epoch": 9.511434511434512, + "grad_norm": 0.0013959248317405581, + "learning_rate": 0.2904245832079922, + "loss": 0.251, + "num_input_tokens_seen": 3489632, + "step": 4575 + }, + { + "epoch": 9.521829521829522, + "grad_norm": 0.0006215545581653714, + "learning_rate": 0.29040386355789377, + "loss": 0.2819, + "num_input_tokens_seen": 3493152, + "step": 4580 + }, + { + "epoch": 9.532224532224532, + "grad_norm": 0.0006808478501625359, + "learning_rate": 0.29038312225575524, + "loss": 0.2677, + "num_input_tokens_seen": 3496992, + "step": 4585 + }, + { + "epoch": 9.542619542619542, + "grad_norm": 0.0037674764171242714, + "learning_rate": 0.29036235930477505, + "loss": 0.3004, + "num_input_tokens_seen": 3500768, + "step": 4590 + }, + { + "epoch": 9.553014553014552, + "grad_norm": 0.004937266930937767, + "learning_rate": 0.29034157470815514, + "loss": 0.3232, + "num_input_tokens_seen": 3504640, + "step": 4595 + }, + { + "epoch": 9.563409563409563, + "grad_norm": 0.00010515181202208623, + "learning_rate": 0.2903207684691008, + "loss": 0.2678, + "num_input_tokens_seen": 3508192, + "step": 4600 + }, + { + "epoch": 9.563409563409563, + "eval_loss": 0.2624172568321228, + "eval_runtime": 13.4039, + "eval_samples_per_second": 63.862, + "eval_steps_per_second": 15.966, + "num_input_tokens_seen": 3508192, + "step": 4600 + }, + { + "epoch": 9.573804573804575, + "grad_norm": 0.0003542314516380429, + "learning_rate": 0.29029994059082054, + "loss": 0.2704, + "num_input_tokens_seen": 3511936, + "step": 4605 + }, + { + "epoch": 9.584199584199585, + "grad_norm": 0.0004584401031024754, + "learning_rate": 0.2902790910765264, + "loss": 0.2254, + "num_input_tokens_seen": 3515840, + "step": 4610 + }, + { + "epoch": 9.594594594594595, + "grad_norm": 0.002230600453913212, + "learning_rate": 0.29025821992943346, + "loss": 0.2854, + "num_input_tokens_seen": 3519648, + "step": 4615 + }, + { + "epoch": 9.604989604989605, + "grad_norm": 0.000521887035574764, + "learning_rate": 0.29023732715276046, + "loss": 0.2686, + "num_input_tokens_seen": 3523520, + "step": 4620 + }, + { + "epoch": 9.615384615384615, + "grad_norm": 0.00011615233233897015, + "learning_rate": 0.2902164127497293, + "loss": 0.2881, + "num_input_tokens_seen": 3527232, + "step": 4625 + }, + { + "epoch": 9.625779625779625, + "grad_norm": 0.00018442737928126007, + "learning_rate": 0.2901954767235652, + "loss": 0.2765, + "num_input_tokens_seen": 3531040, + "step": 4630 + }, + { + "epoch": 9.636174636174637, + "grad_norm": 0.0010019656037911773, + "learning_rate": 0.2901745190774968, + "loss": 0.2549, + "num_input_tokens_seen": 3534976, + "step": 4635 + }, + { + "epoch": 9.646569646569647, + "grad_norm": 0.0009959720773622394, + "learning_rate": 0.290153539814756, + "loss": 0.263, + "num_input_tokens_seen": 3538752, + "step": 4640 + }, + { + "epoch": 9.656964656964657, + "grad_norm": 0.000723016622941941, + "learning_rate": 0.2901325389385781, + "loss": 0.2362, + "num_input_tokens_seen": 3542432, + "step": 4645 + }, + { + "epoch": 9.667359667359667, + "grad_norm": 0.0018010125495493412, + "learning_rate": 0.2901115164522016, + "loss": 0.2763, + "num_input_tokens_seen": 3546240, + "step": 4650 + }, + { + "epoch": 9.677754677754677, + "grad_norm": 0.001170750823803246, + "learning_rate": 0.29009047235886865, + "loss": 0.2732, + "num_input_tokens_seen": 3550208, + "step": 4655 + }, + { + "epoch": 9.688149688149688, + "grad_norm": 0.0016040452755987644, + "learning_rate": 0.2900694066618243, + "loss": 0.2917, + "num_input_tokens_seen": 3554080, + "step": 4660 + }, + { + "epoch": 9.698544698544698, + "grad_norm": 0.001458129845559597, + "learning_rate": 0.2900483193643172, + "loss": 0.2793, + "num_input_tokens_seen": 3557984, + "step": 4665 + }, + { + "epoch": 9.70893970893971, + "grad_norm": 0.001126165152527392, + "learning_rate": 0.29002721046959934, + "loss": 0.2779, + "num_input_tokens_seen": 3561760, + "step": 4670 + }, + { + "epoch": 9.71933471933472, + "grad_norm": 0.001675745821557939, + "learning_rate": 0.29000607998092587, + "loss": 0.2662, + "num_input_tokens_seen": 3565600, + "step": 4675 + }, + { + "epoch": 9.72972972972973, + "grad_norm": 0.0008758932235650718, + "learning_rate": 0.2899849279015555, + "loss": 0.2738, + "num_input_tokens_seen": 3569408, + "step": 4680 + }, + { + "epoch": 9.74012474012474, + "grad_norm": 0.0011675817659124732, + "learning_rate": 0.28996375423475007, + "loss": 0.2649, + "num_input_tokens_seen": 3573376, + "step": 4685 + }, + { + "epoch": 9.75051975051975, + "grad_norm": 0.0006994726718403399, + "learning_rate": 0.28994255898377486, + "loss": 0.2558, + "num_input_tokens_seen": 3577312, + "step": 4690 + }, + { + "epoch": 9.76091476091476, + "grad_norm": 0.0017045224085450172, + "learning_rate": 0.2899213421518984, + "loss": 0.2265, + "num_input_tokens_seen": 3580992, + "step": 4695 + }, + { + "epoch": 9.771309771309772, + "grad_norm": 0.0012412710348144174, + "learning_rate": 0.2899001037423926, + "loss": 0.2731, + "num_input_tokens_seen": 3584768, + "step": 4700 + }, + { + "epoch": 9.781704781704782, + "grad_norm": 0.001517456374131143, + "learning_rate": 0.28987884375853273, + "loss": 0.2737, + "num_input_tokens_seen": 3588544, + "step": 4705 + }, + { + "epoch": 9.792099792099792, + "grad_norm": 0.0004324537003412843, + "learning_rate": 0.2898575622035974, + "loss": 0.2732, + "num_input_tokens_seen": 3592448, + "step": 4710 + }, + { + "epoch": 9.802494802494802, + "grad_norm": 0.001739976811222732, + "learning_rate": 0.2898362590808683, + "loss": 0.2711, + "num_input_tokens_seen": 3596352, + "step": 4715 + }, + { + "epoch": 9.812889812889813, + "grad_norm": 0.0031527949031442404, + "learning_rate": 0.2898149343936308, + "loss": 0.2866, + "num_input_tokens_seen": 3600224, + "step": 4720 + }, + { + "epoch": 9.823284823284823, + "grad_norm": 0.002125125378370285, + "learning_rate": 0.2897935881451734, + "loss": 0.2814, + "num_input_tokens_seen": 3604032, + "step": 4725 + }, + { + "epoch": 9.833679833679835, + "grad_norm": 0.0005428296863101423, + "learning_rate": 0.28977222033878797, + "loss": 0.2461, + "num_input_tokens_seen": 3607776, + "step": 4730 + }, + { + "epoch": 9.844074844074845, + "grad_norm": 0.0017898917431011796, + "learning_rate": 0.28975083097776966, + "loss": 0.2721, + "num_input_tokens_seen": 3611648, + "step": 4735 + }, + { + "epoch": 9.854469854469855, + "grad_norm": 0.0004148696025367826, + "learning_rate": 0.28972942006541696, + "loss": 0.2744, + "num_input_tokens_seen": 3615488, + "step": 4740 + }, + { + "epoch": 9.864864864864865, + "grad_norm": 0.0014468338340520859, + "learning_rate": 0.2897079876050318, + "loss": 0.266, + "num_input_tokens_seen": 3619136, + "step": 4745 + }, + { + "epoch": 9.875259875259875, + "grad_norm": 8.109946065815166e-05, + "learning_rate": 0.2896865335999192, + "loss": 0.2658, + "num_input_tokens_seen": 3622976, + "step": 4750 + }, + { + "epoch": 9.885654885654885, + "grad_norm": 0.003306103404611349, + "learning_rate": 0.28966505805338777, + "loss": 0.2865, + "num_input_tokens_seen": 3626944, + "step": 4755 + }, + { + "epoch": 9.896049896049895, + "grad_norm": 0.0022545347455888987, + "learning_rate": 0.2896435609687492, + "loss": 0.2836, + "num_input_tokens_seen": 3630944, + "step": 4760 + }, + { + "epoch": 9.906444906444907, + "grad_norm": 0.0022805219050496817, + "learning_rate": 0.2896220423493187, + "loss": 0.2459, + "num_input_tokens_seen": 3634752, + "step": 4765 + }, + { + "epoch": 9.916839916839917, + "grad_norm": 0.00028122018557041883, + "learning_rate": 0.28960050219841466, + "loss": 0.2398, + "num_input_tokens_seen": 3638592, + "step": 4770 + }, + { + "epoch": 9.927234927234927, + "grad_norm": 0.0010744949104264379, + "learning_rate": 0.28957894051935884, + "loss": 0.253, + "num_input_tokens_seen": 3642464, + "step": 4775 + }, + { + "epoch": 9.937629937629938, + "grad_norm": 0.0008272434351965785, + "learning_rate": 0.2895573573154764, + "loss": 0.2655, + "num_input_tokens_seen": 3646336, + "step": 4780 + }, + { + "epoch": 9.948024948024948, + "grad_norm": 0.00025309371994808316, + "learning_rate": 0.28953575259009556, + "loss": 0.2646, + "num_input_tokens_seen": 3650048, + "step": 4785 + }, + { + "epoch": 9.958419958419958, + "grad_norm": 0.0006306581781245768, + "learning_rate": 0.2895141263465482, + "loss": 0.286, + "num_input_tokens_seen": 3653856, + "step": 4790 + }, + { + "epoch": 9.96881496881497, + "grad_norm": 0.0010191835463047028, + "learning_rate": 0.28949247858816934, + "loss": 0.2711, + "num_input_tokens_seen": 3657664, + "step": 4795 + }, + { + "epoch": 9.97920997920998, + "grad_norm": 0.00030977826099842787, + "learning_rate": 0.2894708093182973, + "loss": 0.249, + "num_input_tokens_seen": 3661568, + "step": 4800 + }, + { + "epoch": 9.97920997920998, + "eval_loss": 0.2519153952598572, + "eval_runtime": 13.4137, + "eval_samples_per_second": 63.815, + "eval_steps_per_second": 15.954, + "num_input_tokens_seen": 3661568, + "step": 4800 + }, + { + "epoch": 9.98960498960499, + "grad_norm": 0.0008100526756606996, + "learning_rate": 0.2894491185402737, + "loss": 0.2487, + "num_input_tokens_seen": 3665312, + "step": 4805 + }, + { + "epoch": 10.0, + "grad_norm": 0.0007240594713948667, + "learning_rate": 0.2894274062574437, + "loss": 0.268, + "num_input_tokens_seen": 3669168, + "step": 4810 + }, + { + "epoch": 10.01039501039501, + "grad_norm": 0.0007532486924901605, + "learning_rate": 0.2894056724731554, + "loss": 0.2423, + "num_input_tokens_seen": 3672976, + "step": 4815 + }, + { + "epoch": 10.02079002079002, + "grad_norm": 0.0018373773200437427, + "learning_rate": 0.28938391719076056, + "loss": 0.3322, + "num_input_tokens_seen": 3676848, + "step": 4820 + }, + { + "epoch": 10.03118503118503, + "grad_norm": 0.0018719711806625128, + "learning_rate": 0.28936214041361413, + "loss": 0.2777, + "num_input_tokens_seen": 3680624, + "step": 4825 + }, + { + "epoch": 10.041580041580042, + "grad_norm": 0.0002641514874994755, + "learning_rate": 0.2893403421450743, + "loss": 0.2711, + "num_input_tokens_seen": 3684528, + "step": 4830 + }, + { + "epoch": 10.051975051975052, + "grad_norm": 0.0002581313601695001, + "learning_rate": 0.2893185223885026, + "loss": 0.3069, + "num_input_tokens_seen": 3688368, + "step": 4835 + }, + { + "epoch": 10.062370062370062, + "grad_norm": 0.0007623252458870411, + "learning_rate": 0.289296681147264, + "loss": 0.3089, + "num_input_tokens_seen": 3692144, + "step": 4840 + }, + { + "epoch": 10.072765072765073, + "grad_norm": 0.00021265962277539074, + "learning_rate": 0.28927481842472663, + "loss": 0.3059, + "num_input_tokens_seen": 3695984, + "step": 4845 + }, + { + "epoch": 10.083160083160083, + "grad_norm": 0.0001783665647963062, + "learning_rate": 0.28925293422426207, + "loss": 0.2566, + "num_input_tokens_seen": 3699664, + "step": 4850 + }, + { + "epoch": 10.093555093555093, + "grad_norm": 0.0008891077013686299, + "learning_rate": 0.28923102854924504, + "loss": 0.2569, + "num_input_tokens_seen": 3703344, + "step": 4855 + }, + { + "epoch": 10.103950103950105, + "grad_norm": 0.001206408953294158, + "learning_rate": 0.2892091014030537, + "loss": 0.3077, + "num_input_tokens_seen": 3707248, + "step": 4860 + }, + { + "epoch": 10.114345114345115, + "grad_norm": 0.004202185198664665, + "learning_rate": 0.2891871527890696, + "loss": 0.3181, + "num_input_tokens_seen": 3710992, + "step": 4865 + }, + { + "epoch": 10.124740124740125, + "grad_norm": 0.0012446991167962551, + "learning_rate": 0.2891651827106773, + "loss": 0.2739, + "num_input_tokens_seen": 3714704, + "step": 4870 + }, + { + "epoch": 10.135135135135135, + "grad_norm": 0.00018531220848672092, + "learning_rate": 0.2891431911712651, + "loss": 0.2738, + "num_input_tokens_seen": 3718480, + "step": 4875 + }, + { + "epoch": 10.145530145530145, + "grad_norm": 0.00020502042025327682, + "learning_rate": 0.2891211781742241, + "loss": 0.296, + "num_input_tokens_seen": 3722320, + "step": 4880 + }, + { + "epoch": 10.155925155925155, + "grad_norm": 0.0010973821626976132, + "learning_rate": 0.2890991437229492, + "loss": 0.2642, + "num_input_tokens_seen": 3726128, + "step": 4885 + }, + { + "epoch": 10.166320166320165, + "grad_norm": 0.0007894599111750722, + "learning_rate": 0.2890770878208383, + "loss": 0.2794, + "num_input_tokens_seen": 3729872, + "step": 4890 + }, + { + "epoch": 10.176715176715177, + "grad_norm": 0.0010003134375438094, + "learning_rate": 0.28905501047129273, + "loss": 0.2636, + "num_input_tokens_seen": 3733776, + "step": 4895 + }, + { + "epoch": 10.187110187110187, + "grad_norm": 0.002415601396933198, + "learning_rate": 0.289032911677717, + "loss": 0.3135, + "num_input_tokens_seen": 3737616, + "step": 4900 + }, + { + "epoch": 10.197505197505198, + "grad_norm": 0.0018411775818094611, + "learning_rate": 0.28901079144351915, + "loss": 0.2818, + "num_input_tokens_seen": 3741328, + "step": 4905 + }, + { + "epoch": 10.207900207900208, + "grad_norm": 0.0011528501054272056, + "learning_rate": 0.2889886497721103, + "loss": 0.2914, + "num_input_tokens_seen": 3745136, + "step": 4910 + }, + { + "epoch": 10.218295218295218, + "grad_norm": 8.849740697769448e-05, + "learning_rate": 0.28896648666690505, + "loss": 0.2766, + "num_input_tokens_seen": 3748912, + "step": 4915 + }, + { + "epoch": 10.228690228690228, + "grad_norm": 0.00015786263975314796, + "learning_rate": 0.2889443021313212, + "loss": 0.2747, + "num_input_tokens_seen": 3752688, + "step": 4920 + }, + { + "epoch": 10.23908523908524, + "grad_norm": 0.0003860458673443645, + "learning_rate": 0.28892209616877984, + "loss": 0.2804, + "num_input_tokens_seen": 3756432, + "step": 4925 + }, + { + "epoch": 10.24948024948025, + "grad_norm": 0.0012466453481465578, + "learning_rate": 0.28889986878270546, + "loss": 0.2688, + "num_input_tokens_seen": 3760272, + "step": 4930 + }, + { + "epoch": 10.25987525987526, + "grad_norm": 0.001825141254812479, + "learning_rate": 0.28887761997652583, + "loss": 0.2721, + "num_input_tokens_seen": 3764144, + "step": 4935 + }, + { + "epoch": 10.27027027027027, + "grad_norm": 0.0011676604626700282, + "learning_rate": 0.2888553497536719, + "loss": 0.265, + "num_input_tokens_seen": 3767984, + "step": 4940 + }, + { + "epoch": 10.28066528066528, + "grad_norm": 0.00023977972159627825, + "learning_rate": 0.2888330581175781, + "loss": 0.258, + "num_input_tokens_seen": 3771760, + "step": 4945 + }, + { + "epoch": 10.29106029106029, + "grad_norm": 0.0003496152930893004, + "learning_rate": 0.28881074507168203, + "loss": 0.2628, + "num_input_tokens_seen": 3775664, + "step": 4950 + }, + { + "epoch": 10.301455301455302, + "grad_norm": 0.0007483397494070232, + "learning_rate": 0.2887884106194247, + "loss": 0.2933, + "num_input_tokens_seen": 3779408, + "step": 4955 + }, + { + "epoch": 10.311850311850312, + "grad_norm": 0.0020368245895951986, + "learning_rate": 0.28876605476425027, + "loss": 0.2879, + "num_input_tokens_seen": 3783248, + "step": 4960 + }, + { + "epoch": 10.322245322245323, + "grad_norm": 0.0022097816690802574, + "learning_rate": 0.2887436775096064, + "loss": 0.3185, + "num_input_tokens_seen": 3787024, + "step": 4965 + }, + { + "epoch": 10.332640332640333, + "grad_norm": 0.0009934940608218312, + "learning_rate": 0.2887212788589439, + "loss": 0.2942, + "num_input_tokens_seen": 3790800, + "step": 4970 + }, + { + "epoch": 10.343035343035343, + "grad_norm": 0.0009917601710185409, + "learning_rate": 0.2886988588157169, + "loss": 0.2874, + "num_input_tokens_seen": 3794672, + "step": 4975 + }, + { + "epoch": 10.353430353430353, + "grad_norm": 0.0005199462757445872, + "learning_rate": 0.28867641738338284, + "loss": 0.2899, + "num_input_tokens_seen": 3798352, + "step": 4980 + }, + { + "epoch": 10.363825363825363, + "grad_norm": 0.0004383198684081435, + "learning_rate": 0.2886539545654026, + "loss": 0.2463, + "num_input_tokens_seen": 3802000, + "step": 4985 + }, + { + "epoch": 10.374220374220375, + "grad_norm": 0.001138963270932436, + "learning_rate": 0.28863147036524006, + "loss": 0.2847, + "num_input_tokens_seen": 3805840, + "step": 4990 + }, + { + "epoch": 10.384615384615385, + "grad_norm": 0.00042424467392265797, + "learning_rate": 0.2886089647863626, + "loss": 0.2859, + "num_input_tokens_seen": 3809680, + "step": 4995 + }, + { + "epoch": 10.395010395010395, + "grad_norm": 0.000569929659832269, + "learning_rate": 0.288586437832241, + "loss": 0.3018, + "num_input_tokens_seen": 3813552, + "step": 5000 + }, + { + "epoch": 10.395010395010395, + "eval_loss": 0.27405840158462524, + "eval_runtime": 13.4322, + "eval_samples_per_second": 63.727, + "eval_steps_per_second": 15.932, + "num_input_tokens_seen": 3813552, + "step": 5000 + }, + { + "epoch": 10.405405405405405, + "grad_norm": 0.0006497822469100356, + "learning_rate": 0.28856388950634904, + "loss": 0.2645, + "num_input_tokens_seen": 3817328, + "step": 5005 + }, + { + "epoch": 10.415800415800415, + "grad_norm": 0.0019462925847619772, + "learning_rate": 0.288541319812164, + "loss": 0.3245, + "num_input_tokens_seen": 3821136, + "step": 5010 + }, + { + "epoch": 10.426195426195425, + "grad_norm": 0.0034500728361308575, + "learning_rate": 0.2885187287531665, + "loss": 0.3026, + "num_input_tokens_seen": 3825168, + "step": 5015 + }, + { + "epoch": 10.436590436590437, + "grad_norm": 0.0009363286662846804, + "learning_rate": 0.2884961163328402, + "loss": 0.2867, + "num_input_tokens_seen": 3828848, + "step": 5020 + }, + { + "epoch": 10.446985446985448, + "grad_norm": 0.0009797376114875078, + "learning_rate": 0.28847348255467237, + "loss": 0.3134, + "num_input_tokens_seen": 3832720, + "step": 5025 + }, + { + "epoch": 10.457380457380458, + "grad_norm": 0.0008015984203666449, + "learning_rate": 0.28845082742215333, + "loss": 0.2627, + "num_input_tokens_seen": 3836528, + "step": 5030 + }, + { + "epoch": 10.467775467775468, + "grad_norm": 0.0001685633324086666, + "learning_rate": 0.2884281509387769, + "loss": 0.274, + "num_input_tokens_seen": 3840304, + "step": 5035 + }, + { + "epoch": 10.478170478170478, + "grad_norm": 0.0004156519134994596, + "learning_rate": 0.2884054531080399, + "loss": 0.2737, + "num_input_tokens_seen": 3844240, + "step": 5040 + }, + { + "epoch": 10.488565488565488, + "grad_norm": 0.0005527977482415736, + "learning_rate": 0.28838273393344277, + "loss": 0.2226, + "num_input_tokens_seen": 3848080, + "step": 5045 + }, + { + "epoch": 10.4989604989605, + "grad_norm": 0.0012770200846716762, + "learning_rate": 0.288359993418489, + "loss": 0.3116, + "num_input_tokens_seen": 3851856, + "step": 5050 + }, + { + "epoch": 10.50935550935551, + "grad_norm": 0.0024924289900809526, + "learning_rate": 0.28833723156668556, + "loss": 0.3031, + "num_input_tokens_seen": 3855760, + "step": 5055 + }, + { + "epoch": 10.51975051975052, + "grad_norm": 0.0009867752669379115, + "learning_rate": 0.2883144483815425, + "loss": 0.294, + "num_input_tokens_seen": 3859664, + "step": 5060 + }, + { + "epoch": 10.53014553014553, + "grad_norm": 0.0009886941406875849, + "learning_rate": 0.28829164386657335, + "loss": 0.2859, + "num_input_tokens_seen": 3863568, + "step": 5065 + }, + { + "epoch": 10.54054054054054, + "grad_norm": 0.0011199767468497157, + "learning_rate": 0.28826881802529486, + "loss": 0.2485, + "num_input_tokens_seen": 3867408, + "step": 5070 + }, + { + "epoch": 10.55093555093555, + "grad_norm": 0.00019102166697848588, + "learning_rate": 0.28824597086122705, + "loss": 0.2301, + "num_input_tokens_seen": 3871152, + "step": 5075 + }, + { + "epoch": 10.56133056133056, + "grad_norm": 0.00044801447074860334, + "learning_rate": 0.28822310237789317, + "loss": 0.3111, + "num_input_tokens_seen": 3874896, + "step": 5080 + }, + { + "epoch": 10.571725571725572, + "grad_norm": 0.0004440463671926409, + "learning_rate": 0.2882002125788199, + "loss": 0.2822, + "num_input_tokens_seen": 3878640, + "step": 5085 + }, + { + "epoch": 10.582120582120583, + "grad_norm": 0.0011579656274989247, + "learning_rate": 0.2881773014675371, + "loss": 0.2793, + "num_input_tokens_seen": 3882416, + "step": 5090 + }, + { + "epoch": 10.592515592515593, + "grad_norm": 0.0009008412598632276, + "learning_rate": 0.288154369047578, + "loss": 0.2753, + "num_input_tokens_seen": 3886288, + "step": 5095 + }, + { + "epoch": 10.602910602910603, + "grad_norm": 0.0005793485324829817, + "learning_rate": 0.28813141532247905, + "loss": 0.2861, + "num_input_tokens_seen": 3890192, + "step": 5100 + }, + { + "epoch": 10.613305613305613, + "grad_norm": 0.00036320521030575037, + "learning_rate": 0.28810844029578, + "loss": 0.2742, + "num_input_tokens_seen": 3894096, + "step": 5105 + }, + { + "epoch": 10.623700623700623, + "grad_norm": 0.0015469376230612397, + "learning_rate": 0.2880854439710238, + "loss": 0.2659, + "num_input_tokens_seen": 3897904, + "step": 5110 + }, + { + "epoch": 10.634095634095633, + "grad_norm": 0.0001771929964888841, + "learning_rate": 0.28806242635175694, + "loss": 0.2747, + "num_input_tokens_seen": 3901776, + "step": 5115 + }, + { + "epoch": 10.644490644490645, + "grad_norm": 0.0004131075693294406, + "learning_rate": 0.2880393874415289, + "loss": 0.2845, + "num_input_tokens_seen": 3905552, + "step": 5120 + }, + { + "epoch": 10.654885654885655, + "grad_norm": 0.0006141449557617307, + "learning_rate": 0.2880163272438926, + "loss": 0.2829, + "num_input_tokens_seen": 3909360, + "step": 5125 + }, + { + "epoch": 10.665280665280665, + "grad_norm": 0.0007640831172466278, + "learning_rate": 0.2879932457624042, + "loss": 0.258, + "num_input_tokens_seen": 3913200, + "step": 5130 + }, + { + "epoch": 10.675675675675675, + "grad_norm": 0.0011102590942755342, + "learning_rate": 0.2879701430006232, + "loss": 0.2748, + "num_input_tokens_seen": 3917104, + "step": 5135 + }, + { + "epoch": 10.686070686070686, + "grad_norm": 0.0016640633111819625, + "learning_rate": 0.28794701896211233, + "loss": 0.287, + "num_input_tokens_seen": 3920944, + "step": 5140 + }, + { + "epoch": 10.696465696465696, + "grad_norm": 0.0001840692275436595, + "learning_rate": 0.28792387365043753, + "loss": 0.2702, + "num_input_tokens_seen": 3924656, + "step": 5145 + }, + { + "epoch": 10.706860706860708, + "grad_norm": 0.0005854679620824754, + "learning_rate": 0.28790070706916815, + "loss": 0.3082, + "num_input_tokens_seen": 3928464, + "step": 5150 + }, + { + "epoch": 10.717255717255718, + "grad_norm": 0.0020343204960227013, + "learning_rate": 0.2878775192218768, + "loss": 0.2626, + "num_input_tokens_seen": 3932592, + "step": 5155 + }, + { + "epoch": 10.727650727650728, + "grad_norm": 0.000736934831365943, + "learning_rate": 0.2878543101121393, + "loss": 0.2649, + "num_input_tokens_seen": 3936464, + "step": 5160 + }, + { + "epoch": 10.738045738045738, + "grad_norm": 0.00013142768875695765, + "learning_rate": 0.28783107974353483, + "loss": 0.256, + "num_input_tokens_seen": 3940304, + "step": 5165 + }, + { + "epoch": 10.748440748440748, + "grad_norm": 0.0011673502158373594, + "learning_rate": 0.2878078281196457, + "loss": 0.2795, + "num_input_tokens_seen": 3944048, + "step": 5170 + }, + { + "epoch": 10.758835758835758, + "grad_norm": 0.001666890224441886, + "learning_rate": 0.28778455524405777, + "loss": 0.2791, + "num_input_tokens_seen": 3947824, + "step": 5175 + }, + { + "epoch": 10.76923076923077, + "grad_norm": 0.0003241396916564554, + "learning_rate": 0.2877612611203598, + "loss": 0.2596, + "num_input_tokens_seen": 3951568, + "step": 5180 + }, + { + "epoch": 10.77962577962578, + "grad_norm": 0.00018004632147494704, + "learning_rate": 0.28773794575214423, + "loss": 0.27, + "num_input_tokens_seen": 3955408, + "step": 5185 + }, + { + "epoch": 10.79002079002079, + "grad_norm": 0.0014246640494093299, + "learning_rate": 0.28771460914300645, + "loss": 0.2948, + "num_input_tokens_seen": 3959280, + "step": 5190 + }, + { + "epoch": 10.8004158004158, + "grad_norm": 0.00027323196991346776, + "learning_rate": 0.2876912512965454, + "loss": 0.2778, + "num_input_tokens_seen": 3963216, + "step": 5195 + }, + { + "epoch": 10.81081081081081, + "grad_norm": 0.0006623450899496675, + "learning_rate": 0.287667872216363, + "loss": 0.2871, + "num_input_tokens_seen": 3967024, + "step": 5200 + }, + { + "epoch": 10.81081081081081, + "eval_loss": 0.2504214942455292, + "eval_runtime": 13.433, + "eval_samples_per_second": 63.723, + "eval_steps_per_second": 15.931, + "num_input_tokens_seen": 3967024, + "step": 5200 + }, + { + "epoch": 10.82120582120582, + "grad_norm": 0.0001593843480804935, + "learning_rate": 0.2876444719060647, + "loss": 0.2564, + "num_input_tokens_seen": 3970896, + "step": 5205 + }, + { + "epoch": 10.83160083160083, + "grad_norm": 0.002191289095208049, + "learning_rate": 0.287621050369259, + "loss": 0.2633, + "num_input_tokens_seen": 3974832, + "step": 5210 + }, + { + "epoch": 10.841995841995843, + "grad_norm": 0.00039982551243156195, + "learning_rate": 0.28759760760955794, + "loss": 0.2646, + "num_input_tokens_seen": 3978704, + "step": 5215 + }, + { + "epoch": 10.852390852390853, + "grad_norm": 0.00032265440677292645, + "learning_rate": 0.2875741436305766, + "loss": 0.2671, + "num_input_tokens_seen": 3982448, + "step": 5220 + }, + { + "epoch": 10.862785862785863, + "grad_norm": 0.0006903450121171772, + "learning_rate": 0.28755065843593347, + "loss": 0.2647, + "num_input_tokens_seen": 3986160, + "step": 5225 + }, + { + "epoch": 10.873180873180873, + "grad_norm": 0.0008125486201606691, + "learning_rate": 0.2875271520292502, + "loss": 0.2784, + "num_input_tokens_seen": 3990000, + "step": 5230 + }, + { + "epoch": 10.883575883575883, + "grad_norm": 0.0006441565928980708, + "learning_rate": 0.28750362441415184, + "loss": 0.2813, + "num_input_tokens_seen": 3993744, + "step": 5235 + }, + { + "epoch": 10.893970893970893, + "grad_norm": 0.0004949583671987057, + "learning_rate": 0.28748007559426664, + "loss": 0.2646, + "num_input_tokens_seen": 3997552, + "step": 5240 + }, + { + "epoch": 10.904365904365905, + "grad_norm": 0.00015367277956102043, + "learning_rate": 0.2874565055732261, + "loss": 0.2807, + "num_input_tokens_seen": 4001392, + "step": 5245 + }, + { + "epoch": 10.914760914760915, + "grad_norm": 0.0006494914414361119, + "learning_rate": 0.28743291435466495, + "loss": 0.2587, + "num_input_tokens_seen": 4005232, + "step": 5250 + }, + { + "epoch": 10.925155925155925, + "grad_norm": 0.0008468227460980415, + "learning_rate": 0.2874093019422214, + "loss": 0.2684, + "num_input_tokens_seen": 4009008, + "step": 5255 + }, + { + "epoch": 10.935550935550935, + "grad_norm": 0.00025040569016709924, + "learning_rate": 0.28738566833953666, + "loss": 0.2502, + "num_input_tokens_seen": 4012848, + "step": 5260 + }, + { + "epoch": 10.945945945945946, + "grad_norm": 0.00017516358639113605, + "learning_rate": 0.28736201355025537, + "loss": 0.279, + "num_input_tokens_seen": 4016560, + "step": 5265 + }, + { + "epoch": 10.956340956340956, + "grad_norm": 0.0001797443110262975, + "learning_rate": 0.28733833757802535, + "loss": 0.2495, + "num_input_tokens_seen": 4020400, + "step": 5270 + }, + { + "epoch": 10.966735966735968, + "grad_norm": 8.787220576778054e-05, + "learning_rate": 0.28731464042649785, + "loss": 0.2736, + "num_input_tokens_seen": 4024144, + "step": 5275 + }, + { + "epoch": 10.977130977130978, + "grad_norm": 0.0016055147862061858, + "learning_rate": 0.2872909220993271, + "loss": 0.2709, + "num_input_tokens_seen": 4027920, + "step": 5280 + }, + { + "epoch": 10.987525987525988, + "grad_norm": 0.002014842117205262, + "learning_rate": 0.287267182600171, + "loss": 0.2791, + "num_input_tokens_seen": 4031664, + "step": 5285 + }, + { + "epoch": 10.997920997920998, + "grad_norm": 0.0012715302873402834, + "learning_rate": 0.2872434219326902, + "loss": 0.2464, + "num_input_tokens_seen": 4035408, + "step": 5290 + }, + { + "epoch": 11.008316008316008, + "grad_norm": 0.0002326093817828223, + "learning_rate": 0.28721964010054907, + "loss": 0.2573, + "num_input_tokens_seen": 4039168, + "step": 5295 + }, + { + "epoch": 11.018711018711018, + "grad_norm": 0.0003226780681870878, + "learning_rate": 0.28719583710741503, + "loss": 0.2841, + "num_input_tokens_seen": 4043008, + "step": 5300 + }, + { + "epoch": 11.029106029106028, + "grad_norm": 0.0005045076250098646, + "learning_rate": 0.28717201295695877, + "loss": 0.2898, + "num_input_tokens_seen": 4046880, + "step": 5305 + }, + { + "epoch": 11.03950103950104, + "grad_norm": 0.00042858265805989504, + "learning_rate": 0.28714816765285434, + "loss": 0.2714, + "num_input_tokens_seen": 4050720, + "step": 5310 + }, + { + "epoch": 11.04989604989605, + "grad_norm": 0.000739100796636194, + "learning_rate": 0.28712430119877896, + "loss": 0.2866, + "num_input_tokens_seen": 4054560, + "step": 5315 + }, + { + "epoch": 11.06029106029106, + "grad_norm": 0.0007063311641104519, + "learning_rate": 0.28710041359841304, + "loss": 0.2742, + "num_input_tokens_seen": 4058304, + "step": 5320 + }, + { + "epoch": 11.07068607068607, + "grad_norm": 0.0016039853217080235, + "learning_rate": 0.28707650485544056, + "loss": 0.2628, + "num_input_tokens_seen": 4062272, + "step": 5325 + }, + { + "epoch": 11.08108108108108, + "grad_norm": 0.0008827102719806135, + "learning_rate": 0.28705257497354836, + "loss": 0.2765, + "num_input_tokens_seen": 4066112, + "step": 5330 + }, + { + "epoch": 11.09147609147609, + "grad_norm": 0.0001345526980003342, + "learning_rate": 0.28702862395642675, + "loss": 0.2678, + "num_input_tokens_seen": 4069952, + "step": 5335 + }, + { + "epoch": 11.101871101871103, + "grad_norm": 0.0012599313631653786, + "learning_rate": 0.28700465180776935, + "loss": 0.2699, + "num_input_tokens_seen": 4073824, + "step": 5340 + }, + { + "epoch": 11.112266112266113, + "grad_norm": 0.0002967861364595592, + "learning_rate": 0.2869806585312729, + "loss": 0.2695, + "num_input_tokens_seen": 4077664, + "step": 5345 + }, + { + "epoch": 11.122661122661123, + "grad_norm": 0.001670153345912695, + "learning_rate": 0.28695664413063754, + "loss": 0.2905, + "num_input_tokens_seen": 4081600, + "step": 5350 + }, + { + "epoch": 11.133056133056133, + "grad_norm": 0.0021830438636243343, + "learning_rate": 0.28693260860956654, + "loss": 0.3019, + "num_input_tokens_seen": 4085472, + "step": 5355 + }, + { + "epoch": 11.143451143451143, + "grad_norm": 0.0011994813103228807, + "learning_rate": 0.2869085519717665, + "loss": 0.2787, + "num_input_tokens_seen": 4089280, + "step": 5360 + }, + { + "epoch": 11.153846153846153, + "grad_norm": 0.00024140537425410002, + "learning_rate": 0.28688447422094726, + "loss": 0.2739, + "num_input_tokens_seen": 4093056, + "step": 5365 + }, + { + "epoch": 11.164241164241163, + "grad_norm": 0.0010804218472912908, + "learning_rate": 0.2868603753608219, + "loss": 0.2628, + "num_input_tokens_seen": 4096928, + "step": 5370 + }, + { + "epoch": 11.174636174636175, + "grad_norm": 0.00030141184106469154, + "learning_rate": 0.28683625539510665, + "loss": 0.2738, + "num_input_tokens_seen": 4100768, + "step": 5375 + }, + { + "epoch": 11.185031185031185, + "grad_norm": 0.0006643576198257506, + "learning_rate": 0.28681211432752135, + "loss": 0.2672, + "num_input_tokens_seen": 4104608, + "step": 5380 + }, + { + "epoch": 11.195426195426196, + "grad_norm": 0.0011604147730395198, + "learning_rate": 0.2867879521617887, + "loss": 0.2773, + "num_input_tokens_seen": 4108320, + "step": 5385 + }, + { + "epoch": 11.205821205821206, + "grad_norm": 0.0005744689260609448, + "learning_rate": 0.28676376890163485, + "loss": 0.2748, + "num_input_tokens_seen": 4112224, + "step": 5390 + }, + { + "epoch": 11.216216216216216, + "grad_norm": 0.0015742257237434387, + "learning_rate": 0.2867395645507891, + "loss": 0.2811, + "num_input_tokens_seen": 4116032, + "step": 5395 + }, + { + "epoch": 11.226611226611226, + "grad_norm": 0.0003916473069693893, + "learning_rate": 0.2867153391129842, + "loss": 0.2634, + "num_input_tokens_seen": 4120032, + "step": 5400 + }, + { + "epoch": 11.226611226611226, + "eval_loss": 0.24936148524284363, + "eval_runtime": 13.4616, + "eval_samples_per_second": 63.589, + "eval_steps_per_second": 15.897, + "num_input_tokens_seen": 4120032, + "step": 5400 + }, + { + "epoch": 11.237006237006238, + "grad_norm": 0.0012456130934879184, + "learning_rate": 0.28669109259195585, + "loss": 0.2167, + "num_input_tokens_seen": 4123680, + "step": 5405 + }, + { + "epoch": 11.247401247401248, + "grad_norm": 0.000903849839232862, + "learning_rate": 0.2866668249914433, + "loss": 0.2951, + "num_input_tokens_seen": 4127712, + "step": 5410 + }, + { + "epoch": 11.257796257796258, + "grad_norm": 0.0006740223034285009, + "learning_rate": 0.2866425363151889, + "loss": 0.2703, + "num_input_tokens_seen": 4131360, + "step": 5415 + }, + { + "epoch": 11.268191268191268, + "grad_norm": 0.00018547267245594412, + "learning_rate": 0.2866182265669382, + "loss": 0.2725, + "num_input_tokens_seen": 4135200, + "step": 5420 + }, + { + "epoch": 11.278586278586278, + "grad_norm": 0.0008092054631561041, + "learning_rate": 0.28659389575044014, + "loss": 0.2718, + "num_input_tokens_seen": 4139040, + "step": 5425 + }, + { + "epoch": 11.288981288981288, + "grad_norm": 0.0010923068039119244, + "learning_rate": 0.28656954386944683, + "loss": 0.296, + "num_input_tokens_seen": 4142880, + "step": 5430 + }, + { + "epoch": 11.299376299376299, + "grad_norm": 9.251588198821992e-05, + "learning_rate": 0.28654517092771353, + "loss": 0.2951, + "num_input_tokens_seen": 4146496, + "step": 5435 + }, + { + "epoch": 11.30977130977131, + "grad_norm": 0.0004181528929620981, + "learning_rate": 0.286520776928999, + "loss": 0.2586, + "num_input_tokens_seen": 4150272, + "step": 5440 + }, + { + "epoch": 11.32016632016632, + "grad_norm": 0.0017176421824842691, + "learning_rate": 0.286496361877065, + "loss": 0.2982, + "num_input_tokens_seen": 4154080, + "step": 5445 + }, + { + "epoch": 11.33056133056133, + "grad_norm": 0.001848685904406011, + "learning_rate": 0.28647192577567676, + "loss": 0.2637, + "num_input_tokens_seen": 4157984, + "step": 5450 + }, + { + "epoch": 11.34095634095634, + "grad_norm": 0.0011772282887250185, + "learning_rate": 0.28644746862860254, + "loss": 0.2825, + "num_input_tokens_seen": 4161792, + "step": 5455 + }, + { + "epoch": 11.35135135135135, + "grad_norm": 0.0002183064352720976, + "learning_rate": 0.2864229904396139, + "loss": 0.2804, + "num_input_tokens_seen": 4165600, + "step": 5460 + }, + { + "epoch": 11.361746361746361, + "grad_norm": 0.0003906710771843791, + "learning_rate": 0.28639849121248573, + "loss": 0.2626, + "num_input_tokens_seen": 4169440, + "step": 5465 + }, + { + "epoch": 11.372141372141373, + "grad_norm": 0.0009167538373731077, + "learning_rate": 0.28637397095099615, + "loss": 0.2665, + "num_input_tokens_seen": 4173280, + "step": 5470 + }, + { + "epoch": 11.382536382536383, + "grad_norm": 0.000585370697081089, + "learning_rate": 0.28634942965892646, + "loss": 0.2475, + "num_input_tokens_seen": 4177120, + "step": 5475 + }, + { + "epoch": 11.392931392931393, + "grad_norm": 0.0013894661096855998, + "learning_rate": 0.28632486734006124, + "loss": 0.2801, + "num_input_tokens_seen": 4180960, + "step": 5480 + }, + { + "epoch": 11.403326403326403, + "grad_norm": 0.0002715508744586259, + "learning_rate": 0.28630028399818835, + "loss": 0.2564, + "num_input_tokens_seen": 4184832, + "step": 5485 + }, + { + "epoch": 11.413721413721413, + "grad_norm": 0.0013098849449306726, + "learning_rate": 0.2862756796370987, + "loss": 0.2695, + "num_input_tokens_seen": 4188736, + "step": 5490 + }, + { + "epoch": 11.424116424116423, + "grad_norm": 9.128142846748233e-05, + "learning_rate": 0.2862510542605868, + "loss": 0.2454, + "num_input_tokens_seen": 4192384, + "step": 5495 + }, + { + "epoch": 11.434511434511435, + "grad_norm": 0.0007413947605527937, + "learning_rate": 0.2862264078724501, + "loss": 0.2604, + "num_input_tokens_seen": 4196256, + "step": 5500 + }, + { + "epoch": 11.444906444906445, + "grad_norm": 0.002100604586303234, + "learning_rate": 0.28620174047648933, + "loss": 0.2803, + "num_input_tokens_seen": 4199904, + "step": 5505 + }, + { + "epoch": 11.455301455301456, + "grad_norm": 0.0005267172819003463, + "learning_rate": 0.2861770520765086, + "loss": 0.2407, + "num_input_tokens_seen": 4203648, + "step": 5510 + }, + { + "epoch": 11.465696465696466, + "grad_norm": 0.0012598353205248713, + "learning_rate": 0.2861523426763151, + "loss": 0.3304, + "num_input_tokens_seen": 4207488, + "step": 5515 + }, + { + "epoch": 11.476091476091476, + "grad_norm": 0.0008330654818564653, + "learning_rate": 0.2861276122797194, + "loss": 0.2494, + "num_input_tokens_seen": 4211200, + "step": 5520 + }, + { + "epoch": 11.486486486486486, + "grad_norm": 0.0005036956281401217, + "learning_rate": 0.28610286089053516, + "loss": 0.2601, + "num_input_tokens_seen": 4215072, + "step": 5525 + }, + { + "epoch": 11.496881496881496, + "grad_norm": 0.00010328643838874996, + "learning_rate": 0.28607808851257943, + "loss": 0.2409, + "num_input_tokens_seen": 4218944, + "step": 5530 + }, + { + "epoch": 11.507276507276508, + "grad_norm": 0.0005854734918102622, + "learning_rate": 0.28605329514967237, + "loss": 0.2761, + "num_input_tokens_seen": 4222816, + "step": 5535 + }, + { + "epoch": 11.517671517671518, + "grad_norm": 0.0015499501023441553, + "learning_rate": 0.2860284808056374, + "loss": 0.2498, + "num_input_tokens_seen": 4226688, + "step": 5540 + }, + { + "epoch": 11.528066528066528, + "grad_norm": 0.0008388102287426591, + "learning_rate": 0.28600364548430135, + "loss": 0.2873, + "num_input_tokens_seen": 4230496, + "step": 5545 + }, + { + "epoch": 11.538461538461538, + "grad_norm": 0.0006467925850301981, + "learning_rate": 0.28597878918949393, + "loss": 0.2689, + "num_input_tokens_seen": 4234272, + "step": 5550 + }, + { + "epoch": 11.548856548856548, + "grad_norm": 7.578585064038634e-05, + "learning_rate": 0.2859539119250485, + "loss": 0.2648, + "num_input_tokens_seen": 4238272, + "step": 5555 + }, + { + "epoch": 11.559251559251559, + "grad_norm": 0.0008321469649672508, + "learning_rate": 0.2859290136948013, + "loss": 0.2671, + "num_input_tokens_seen": 4241920, + "step": 5560 + }, + { + "epoch": 11.56964656964657, + "grad_norm": 0.000922494160477072, + "learning_rate": 0.28590409450259197, + "loss": 0.2747, + "num_input_tokens_seen": 4245632, + "step": 5565 + }, + { + "epoch": 11.58004158004158, + "grad_norm": 0.0004655584925785661, + "learning_rate": 0.28587915435226346, + "loss": 0.271, + "num_input_tokens_seen": 4249472, + "step": 5570 + }, + { + "epoch": 11.59043659043659, + "grad_norm": 0.001390872523188591, + "learning_rate": 0.2858541932476617, + "loss": 0.2483, + "num_input_tokens_seen": 4253376, + "step": 5575 + }, + { + "epoch": 11.6008316008316, + "grad_norm": 0.0007518345373682678, + "learning_rate": 0.2858292111926361, + "loss": 0.3153, + "num_input_tokens_seen": 4257152, + "step": 5580 + }, + { + "epoch": 11.611226611226611, + "grad_norm": 0.0001840638869907707, + "learning_rate": 0.28580420819103924, + "loss": 0.2759, + "num_input_tokens_seen": 4260960, + "step": 5585 + }, + { + "epoch": 11.621621621621621, + "grad_norm": 0.0008532400825060904, + "learning_rate": 0.2857791842467269, + "loss": 0.2764, + "num_input_tokens_seen": 4264832, + "step": 5590 + }, + { + "epoch": 11.632016632016633, + "grad_norm": 0.00016213746857829392, + "learning_rate": 0.2857541393635579, + "loss": 0.2697, + "num_input_tokens_seen": 4268704, + "step": 5595 + }, + { + "epoch": 11.642411642411643, + "grad_norm": 0.0005190189694985747, + "learning_rate": 0.2857290735453948, + "loss": 0.2821, + "num_input_tokens_seen": 4272608, + "step": 5600 + }, + { + "epoch": 11.642411642411643, + "eval_loss": 0.2751775085926056, + "eval_runtime": 13.4019, + "eval_samples_per_second": 63.872, + "eval_steps_per_second": 15.968, + "num_input_tokens_seen": 4272608, + "step": 5600 + }, + { + "epoch": 11.652806652806653, + "grad_norm": 0.0018004627199843526, + "learning_rate": 0.28570398679610276, + "loss": 0.2563, + "num_input_tokens_seen": 4276480, + "step": 5605 + }, + { + "epoch": 11.663201663201663, + "grad_norm": 0.0011417015921324492, + "learning_rate": 0.2856788791195506, + "loss": 0.2325, + "num_input_tokens_seen": 4280160, + "step": 5610 + }, + { + "epoch": 11.673596673596673, + "grad_norm": 0.0016824102494865656, + "learning_rate": 0.28565375051961023, + "loss": 0.3769, + "num_input_tokens_seen": 4284032, + "step": 5615 + }, + { + "epoch": 11.683991683991684, + "grad_norm": 0.0011130665661767125, + "learning_rate": 0.28562860100015686, + "loss": 0.2992, + "num_input_tokens_seen": 4288000, + "step": 5620 + }, + { + "epoch": 11.694386694386694, + "grad_norm": 0.0013533191522583365, + "learning_rate": 0.2856034305650687, + "loss": 0.3112, + "num_input_tokens_seen": 4291776, + "step": 5625 + }, + { + "epoch": 11.704781704781706, + "grad_norm": 0.0015348048182204366, + "learning_rate": 0.28557823921822756, + "loss": 0.2866, + "num_input_tokens_seen": 4295552, + "step": 5630 + }, + { + "epoch": 11.715176715176716, + "grad_norm": 0.002123083220794797, + "learning_rate": 0.2855530269635181, + "loss": 0.2612, + "num_input_tokens_seen": 4299232, + "step": 5635 + }, + { + "epoch": 11.725571725571726, + "grad_norm": 0.000828628137242049, + "learning_rate": 0.2855277938048284, + "loss": 0.2684, + "num_input_tokens_seen": 4303168, + "step": 5640 + }, + { + "epoch": 11.735966735966736, + "grad_norm": 0.00045358893112279475, + "learning_rate": 0.2855025397460498, + "loss": 0.2631, + "num_input_tokens_seen": 4306848, + "step": 5645 + }, + { + "epoch": 11.746361746361746, + "grad_norm": 0.001199671532958746, + "learning_rate": 0.28547726479107666, + "loss": 0.2802, + "num_input_tokens_seen": 4310560, + "step": 5650 + }, + { + "epoch": 11.756756756756756, + "grad_norm": 0.00013404253695625812, + "learning_rate": 0.2854519689438068, + "loss": 0.2811, + "num_input_tokens_seen": 4314272, + "step": 5655 + }, + { + "epoch": 11.767151767151766, + "grad_norm": 0.00024221248168032616, + "learning_rate": 0.2854266522081412, + "loss": 0.2832, + "num_input_tokens_seen": 4318016, + "step": 5660 + }, + { + "epoch": 11.777546777546778, + "grad_norm": 0.0004617544182110578, + "learning_rate": 0.28540131458798385, + "loss": 0.2517, + "num_input_tokens_seen": 4321760, + "step": 5665 + }, + { + "epoch": 11.787941787941788, + "grad_norm": 0.0011133091757073998, + "learning_rate": 0.28537595608724226, + "loss": 0.2815, + "num_input_tokens_seen": 4325696, + "step": 5670 + }, + { + "epoch": 11.798336798336798, + "grad_norm": 0.00035040912916883826, + "learning_rate": 0.28535057670982705, + "loss": 0.2786, + "num_input_tokens_seen": 4329376, + "step": 5675 + }, + { + "epoch": 11.808731808731808, + "grad_norm": 0.0004917478654533625, + "learning_rate": 0.285325176459652, + "loss": 0.2829, + "num_input_tokens_seen": 4333184, + "step": 5680 + }, + { + "epoch": 11.819126819126819, + "grad_norm": 0.0002756681642495096, + "learning_rate": 0.28529975534063406, + "loss": 0.275, + "num_input_tokens_seen": 4337120, + "step": 5685 + }, + { + "epoch": 11.829521829521829, + "grad_norm": 0.0005553779192268848, + "learning_rate": 0.2852743133566936, + "loss": 0.2245, + "num_input_tokens_seen": 4340832, + "step": 5690 + }, + { + "epoch": 11.83991683991684, + "grad_norm": 0.001064133713953197, + "learning_rate": 0.2852488505117541, + "loss": 0.3045, + "num_input_tokens_seen": 4344672, + "step": 5695 + }, + { + "epoch": 11.85031185031185, + "grad_norm": 0.0018903706222772598, + "learning_rate": 0.28522336680974214, + "loss": 0.2704, + "num_input_tokens_seen": 4348416, + "step": 5700 + }, + { + "epoch": 11.86070686070686, + "grad_norm": 0.00017086938896682113, + "learning_rate": 0.2851978622545877, + "loss": 0.2704, + "num_input_tokens_seen": 4352192, + "step": 5705 + }, + { + "epoch": 11.871101871101871, + "grad_norm": 0.001326584373600781, + "learning_rate": 0.285172336850224, + "loss": 0.2706, + "num_input_tokens_seen": 4356032, + "step": 5710 + }, + { + "epoch": 11.881496881496881, + "grad_norm": 0.0006213866290636361, + "learning_rate": 0.2851467906005871, + "loss": 0.2734, + "num_input_tokens_seen": 4359872, + "step": 5715 + }, + { + "epoch": 11.891891891891891, + "grad_norm": 0.00016239417891483754, + "learning_rate": 0.28512122350961683, + "loss": 0.2764, + "num_input_tokens_seen": 4363712, + "step": 5720 + }, + { + "epoch": 11.902286902286903, + "grad_norm": 0.0001600651303306222, + "learning_rate": 0.2850956355812559, + "loss": 0.2657, + "num_input_tokens_seen": 4367552, + "step": 5725 + }, + { + "epoch": 11.912681912681913, + "grad_norm": 0.00046010466758161783, + "learning_rate": 0.28507002681945015, + "loss": 0.2651, + "num_input_tokens_seen": 4371264, + "step": 5730 + }, + { + "epoch": 11.923076923076923, + "grad_norm": 0.0014085611328482628, + "learning_rate": 0.28504439722814895, + "loss": 0.2759, + "num_input_tokens_seen": 4375008, + "step": 5735 + }, + { + "epoch": 11.933471933471933, + "grad_norm": 0.00039940845454111695, + "learning_rate": 0.28501874681130457, + "loss": 0.2765, + "num_input_tokens_seen": 4378976, + "step": 5740 + }, + { + "epoch": 11.943866943866944, + "grad_norm": 0.00048410348244942725, + "learning_rate": 0.2849930755728727, + "loss": 0.2868, + "num_input_tokens_seen": 4382624, + "step": 5745 + }, + { + "epoch": 11.954261954261954, + "grad_norm": 0.0006358897080644965, + "learning_rate": 0.28496738351681217, + "loss": 0.2685, + "num_input_tokens_seen": 4386624, + "step": 5750 + }, + { + "epoch": 11.964656964656964, + "grad_norm": 0.00012058711581630632, + "learning_rate": 0.284941670647085, + "loss": 0.2684, + "num_input_tokens_seen": 4390368, + "step": 5755 + }, + { + "epoch": 11.975051975051976, + "grad_norm": 0.00016567543207202107, + "learning_rate": 0.2849159369676563, + "loss": 0.2604, + "num_input_tokens_seen": 4394048, + "step": 5760 + }, + { + "epoch": 11.985446985446986, + "grad_norm": 0.0013829134404659271, + "learning_rate": 0.2848901824824948, + "loss": 0.2799, + "num_input_tokens_seen": 4397824, + "step": 5765 + }, + { + "epoch": 11.995841995841996, + "grad_norm": 0.0005715001025237143, + "learning_rate": 0.284864407195572, + "loss": 0.2706, + "num_input_tokens_seen": 4401696, + "step": 5770 + }, + { + "epoch": 12.006237006237006, + "grad_norm": 0.0015061214799061418, + "learning_rate": 0.28483861111086284, + "loss": 0.2637, + "num_input_tokens_seen": 4405368, + "step": 5775 + }, + { + "epoch": 12.016632016632016, + "grad_norm": 0.0008287896635010839, + "learning_rate": 0.2848127942323453, + "loss": 0.2837, + "num_input_tokens_seen": 4409304, + "step": 5780 + }, + { + "epoch": 12.027027027027026, + "grad_norm": 0.0006331239128485322, + "learning_rate": 0.2847869565640007, + "loss": 0.2772, + "num_input_tokens_seen": 4413016, + "step": 5785 + }, + { + "epoch": 12.037422037422038, + "grad_norm": 0.0012379608815535903, + "learning_rate": 0.2847610981098136, + "loss": 0.2783, + "num_input_tokens_seen": 4416696, + "step": 5790 + }, + { + "epoch": 12.047817047817048, + "grad_norm": 0.0012926555937156081, + "learning_rate": 0.2847352188737716, + "loss": 0.2823, + "num_input_tokens_seen": 4420440, + "step": 5795 + }, + { + "epoch": 12.058212058212058, + "grad_norm": 0.0013617328368127346, + "learning_rate": 0.2847093188598658, + "loss": 0.2883, + "num_input_tokens_seen": 4424280, + "step": 5800 + }, + { + "epoch": 12.058212058212058, + "eval_loss": 0.26770585775375366, + "eval_runtime": 13.4056, + "eval_samples_per_second": 63.854, + "eval_steps_per_second": 15.963, + "num_input_tokens_seen": 4424280, + "step": 5800 + }, + { + "epoch": 12.068607068607069, + "grad_norm": 0.00018180614279117435, + "learning_rate": 0.28468339807209003, + "loss": 0.2756, + "num_input_tokens_seen": 4428088, + "step": 5805 + }, + { + "epoch": 12.079002079002079, + "grad_norm": 0.00022024096688255668, + "learning_rate": 0.2846574565144418, + "loss": 0.2603, + "num_input_tokens_seen": 4431928, + "step": 5810 + }, + { + "epoch": 12.089397089397089, + "grad_norm": 9.740667883306742e-05, + "learning_rate": 0.28463149419092154, + "loss": 0.2587, + "num_input_tokens_seen": 4435864, + "step": 5815 + }, + { + "epoch": 12.0997920997921, + "grad_norm": 0.0004528888675849885, + "learning_rate": 0.284605511105533, + "loss": 0.265, + "num_input_tokens_seen": 4439544, + "step": 5820 + }, + { + "epoch": 12.11018711018711, + "grad_norm": 0.0007771481177769601, + "learning_rate": 0.28457950726228315, + "loss": 0.2695, + "num_input_tokens_seen": 4443224, + "step": 5825 + }, + { + "epoch": 12.120582120582121, + "grad_norm": 0.00012699983199127018, + "learning_rate": 0.28455348266518193, + "loss": 0.278, + "num_input_tokens_seen": 4447000, + "step": 5830 + }, + { + "epoch": 12.130977130977131, + "grad_norm": 0.0001094221806852147, + "learning_rate": 0.28452743731824287, + "loss": 0.2614, + "num_input_tokens_seen": 4450744, + "step": 5835 + }, + { + "epoch": 12.141372141372141, + "grad_norm": 0.0009366910089738667, + "learning_rate": 0.28450137122548236, + "loss": 0.2588, + "num_input_tokens_seen": 4454456, + "step": 5840 + }, + { + "epoch": 12.151767151767151, + "grad_norm": 0.0002853893965948373, + "learning_rate": 0.2844752843909201, + "loss": 0.2881, + "num_input_tokens_seen": 4458200, + "step": 5845 + }, + { + "epoch": 12.162162162162161, + "grad_norm": 0.00043805918539874256, + "learning_rate": 0.28444917681857923, + "loss": 0.2785, + "num_input_tokens_seen": 4461912, + "step": 5850 + }, + { + "epoch": 12.172557172557173, + "grad_norm": 0.0012598710600286722, + "learning_rate": 0.28442304851248557, + "loss": 0.288, + "num_input_tokens_seen": 4465720, + "step": 5855 + }, + { + "epoch": 12.182952182952183, + "grad_norm": 0.0009912147652357817, + "learning_rate": 0.2843968994766686, + "loss": 0.2593, + "num_input_tokens_seen": 4469624, + "step": 5860 + }, + { + "epoch": 12.193347193347194, + "grad_norm": 0.000685918377712369, + "learning_rate": 0.28437072971516075, + "loss": 0.2805, + "num_input_tokens_seen": 4473464, + "step": 5865 + }, + { + "epoch": 12.203742203742204, + "grad_norm": 0.001031465595588088, + "learning_rate": 0.2843445392319979, + "loss": 0.2691, + "num_input_tokens_seen": 4477240, + "step": 5870 + }, + { + "epoch": 12.214137214137214, + "grad_norm": 0.00011284740321571007, + "learning_rate": 0.28431832803121865, + "loss": 0.2729, + "num_input_tokens_seen": 4480952, + "step": 5875 + }, + { + "epoch": 12.224532224532224, + "grad_norm": 0.0010026207892224193, + "learning_rate": 0.28429209611686534, + "loss": 0.268, + "num_input_tokens_seen": 4484824, + "step": 5880 + }, + { + "epoch": 12.234927234927236, + "grad_norm": 0.00038795857108198106, + "learning_rate": 0.28426584349298323, + "loss": 0.2536, + "num_input_tokens_seen": 4488536, + "step": 5885 + }, + { + "epoch": 12.245322245322246, + "grad_norm": 0.0004827986122108996, + "learning_rate": 0.2842395701636207, + "loss": 0.2659, + "num_input_tokens_seen": 4492472, + "step": 5890 + }, + { + "epoch": 12.255717255717256, + "grad_norm": 0.0006701254169456661, + "learning_rate": 0.28421327613282954, + "loss": 0.2766, + "num_input_tokens_seen": 4496056, + "step": 5895 + }, + { + "epoch": 12.266112266112266, + "grad_norm": 0.000321400526445359, + "learning_rate": 0.28418696140466454, + "loss": 0.286, + "num_input_tokens_seen": 4499864, + "step": 5900 + }, + { + "epoch": 12.276507276507276, + "grad_norm": 0.0003939138550776988, + "learning_rate": 0.2841606259831838, + "loss": 0.2709, + "num_input_tokens_seen": 4503512, + "step": 5905 + }, + { + "epoch": 12.286902286902286, + "grad_norm": 0.001314492430537939, + "learning_rate": 0.2841342698724486, + "loss": 0.2688, + "num_input_tokens_seen": 4507224, + "step": 5910 + }, + { + "epoch": 12.297297297297296, + "grad_norm": 0.0020476027857512236, + "learning_rate": 0.28410789307652334, + "loss": 0.2821, + "num_input_tokens_seen": 4511128, + "step": 5915 + }, + { + "epoch": 12.307692307692308, + "grad_norm": 0.0008405789267271757, + "learning_rate": 0.2840814955994756, + "loss": 0.2699, + "num_input_tokens_seen": 4514840, + "step": 5920 + }, + { + "epoch": 12.318087318087318, + "grad_norm": 0.00016065254749264568, + "learning_rate": 0.2840550774453763, + "loss": 0.2848, + "num_input_tokens_seen": 4518648, + "step": 5925 + }, + { + "epoch": 12.328482328482329, + "grad_norm": 0.0002067928871838376, + "learning_rate": 0.28402863861829947, + "loss": 0.2772, + "num_input_tokens_seen": 4522392, + "step": 5930 + }, + { + "epoch": 12.338877338877339, + "grad_norm": 0.0013550748117268085, + "learning_rate": 0.2840021791223222, + "loss": 0.2733, + "num_input_tokens_seen": 4526200, + "step": 5935 + }, + { + "epoch": 12.349272349272349, + "grad_norm": 0.0005376901826821268, + "learning_rate": 0.2839756989615249, + "loss": 0.273, + "num_input_tokens_seen": 4529912, + "step": 5940 + }, + { + "epoch": 12.359667359667359, + "grad_norm": 0.0006089470698498189, + "learning_rate": 0.28394919813999125, + "loss": 0.2502, + "num_input_tokens_seen": 4533720, + "step": 5945 + }, + { + "epoch": 12.37006237006237, + "grad_norm": 0.0020214782562106848, + "learning_rate": 0.28392267666180787, + "loss": 0.2825, + "num_input_tokens_seen": 4537528, + "step": 5950 + }, + { + "epoch": 12.380457380457381, + "grad_norm": 0.000504746160004288, + "learning_rate": 0.2838961345310648, + "loss": 0.2656, + "num_input_tokens_seen": 4541400, + "step": 5955 + }, + { + "epoch": 12.390852390852391, + "grad_norm": 0.0003566770174074918, + "learning_rate": 0.2838695717518552, + "loss": 0.2994, + "num_input_tokens_seen": 4545336, + "step": 5960 + }, + { + "epoch": 12.401247401247401, + "grad_norm": 0.00028244126588106155, + "learning_rate": 0.28384298832827526, + "loss": 0.2663, + "num_input_tokens_seen": 4549240, + "step": 5965 + }, + { + "epoch": 12.411642411642411, + "grad_norm": 0.00011501672270242125, + "learning_rate": 0.28381638426442457, + "loss": 0.2678, + "num_input_tokens_seen": 4552984, + "step": 5970 + }, + { + "epoch": 12.422037422037421, + "grad_norm": 0.0004110147128812969, + "learning_rate": 0.2837897595644057, + "loss": 0.2756, + "num_input_tokens_seen": 4556632, + "step": 5975 + }, + { + "epoch": 12.432432432432432, + "grad_norm": 0.0007743166061118245, + "learning_rate": 0.28376311423232475, + "loss": 0.2669, + "num_input_tokens_seen": 4560376, + "step": 5980 + }, + { + "epoch": 12.442827442827443, + "grad_norm": 0.00017317116726189852, + "learning_rate": 0.2837364482722905, + "loss": 0.272, + "num_input_tokens_seen": 4564120, + "step": 5985 + }, + { + "epoch": 12.453222453222454, + "grad_norm": 0.0007913258741609752, + "learning_rate": 0.28370976168841533, + "loss": 0.2824, + "num_input_tokens_seen": 4567864, + "step": 5990 + }, + { + "epoch": 12.463617463617464, + "grad_norm": 0.0009628135594539344, + "learning_rate": 0.2836830544848146, + "loss": 0.2629, + "num_input_tokens_seen": 4571576, + "step": 5995 + }, + { + "epoch": 12.474012474012474, + "grad_norm": 7.43231939850375e-05, + "learning_rate": 0.2836563266656069, + "loss": 0.271, + "num_input_tokens_seen": 4575480, + "step": 6000 + }, + { + "epoch": 12.474012474012474, + "eval_loss": 0.24917590618133545, + "eval_runtime": 13.3975, + "eval_samples_per_second": 63.892, + "eval_steps_per_second": 15.973, + "num_input_tokens_seen": 4575480, + "step": 6000 + }, + { + "epoch": 12.484407484407484, + "grad_norm": 9.255088662030175e-05, + "learning_rate": 0.283629578234914, + "loss": 0.2529, + "num_input_tokens_seen": 4579224, + "step": 6005 + }, + { + "epoch": 12.494802494802494, + "grad_norm": 0.0004901768988929689, + "learning_rate": 0.2836028091968608, + "loss": 0.2905, + "num_input_tokens_seen": 4583192, + "step": 6010 + }, + { + "epoch": 12.505197505197506, + "grad_norm": 0.0010116827907040715, + "learning_rate": 0.28357601955557554, + "loss": 0.2489, + "num_input_tokens_seen": 4586872, + "step": 6015 + }, + { + "epoch": 12.515592515592516, + "grad_norm": 0.0005950178601779044, + "learning_rate": 0.2835492093151894, + "loss": 0.3089, + "num_input_tokens_seen": 4590712, + "step": 6020 + }, + { + "epoch": 12.525987525987526, + "grad_norm": 0.0007853138376958668, + "learning_rate": 0.2835223784798369, + "loss": 0.282, + "num_input_tokens_seen": 4594488, + "step": 6025 + }, + { + "epoch": 12.536382536382536, + "grad_norm": 0.0018702605739235878, + "learning_rate": 0.2834955270536557, + "loss": 0.2796, + "num_input_tokens_seen": 4598264, + "step": 6030 + }, + { + "epoch": 12.546777546777546, + "grad_norm": 0.00075818708864972, + "learning_rate": 0.2834686550407866, + "loss": 0.2789, + "num_input_tokens_seen": 4602104, + "step": 6035 + }, + { + "epoch": 12.557172557172557, + "grad_norm": 0.00012929151125717908, + "learning_rate": 0.28344176244537367, + "loss": 0.2611, + "num_input_tokens_seen": 4605976, + "step": 6040 + }, + { + "epoch": 12.567567567567568, + "grad_norm": 0.0014215228147804737, + "learning_rate": 0.28341484927156396, + "loss": 0.2702, + "num_input_tokens_seen": 4609848, + "step": 6045 + }, + { + "epoch": 12.577962577962579, + "grad_norm": 0.0009078827570192516, + "learning_rate": 0.28338791552350795, + "loss": 0.2554, + "num_input_tokens_seen": 4613656, + "step": 6050 + }, + { + "epoch": 12.588357588357589, + "grad_norm": 0.0001375266001559794, + "learning_rate": 0.28336096120535914, + "loss": 0.2652, + "num_input_tokens_seen": 4617464, + "step": 6055 + }, + { + "epoch": 12.598752598752599, + "grad_norm": 0.00038482018862850964, + "learning_rate": 0.2833339863212741, + "loss": 0.2433, + "num_input_tokens_seen": 4621336, + "step": 6060 + }, + { + "epoch": 12.609147609147609, + "grad_norm": 0.0005166112096048892, + "learning_rate": 0.28330699087541283, + "loss": 0.2808, + "num_input_tokens_seen": 4625208, + "step": 6065 + }, + { + "epoch": 12.619542619542619, + "grad_norm": 0.0001253746886504814, + "learning_rate": 0.2832799748719384, + "loss": 0.269, + "num_input_tokens_seen": 4629208, + "step": 6070 + }, + { + "epoch": 12.62993762993763, + "grad_norm": 0.00010340005246689543, + "learning_rate": 0.28325293831501686, + "loss": 0.2782, + "num_input_tokens_seen": 4632824, + "step": 6075 + }, + { + "epoch": 12.640332640332641, + "grad_norm": 0.0002453316992614418, + "learning_rate": 0.2832258812088177, + "loss": 0.2574, + "num_input_tokens_seen": 4636632, + "step": 6080 + }, + { + "epoch": 12.650727650727651, + "grad_norm": 0.0009290928137488663, + "learning_rate": 0.2831988035575134, + "loss": 0.2524, + "num_input_tokens_seen": 4640440, + "step": 6085 + }, + { + "epoch": 12.661122661122661, + "grad_norm": 0.0016221727710217237, + "learning_rate": 0.28317170536527975, + "loss": 0.2982, + "num_input_tokens_seen": 4644408, + "step": 6090 + }, + { + "epoch": 12.671517671517671, + "grad_norm": 0.0012651460710912943, + "learning_rate": 0.2831445866362956, + "loss": 0.2707, + "num_input_tokens_seen": 4648248, + "step": 6095 + }, + { + "epoch": 12.681912681912682, + "grad_norm": 0.0007314207032322884, + "learning_rate": 0.2831174473747429, + "loss": 0.2838, + "num_input_tokens_seen": 4651992, + "step": 6100 + }, + { + "epoch": 12.692307692307692, + "grad_norm": 0.001460422296077013, + "learning_rate": 0.2830902875848071, + "loss": 0.2603, + "num_input_tokens_seen": 4655576, + "step": 6105 + }, + { + "epoch": 12.702702702702704, + "grad_norm": 0.00042679597390815616, + "learning_rate": 0.28306310727067635, + "loss": 0.2771, + "num_input_tokens_seen": 4659416, + "step": 6110 + }, + { + "epoch": 12.713097713097714, + "grad_norm": 0.000388226326322183, + "learning_rate": 0.2830359064365423, + "loss": 0.2669, + "num_input_tokens_seen": 4663288, + "step": 6115 + }, + { + "epoch": 12.723492723492724, + "grad_norm": 0.0001736678823363036, + "learning_rate": 0.28300868508659965, + "loss": 0.2785, + "num_input_tokens_seen": 4667096, + "step": 6120 + }, + { + "epoch": 12.733887733887734, + "grad_norm": 0.00040306313894689083, + "learning_rate": 0.28298144322504626, + "loss": 0.2582, + "num_input_tokens_seen": 4670840, + "step": 6125 + }, + { + "epoch": 12.744282744282744, + "grad_norm": 0.0004889412084594369, + "learning_rate": 0.2829541808560832, + "loss": 0.2695, + "num_input_tokens_seen": 4674552, + "step": 6130 + }, + { + "epoch": 12.754677754677754, + "grad_norm": 0.00038278326974250376, + "learning_rate": 0.2829268979839146, + "loss": 0.2436, + "num_input_tokens_seen": 4678456, + "step": 6135 + }, + { + "epoch": 12.765072765072766, + "grad_norm": 0.00026976349181495607, + "learning_rate": 0.2828995946127479, + "loss": 0.2698, + "num_input_tokens_seen": 4682232, + "step": 6140 + }, + { + "epoch": 12.775467775467776, + "grad_norm": 0.0020005572587251663, + "learning_rate": 0.2828722707467936, + "loss": 0.2864, + "num_input_tokens_seen": 4686072, + "step": 6145 + }, + { + "epoch": 12.785862785862786, + "grad_norm": 0.00045545899774879217, + "learning_rate": 0.2828449263902653, + "loss": 0.2595, + "num_input_tokens_seen": 4689912, + "step": 6150 + }, + { + "epoch": 12.796257796257796, + "grad_norm": 0.00031507579842582345, + "learning_rate": 0.28281756154738, + "loss": 0.3103, + "num_input_tokens_seen": 4693688, + "step": 6155 + }, + { + "epoch": 12.806652806652806, + "grad_norm": 0.0013923244550824165, + "learning_rate": 0.28279017622235764, + "loss": 0.2664, + "num_input_tokens_seen": 4697496, + "step": 6160 + }, + { + "epoch": 12.817047817047817, + "grad_norm": 0.0006714232731610537, + "learning_rate": 0.28276277041942127, + "loss": 0.2586, + "num_input_tokens_seen": 4701368, + "step": 6165 + }, + { + "epoch": 12.827442827442827, + "grad_norm": 0.0016950289718806744, + "learning_rate": 0.2827353441427974, + "loss": 0.2783, + "num_input_tokens_seen": 4705304, + "step": 6170 + }, + { + "epoch": 12.837837837837839, + "grad_norm": 0.0001451850839657709, + "learning_rate": 0.2827078973967153, + "loss": 0.2685, + "num_input_tokens_seen": 4709240, + "step": 6175 + }, + { + "epoch": 12.848232848232849, + "grad_norm": 0.00039057733374647796, + "learning_rate": 0.2826804301854078, + "loss": 0.2766, + "num_input_tokens_seen": 4713080, + "step": 6180 + }, + { + "epoch": 12.858627858627859, + "grad_norm": 0.0007726919720880687, + "learning_rate": 0.2826529425131105, + "loss": 0.2747, + "num_input_tokens_seen": 4717112, + "step": 6185 + }, + { + "epoch": 12.869022869022869, + "grad_norm": 0.0006966284126974642, + "learning_rate": 0.2826254343840625, + "loss": 0.2658, + "num_input_tokens_seen": 4721112, + "step": 6190 + }, + { + "epoch": 12.879417879417879, + "grad_norm": 0.00011544146400410682, + "learning_rate": 0.2825979058025059, + "loss": 0.2738, + "num_input_tokens_seen": 4724856, + "step": 6195 + }, + { + "epoch": 12.88981288981289, + "grad_norm": 0.0003345098521094769, + "learning_rate": 0.2825703567726858, + "loss": 0.2644, + "num_input_tokens_seen": 4728792, + "step": 6200 + }, + { + "epoch": 12.88981288981289, + "eval_loss": 0.24961183965206146, + "eval_runtime": 13.4495, + "eval_samples_per_second": 63.645, + "eval_steps_per_second": 15.911, + "num_input_tokens_seen": 4728792, + "step": 6200 + }, + { + "epoch": 12.9002079002079, + "grad_norm": 0.00044291187077760696, + "learning_rate": 0.2825427872988508, + "loss": 0.2614, + "num_input_tokens_seen": 4732600, + "step": 6205 + }, + { + "epoch": 12.910602910602911, + "grad_norm": 0.0003708438016474247, + "learning_rate": 0.28251519738525227, + "loss": 0.2874, + "num_input_tokens_seen": 4736472, + "step": 6210 + }, + { + "epoch": 12.920997920997921, + "grad_norm": 0.00046882391325198114, + "learning_rate": 0.28248758703614507, + "loss": 0.2839, + "num_input_tokens_seen": 4740248, + "step": 6215 + }, + { + "epoch": 12.931392931392931, + "grad_norm": 0.00145276531111449, + "learning_rate": 0.28245995625578696, + "loss": 0.2751, + "num_input_tokens_seen": 4744216, + "step": 6220 + }, + { + "epoch": 12.941787941787942, + "grad_norm": 0.00027773386682383716, + "learning_rate": 0.282432305048439, + "loss": 0.2785, + "num_input_tokens_seen": 4747960, + "step": 6225 + }, + { + "epoch": 12.952182952182952, + "grad_norm": 0.0012273472966626287, + "learning_rate": 0.28240463341836536, + "loss": 0.2698, + "num_input_tokens_seen": 4751704, + "step": 6230 + }, + { + "epoch": 12.962577962577962, + "grad_norm": 0.0018733071628957987, + "learning_rate": 0.2823769413698334, + "loss": 0.2862, + "num_input_tokens_seen": 4755480, + "step": 6235 + }, + { + "epoch": 12.972972972972974, + "grad_norm": 0.0008979453705251217, + "learning_rate": 0.2823492289071135, + "loss": 0.2502, + "num_input_tokens_seen": 4759128, + "step": 6240 + }, + { + "epoch": 12.983367983367984, + "grad_norm": 0.00033327305573038757, + "learning_rate": 0.2823214960344793, + "loss": 0.2532, + "num_input_tokens_seen": 4763160, + "step": 6245 + }, + { + "epoch": 12.993762993762994, + "grad_norm": 9.502907778369263e-05, + "learning_rate": 0.28229374275620756, + "loss": 0.2426, + "num_input_tokens_seen": 4767096, + "step": 6250 + }, + { + "epoch": 13.004158004158004, + "grad_norm": 0.000791467959061265, + "learning_rate": 0.28226596907657814, + "loss": 0.275, + "num_input_tokens_seen": 4770896, + "step": 6255 + }, + { + "epoch": 13.014553014553014, + "grad_norm": 0.0010478816693648696, + "learning_rate": 0.28223817499987414, + "loss": 0.2809, + "num_input_tokens_seen": 4774704, + "step": 6260 + }, + { + "epoch": 13.024948024948024, + "grad_norm": 0.0010234445799142122, + "learning_rate": 0.2822103605303818, + "loss": 0.2708, + "num_input_tokens_seen": 4778416, + "step": 6265 + }, + { + "epoch": 13.035343035343036, + "grad_norm": 0.00024356390349566936, + "learning_rate": 0.2821825256723903, + "loss": 0.2847, + "num_input_tokens_seen": 4782192, + "step": 6270 + }, + { + "epoch": 13.045738045738046, + "grad_norm": 0.00011574855307117105, + "learning_rate": 0.2821546704301923, + "loss": 0.2983, + "num_input_tokens_seen": 4786032, + "step": 6275 + }, + { + "epoch": 13.056133056133056, + "grad_norm": 0.0010286258766427636, + "learning_rate": 0.2821267948080834, + "loss": 0.2702, + "num_input_tokens_seen": 4789776, + "step": 6280 + }, + { + "epoch": 13.066528066528067, + "grad_norm": 0.0003775153891183436, + "learning_rate": 0.28209889881036226, + "loss": 0.236, + "num_input_tokens_seen": 4793648, + "step": 6285 + }, + { + "epoch": 13.076923076923077, + "grad_norm": 0.0012265084078535438, + "learning_rate": 0.28207098244133094, + "loss": 0.3034, + "num_input_tokens_seen": 4797648, + "step": 6290 + }, + { + "epoch": 13.087318087318087, + "grad_norm": 0.0017759909387677908, + "learning_rate": 0.2820430457052943, + "loss": 0.2806, + "num_input_tokens_seen": 4801456, + "step": 6295 + }, + { + "epoch": 13.097713097713097, + "grad_norm": 0.0004505858232732862, + "learning_rate": 0.28201508860656077, + "loss": 0.2686, + "num_input_tokens_seen": 4805264, + "step": 6300 + }, + { + "epoch": 13.108108108108109, + "grad_norm": 0.0008845975971780717, + "learning_rate": 0.2819871111494415, + "loss": 0.3032, + "num_input_tokens_seen": 4809008, + "step": 6305 + }, + { + "epoch": 13.118503118503119, + "grad_norm": 0.001012252876535058, + "learning_rate": 0.28195911333825113, + "loss": 0.2753, + "num_input_tokens_seen": 4812848, + "step": 6310 + }, + { + "epoch": 13.128898128898129, + "grad_norm": 0.00036685154191218317, + "learning_rate": 0.28193109517730713, + "loss": 0.2701, + "num_input_tokens_seen": 4816752, + "step": 6315 + }, + { + "epoch": 13.13929313929314, + "grad_norm": 0.0002599575382191688, + "learning_rate": 0.2819030566709303, + "loss": 0.2786, + "num_input_tokens_seen": 4820432, + "step": 6320 + }, + { + "epoch": 13.14968814968815, + "grad_norm": 0.0006284094415605068, + "learning_rate": 0.2818749978234445, + "loss": 0.2745, + "num_input_tokens_seen": 4824272, + "step": 6325 + }, + { + "epoch": 13.16008316008316, + "grad_norm": 0.000569976749829948, + "learning_rate": 0.2818469186391768, + "loss": 0.2682, + "num_input_tokens_seen": 4827984, + "step": 6330 + }, + { + "epoch": 13.170478170478171, + "grad_norm": 0.0008672393742017448, + "learning_rate": 0.28181881912245743, + "loss": 0.2503, + "num_input_tokens_seen": 4831664, + "step": 6335 + }, + { + "epoch": 13.180873180873181, + "grad_norm": 0.0005721260677091777, + "learning_rate": 0.2817906992776195, + "loss": 0.2696, + "num_input_tokens_seen": 4835280, + "step": 6340 + }, + { + "epoch": 13.191268191268192, + "grad_norm": 0.0009665184188634157, + "learning_rate": 0.28176255910899967, + "loss": 0.2821, + "num_input_tokens_seen": 4839088, + "step": 6345 + }, + { + "epoch": 13.201663201663202, + "grad_norm": 0.0006864466122351587, + "learning_rate": 0.2817343986209373, + "loss": 0.2764, + "num_input_tokens_seen": 4842960, + "step": 6350 + }, + { + "epoch": 13.212058212058212, + "grad_norm": 0.0003391333157196641, + "learning_rate": 0.2817062178177753, + "loss": 0.2582, + "num_input_tokens_seen": 4846800, + "step": 6355 + }, + { + "epoch": 13.222453222453222, + "grad_norm": 0.0008498854003846645, + "learning_rate": 0.2816780167038593, + "loss": 0.3029, + "num_input_tokens_seen": 4850640, + "step": 6360 + }, + { + "epoch": 13.232848232848234, + "grad_norm": 0.001787275425158441, + "learning_rate": 0.28164979528353834, + "loss": 0.2902, + "num_input_tokens_seen": 4854352, + "step": 6365 + }, + { + "epoch": 13.243243243243244, + "grad_norm": 0.0008399019716307521, + "learning_rate": 0.28162155356116453, + "loss": 0.2688, + "num_input_tokens_seen": 4858096, + "step": 6370 + }, + { + "epoch": 13.253638253638254, + "grad_norm": 0.0014857238857075572, + "learning_rate": 0.28159329154109314, + "loss": 0.2366, + "num_input_tokens_seen": 4861968, + "step": 6375 + }, + { + "epoch": 13.264033264033264, + "grad_norm": 0.0009478324209339917, + "learning_rate": 0.28156500922768246, + "loss": 0.3054, + "num_input_tokens_seen": 4865776, + "step": 6380 + }, + { + "epoch": 13.274428274428274, + "grad_norm": 0.002258904976770282, + "learning_rate": 0.28153670662529406, + "loss": 0.2809, + "num_input_tokens_seen": 4869552, + "step": 6385 + }, + { + "epoch": 13.284823284823284, + "grad_norm": 0.0005595171242021024, + "learning_rate": 0.28150838373829246, + "loss": 0.2793, + "num_input_tokens_seen": 4873488, + "step": 6390 + }, + { + "epoch": 13.295218295218294, + "grad_norm": 0.0008201505988836288, + "learning_rate": 0.2814800405710455, + "loss": 0.2712, + "num_input_tokens_seen": 4877200, + "step": 6395 + }, + { + "epoch": 13.305613305613306, + "grad_norm": 0.000540904002264142, + "learning_rate": 0.2814516771279239, + "loss": 0.2439, + "num_input_tokens_seen": 4880880, + "step": 6400 + }, + { + "epoch": 13.305613305613306, + "eval_loss": 0.2496749460697174, + "eval_runtime": 13.4111, + "eval_samples_per_second": 63.828, + "eval_steps_per_second": 15.957, + "num_input_tokens_seen": 4880880, + "step": 6400 + }, + { + "epoch": 13.316008316008316, + "grad_norm": 0.00034471412072889507, + "learning_rate": 0.28142329341330186, + "loss": 0.2857, + "num_input_tokens_seen": 4884560, + "step": 6405 + }, + { + "epoch": 13.326403326403327, + "grad_norm": 7.143099355744198e-05, + "learning_rate": 0.2813948894315564, + "loss": 0.2722, + "num_input_tokens_seen": 4888400, + "step": 6410 + }, + { + "epoch": 13.336798336798337, + "grad_norm": 0.0003664206597022712, + "learning_rate": 0.2813664651870677, + "loss": 0.29, + "num_input_tokens_seen": 4892176, + "step": 6415 + }, + { + "epoch": 13.347193347193347, + "grad_norm": 0.0004640058905351907, + "learning_rate": 0.28133802068421926, + "loss": 0.2594, + "num_input_tokens_seen": 4895952, + "step": 6420 + }, + { + "epoch": 13.357588357588357, + "grad_norm": 0.00030228865216486156, + "learning_rate": 0.28130955592739754, + "loss": 0.2802, + "num_input_tokens_seen": 4899856, + "step": 6425 + }, + { + "epoch": 13.367983367983369, + "grad_norm": 0.00037279140087775886, + "learning_rate": 0.2812810709209922, + "loss": 0.2667, + "num_input_tokens_seen": 4903728, + "step": 6430 + }, + { + "epoch": 13.378378378378379, + "grad_norm": 0.0009131809347309172, + "learning_rate": 0.2812525656693959, + "loss": 0.2634, + "num_input_tokens_seen": 4907504, + "step": 6435 + }, + { + "epoch": 13.388773388773389, + "grad_norm": 0.0001961613743333146, + "learning_rate": 0.28122404017700453, + "loss": 0.2531, + "num_input_tokens_seen": 4911280, + "step": 6440 + }, + { + "epoch": 13.3991683991684, + "grad_norm": 0.0010716636897996068, + "learning_rate": 0.2811954944482171, + "loss": 0.2886, + "num_input_tokens_seen": 4915120, + "step": 6445 + }, + { + "epoch": 13.40956340956341, + "grad_norm": 0.00133323366753757, + "learning_rate": 0.2811669284874358, + "loss": 0.3022, + "num_input_tokens_seen": 4918928, + "step": 6450 + }, + { + "epoch": 13.41995841995842, + "grad_norm": 0.003253166563808918, + "learning_rate": 0.2811383422990657, + "loss": 0.2374, + "num_input_tokens_seen": 4922672, + "step": 6455 + }, + { + "epoch": 13.43035343035343, + "grad_norm": 0.0007003267528489232, + "learning_rate": 0.2811097358875152, + "loss": 0.2337, + "num_input_tokens_seen": 4926576, + "step": 6460 + }, + { + "epoch": 13.440748440748441, + "grad_norm": 0.000843209505546838, + "learning_rate": 0.2810811092571959, + "loss": 0.327, + "num_input_tokens_seen": 4930384, + "step": 6465 + }, + { + "epoch": 13.451143451143452, + "grad_norm": 0.0005050098989158869, + "learning_rate": 0.28105246241252224, + "loss": 0.2666, + "num_input_tokens_seen": 4934160, + "step": 6470 + }, + { + "epoch": 13.461538461538462, + "grad_norm": 0.0003771214687731117, + "learning_rate": 0.28102379535791194, + "loss": 0.2751, + "num_input_tokens_seen": 4937936, + "step": 6475 + }, + { + "epoch": 13.471933471933472, + "grad_norm": 0.0005760493804700673, + "learning_rate": 0.2809951080977859, + "loss": 0.2444, + "num_input_tokens_seen": 4941808, + "step": 6480 + }, + { + "epoch": 13.482328482328482, + "grad_norm": 0.0004649146576412022, + "learning_rate": 0.28096640063656797, + "loss": 0.3131, + "num_input_tokens_seen": 4945648, + "step": 6485 + }, + { + "epoch": 13.492723492723492, + "grad_norm": 0.0007316998089663684, + "learning_rate": 0.2809376729786852, + "loss": 0.2846, + "num_input_tokens_seen": 4949424, + "step": 6490 + }, + { + "epoch": 13.503118503118504, + "grad_norm": 0.0005719223408959806, + "learning_rate": 0.28090892512856785, + "loss": 0.2812, + "num_input_tokens_seen": 4953456, + "step": 6495 + }, + { + "epoch": 13.513513513513514, + "grad_norm": 0.0005000099190510809, + "learning_rate": 0.2808801570906491, + "loss": 0.2645, + "num_input_tokens_seen": 4957264, + "step": 6500 + }, + { + "epoch": 13.523908523908524, + "grad_norm": 0.000599811552092433, + "learning_rate": 0.2808513688693654, + "loss": 0.2548, + "num_input_tokens_seen": 4961072, + "step": 6505 + }, + { + "epoch": 13.534303534303534, + "grad_norm": 0.0011316632153466344, + "learning_rate": 0.28082256046915627, + "loss": 0.2609, + "num_input_tokens_seen": 4964944, + "step": 6510 + }, + { + "epoch": 13.544698544698544, + "grad_norm": 0.0016541045624762774, + "learning_rate": 0.28079373189446427, + "loss": 0.2714, + "num_input_tokens_seen": 4968848, + "step": 6515 + }, + { + "epoch": 13.555093555093555, + "grad_norm": 0.0005101578426547348, + "learning_rate": 0.28076488314973513, + "loss": 0.2559, + "num_input_tokens_seen": 4972656, + "step": 6520 + }, + { + "epoch": 13.565488565488565, + "grad_norm": 0.00035254814429208636, + "learning_rate": 0.28073601423941774, + "loss": 0.2939, + "num_input_tokens_seen": 4976368, + "step": 6525 + }, + { + "epoch": 13.575883575883577, + "grad_norm": 0.0011359798954799771, + "learning_rate": 0.28070712516796403, + "loss": 0.2729, + "num_input_tokens_seen": 4980144, + "step": 6530 + }, + { + "epoch": 13.586278586278587, + "grad_norm": 0.0012922968016937375, + "learning_rate": 0.28067821593982906, + "loss": 0.2688, + "num_input_tokens_seen": 4984080, + "step": 6535 + }, + { + "epoch": 13.596673596673597, + "grad_norm": 0.0006646570400334895, + "learning_rate": 0.28064928655947097, + "loss": 0.3093, + "num_input_tokens_seen": 4988016, + "step": 6540 + }, + { + "epoch": 13.607068607068607, + "grad_norm": 0.0009308150038123131, + "learning_rate": 0.28062033703135103, + "loss": 0.2895, + "num_input_tokens_seen": 4991920, + "step": 6545 + }, + { + "epoch": 13.617463617463617, + "grad_norm": 0.00025679919053800404, + "learning_rate": 0.2805913673599337, + "loss": 0.2953, + "num_input_tokens_seen": 4995760, + "step": 6550 + }, + { + "epoch": 13.627858627858627, + "grad_norm": 0.0013019065372645855, + "learning_rate": 0.2805623775496864, + "loss": 0.2609, + "num_input_tokens_seen": 4999632, + "step": 6555 + }, + { + "epoch": 13.638253638253639, + "grad_norm": 0.0001876809837995097, + "learning_rate": 0.2805333676050797, + "loss": 0.2864, + "num_input_tokens_seen": 5003536, + "step": 6560 + }, + { + "epoch": 13.64864864864865, + "grad_norm": 0.000774115149397403, + "learning_rate": 0.2805043375305873, + "loss": 0.2753, + "num_input_tokens_seen": 5007440, + "step": 6565 + }, + { + "epoch": 13.65904365904366, + "grad_norm": 0.0013531945878639817, + "learning_rate": 0.2804752873306861, + "loss": 0.2791, + "num_input_tokens_seen": 5011184, + "step": 6570 + }, + { + "epoch": 13.66943866943867, + "grad_norm": 0.0001015575326164253, + "learning_rate": 0.2804462170098559, + "loss": 0.2796, + "num_input_tokens_seen": 5015120, + "step": 6575 + }, + { + "epoch": 13.67983367983368, + "grad_norm": 7.617779920110479e-05, + "learning_rate": 0.2804171265725797, + "loss": 0.2685, + "num_input_tokens_seen": 5019024, + "step": 6580 + }, + { + "epoch": 13.69022869022869, + "grad_norm": 0.00034285796573385596, + "learning_rate": 0.28038801602334373, + "loss": 0.2688, + "num_input_tokens_seen": 5022960, + "step": 6585 + }, + { + "epoch": 13.700623700623701, + "grad_norm": 0.0007522567757405341, + "learning_rate": 0.28035888536663717, + "loss": 0.2592, + "num_input_tokens_seen": 5026864, + "step": 6590 + }, + { + "epoch": 13.711018711018712, + "grad_norm": 0.0010528706479817629, + "learning_rate": 0.2803297346069522, + "loss": 0.2919, + "num_input_tokens_seen": 5030800, + "step": 6595 + }, + { + "epoch": 13.721413721413722, + "grad_norm": 0.0017097215168178082, + "learning_rate": 0.28030056374878437, + "loss": 0.2949, + "num_input_tokens_seen": 5034608, + "step": 6600 + }, + { + "epoch": 13.721413721413722, + "eval_loss": 0.2552819848060608, + "eval_runtime": 13.4764, + "eval_samples_per_second": 63.519, + "eval_steps_per_second": 15.88, + "num_input_tokens_seen": 5034608, + "step": 6600 + }, + { + "epoch": 13.731808731808732, + "grad_norm": 0.0009890925139188766, + "learning_rate": 0.2802713727966321, + "loss": 0.2629, + "num_input_tokens_seen": 5038480, + "step": 6605 + }, + { + "epoch": 13.742203742203742, + "grad_norm": 0.0006453191163018346, + "learning_rate": 0.28024216175499717, + "loss": 0.2541, + "num_input_tokens_seen": 5042160, + "step": 6610 + }, + { + "epoch": 13.752598752598752, + "grad_norm": 0.00013424699136521667, + "learning_rate": 0.2802129306283841, + "loss": 0.244, + "num_input_tokens_seen": 5046000, + "step": 6615 + }, + { + "epoch": 13.762993762993762, + "grad_norm": 0.0012249441351741552, + "learning_rate": 0.28018367942130074, + "loss": 0.274, + "num_input_tokens_seen": 5049648, + "step": 6620 + }, + { + "epoch": 13.773388773388774, + "grad_norm": 0.0006099113961681724, + "learning_rate": 0.28015440813825804, + "loss": 0.2711, + "num_input_tokens_seen": 5053456, + "step": 6625 + }, + { + "epoch": 13.783783783783784, + "grad_norm": 0.0006076281424611807, + "learning_rate": 0.28012511678377006, + "loss": 0.2697, + "num_input_tokens_seen": 5057200, + "step": 6630 + }, + { + "epoch": 13.794178794178794, + "grad_norm": 0.0002612782991491258, + "learning_rate": 0.28009580536235373, + "loss": 0.268, + "num_input_tokens_seen": 5060944, + "step": 6635 + }, + { + "epoch": 13.804573804573804, + "grad_norm": 0.0005154876271262765, + "learning_rate": 0.28006647387852934, + "loss": 0.2761, + "num_input_tokens_seen": 5064816, + "step": 6640 + }, + { + "epoch": 13.814968814968815, + "grad_norm": 0.0008332670549862087, + "learning_rate": 0.28003712233682015, + "loss": 0.2613, + "num_input_tokens_seen": 5068496, + "step": 6645 + }, + { + "epoch": 13.825363825363825, + "grad_norm": 0.00019063743820879608, + "learning_rate": 0.2800077507417526, + "loss": 0.21, + "num_input_tokens_seen": 5072240, + "step": 6650 + }, + { + "epoch": 13.835758835758837, + "grad_norm": 0.0001325707562500611, + "learning_rate": 0.2799783590978561, + "loss": 0.3029, + "num_input_tokens_seen": 5076080, + "step": 6655 + }, + { + "epoch": 13.846153846153847, + "grad_norm": 0.001270609674975276, + "learning_rate": 0.2799489474096632, + "loss": 0.2918, + "num_input_tokens_seen": 5080048, + "step": 6660 + }, + { + "epoch": 13.856548856548857, + "grad_norm": 0.00011493675265228376, + "learning_rate": 0.27991951568170953, + "loss": 0.2847, + "num_input_tokens_seen": 5083696, + "step": 6665 + }, + { + "epoch": 13.866943866943867, + "grad_norm": 0.0009804422734305263, + "learning_rate": 0.2798900639185339, + "loss": 0.3059, + "num_input_tokens_seen": 5087408, + "step": 6670 + }, + { + "epoch": 13.877338877338877, + "grad_norm": 0.0011820865329355001, + "learning_rate": 0.2798605921246781, + "loss": 0.2821, + "num_input_tokens_seen": 5091344, + "step": 6675 + }, + { + "epoch": 13.887733887733887, + "grad_norm": 0.00038007143302820623, + "learning_rate": 0.2798311003046871, + "loss": 0.2785, + "num_input_tokens_seen": 5095152, + "step": 6680 + }, + { + "epoch": 13.898128898128899, + "grad_norm": 0.0001244920858880505, + "learning_rate": 0.2798015884631089, + "loss": 0.2688, + "num_input_tokens_seen": 5099120, + "step": 6685 + }, + { + "epoch": 13.90852390852391, + "grad_norm": 0.00029996802913956344, + "learning_rate": 0.27977205660449445, + "loss": 0.2646, + "num_input_tokens_seen": 5103056, + "step": 6690 + }, + { + "epoch": 13.91891891891892, + "grad_norm": 0.00040143224759958684, + "learning_rate": 0.2797425047333981, + "loss": 0.2527, + "num_input_tokens_seen": 5106928, + "step": 6695 + }, + { + "epoch": 13.92931392931393, + "grad_norm": 0.0004995149793103337, + "learning_rate": 0.27971293285437715, + "loss": 0.292, + "num_input_tokens_seen": 5110704, + "step": 6700 + }, + { + "epoch": 13.93970893970894, + "grad_norm": 0.0003274181508459151, + "learning_rate": 0.2796833409719918, + "loss": 0.2713, + "num_input_tokens_seen": 5114640, + "step": 6705 + }, + { + "epoch": 13.95010395010395, + "grad_norm": 0.001920558512210846, + "learning_rate": 0.27965372909080566, + "loss": 0.2807, + "num_input_tokens_seen": 5118320, + "step": 6710 + }, + { + "epoch": 13.96049896049896, + "grad_norm": 0.0006292856414802372, + "learning_rate": 0.27962409721538506, + "loss": 0.2717, + "num_input_tokens_seen": 5122160, + "step": 6715 + }, + { + "epoch": 13.970893970893972, + "grad_norm": 0.0004902294022031128, + "learning_rate": 0.27959444535029976, + "loss": 0.2631, + "num_input_tokens_seen": 5125904, + "step": 6720 + }, + { + "epoch": 13.981288981288982, + "grad_norm": 0.0005971001810394228, + "learning_rate": 0.27956477350012243, + "loss": 0.2839, + "num_input_tokens_seen": 5129680, + "step": 6725 + }, + { + "epoch": 13.991683991683992, + "grad_norm": 0.00025451448163948953, + "learning_rate": 0.27953508166942875, + "loss": 0.2829, + "num_input_tokens_seen": 5133392, + "step": 6730 + }, + { + "epoch": 14.002079002079002, + "grad_norm": 0.00016253946523647755, + "learning_rate": 0.27950536986279767, + "loss": 0.3089, + "num_input_tokens_seen": 5137056, + "step": 6735 + }, + { + "epoch": 14.012474012474012, + "grad_norm": 0.0005154295940883458, + "learning_rate": 0.2794756380848111, + "loss": 0.26, + "num_input_tokens_seen": 5140832, + "step": 6740 + }, + { + "epoch": 14.022869022869022, + "grad_norm": 0.0002365779655519873, + "learning_rate": 0.279445886340054, + "loss": 0.2481, + "num_input_tokens_seen": 5144640, + "step": 6745 + }, + { + "epoch": 14.033264033264032, + "grad_norm": 0.00011351846478646621, + "learning_rate": 0.27941611463311455, + "loss": 0.2522, + "num_input_tokens_seen": 5148480, + "step": 6750 + }, + { + "epoch": 14.043659043659044, + "grad_norm": 0.0008666909416206181, + "learning_rate": 0.2793863229685839, + "loss": 0.2844, + "num_input_tokens_seen": 5152320, + "step": 6755 + }, + { + "epoch": 14.054054054054054, + "grad_norm": 0.0004768530197907239, + "learning_rate": 0.27935651135105627, + "loss": 0.26, + "num_input_tokens_seen": 5156032, + "step": 6760 + }, + { + "epoch": 14.064449064449065, + "grad_norm": 0.0001247638720087707, + "learning_rate": 0.279326679785129, + "loss": 0.2442, + "num_input_tokens_seen": 5159808, + "step": 6765 + }, + { + "epoch": 14.074844074844075, + "grad_norm": 0.0009793314384296536, + "learning_rate": 0.2792968282754024, + "loss": 0.3086, + "num_input_tokens_seen": 5163520, + "step": 6770 + }, + { + "epoch": 14.085239085239085, + "grad_norm": 0.0006179669871926308, + "learning_rate": 0.2792669568264801, + "loss": 0.2798, + "num_input_tokens_seen": 5167200, + "step": 6775 + }, + { + "epoch": 14.095634095634095, + "grad_norm": 0.0003137282619718462, + "learning_rate": 0.27923706544296856, + "loss": 0.2547, + "num_input_tokens_seen": 5170944, + "step": 6780 + }, + { + "epoch": 14.106029106029107, + "grad_norm": 0.0006193777662701905, + "learning_rate": 0.2792071541294775, + "loss": 0.2805, + "num_input_tokens_seen": 5174816, + "step": 6785 + }, + { + "epoch": 14.116424116424117, + "grad_norm": 0.0009216408943757415, + "learning_rate": 0.27917722289061947, + "loss": 0.2944, + "num_input_tokens_seen": 5178784, + "step": 6790 + }, + { + "epoch": 14.126819126819127, + "grad_norm": 0.00031459599267691374, + "learning_rate": 0.27914727173101034, + "loss": 0.2726, + "num_input_tokens_seen": 5182624, + "step": 6795 + }, + { + "epoch": 14.137214137214137, + "grad_norm": 0.00016076047904789448, + "learning_rate": 0.279117300655269, + "loss": 0.2525, + "num_input_tokens_seen": 5186400, + "step": 6800 + }, + { + "epoch": 14.137214137214137, + "eval_loss": 0.2500765025615692, + "eval_runtime": 13.4269, + "eval_samples_per_second": 63.753, + "eval_steps_per_second": 15.938, + "num_input_tokens_seen": 5186400, + "step": 6800 + }, + { + "epoch": 14.147609147609147, + "grad_norm": 0.0005756820901297033, + "learning_rate": 0.2790873096680173, + "loss": 0.2931, + "num_input_tokens_seen": 5190208, + "step": 6805 + }, + { + "epoch": 14.158004158004157, + "grad_norm": 0.0004946394474245608, + "learning_rate": 0.2790572987738802, + "loss": 0.2771, + "num_input_tokens_seen": 5194176, + "step": 6810 + }, + { + "epoch": 14.16839916839917, + "grad_norm": 0.0003213364689145237, + "learning_rate": 0.27902726797748584, + "loss": 0.2653, + "num_input_tokens_seen": 5197952, + "step": 6815 + }, + { + "epoch": 14.17879417879418, + "grad_norm": 0.000698927731718868, + "learning_rate": 0.2789972172834652, + "loss": 0.2301, + "num_input_tokens_seen": 5201728, + "step": 6820 + }, + { + "epoch": 14.18918918918919, + "grad_norm": 0.0005051936605013907, + "learning_rate": 0.2789671466964527, + "loss": 0.289, + "num_input_tokens_seen": 5205600, + "step": 6825 + }, + { + "epoch": 14.1995841995842, + "grad_norm": 0.00032667024061083794, + "learning_rate": 0.2789370562210854, + "loss": 0.2637, + "num_input_tokens_seen": 5209312, + "step": 6830 + }, + { + "epoch": 14.20997920997921, + "grad_norm": 0.00027768072322942317, + "learning_rate": 0.27890694586200376, + "loss": 0.2727, + "num_input_tokens_seen": 5213184, + "step": 6835 + }, + { + "epoch": 14.22037422037422, + "grad_norm": 0.0005688949604518712, + "learning_rate": 0.2788768156238511, + "loss": 0.2481, + "num_input_tokens_seen": 5217024, + "step": 6840 + }, + { + "epoch": 14.23076923076923, + "grad_norm": 0.00036395780625753105, + "learning_rate": 0.27884666551127385, + "loss": 0.2526, + "num_input_tokens_seen": 5220832, + "step": 6845 + }, + { + "epoch": 14.241164241164242, + "grad_norm": 0.0005436564679257572, + "learning_rate": 0.2788164955289217, + "loss": 0.275, + "num_input_tokens_seen": 5224736, + "step": 6850 + }, + { + "epoch": 14.251559251559252, + "grad_norm": 7.632367487531155e-05, + "learning_rate": 0.27878630568144697, + "loss": 0.2688, + "num_input_tokens_seen": 5228736, + "step": 6855 + }, + { + "epoch": 14.261954261954262, + "grad_norm": 0.00011083899153163657, + "learning_rate": 0.2787560959735056, + "loss": 0.2664, + "num_input_tokens_seen": 5232512, + "step": 6860 + }, + { + "epoch": 14.272349272349272, + "grad_norm": 0.001330072176642716, + "learning_rate": 0.27872586640975616, + "loss": 0.2793, + "num_input_tokens_seen": 5236352, + "step": 6865 + }, + { + "epoch": 14.282744282744282, + "grad_norm": 0.0016209499444812536, + "learning_rate": 0.27869561699486045, + "loss": 0.285, + "num_input_tokens_seen": 5240224, + "step": 6870 + }, + { + "epoch": 14.293139293139292, + "grad_norm": 0.0009718404617160559, + "learning_rate": 0.2786653477334833, + "loss": 0.2843, + "num_input_tokens_seen": 5243968, + "step": 6875 + }, + { + "epoch": 14.303534303534304, + "grad_norm": 0.0006024660542607307, + "learning_rate": 0.2786350586302926, + "loss": 0.2396, + "num_input_tokens_seen": 5247648, + "step": 6880 + }, + { + "epoch": 14.313929313929314, + "grad_norm": 0.0013160171220079064, + "learning_rate": 0.27860474968995935, + "loss": 0.2923, + "num_input_tokens_seen": 5251328, + "step": 6885 + }, + { + "epoch": 14.324324324324325, + "grad_norm": 0.00010290257341694087, + "learning_rate": 0.27857442091715756, + "loss": 0.2544, + "num_input_tokens_seen": 5255008, + "step": 6890 + }, + { + "epoch": 14.334719334719335, + "grad_norm": 0.000662333273794502, + "learning_rate": 0.27854407231656425, + "loss": 0.2655, + "num_input_tokens_seen": 5258720, + "step": 6895 + }, + { + "epoch": 14.345114345114345, + "grad_norm": 0.0007001158664934337, + "learning_rate": 0.2785137038928596, + "loss": 0.2789, + "num_input_tokens_seen": 5262528, + "step": 6900 + }, + { + "epoch": 14.355509355509355, + "grad_norm": 0.0003008940548170358, + "learning_rate": 0.27848331565072687, + "loss": 0.2867, + "num_input_tokens_seen": 5266240, + "step": 6905 + }, + { + "epoch": 14.365904365904367, + "grad_norm": 0.00010646719601936638, + "learning_rate": 0.27845290759485225, + "loss": 0.2696, + "num_input_tokens_seen": 5270080, + "step": 6910 + }, + { + "epoch": 14.376299376299377, + "grad_norm": 0.00040050657116807997, + "learning_rate": 0.278422479729925, + "loss": 0.2413, + "num_input_tokens_seen": 5273856, + "step": 6915 + }, + { + "epoch": 14.386694386694387, + "grad_norm": 0.0005284688086248934, + "learning_rate": 0.2783920320606375, + "loss": 0.2846, + "num_input_tokens_seen": 5277792, + "step": 6920 + }, + { + "epoch": 14.397089397089397, + "grad_norm": 0.0003488158399704844, + "learning_rate": 0.2783615645916852, + "loss": 0.2499, + "num_input_tokens_seen": 5281504, + "step": 6925 + }, + { + "epoch": 14.407484407484407, + "grad_norm": 0.0012154263677075505, + "learning_rate": 0.2783310773277666, + "loss": 0.2734, + "num_input_tokens_seen": 5285280, + "step": 6930 + }, + { + "epoch": 14.417879417879417, + "grad_norm": 0.00025995244504883885, + "learning_rate": 0.2783005702735831, + "loss": 0.269, + "num_input_tokens_seen": 5289312, + "step": 6935 + }, + { + "epoch": 14.428274428274428, + "grad_norm": 0.000442017539171502, + "learning_rate": 0.2782700434338394, + "loss": 0.2717, + "num_input_tokens_seen": 5293248, + "step": 6940 + }, + { + "epoch": 14.43866943866944, + "grad_norm": 0.0014491007896140218, + "learning_rate": 0.278239496813243, + "loss": 0.2769, + "num_input_tokens_seen": 5296992, + "step": 6945 + }, + { + "epoch": 14.44906444906445, + "grad_norm": 5.974783198325895e-05, + "learning_rate": 0.27820893041650463, + "loss": 0.2615, + "num_input_tokens_seen": 5300704, + "step": 6950 + }, + { + "epoch": 14.45945945945946, + "grad_norm": 0.00010668839240679517, + "learning_rate": 0.27817834424833804, + "loss": 0.2766, + "num_input_tokens_seen": 5304416, + "step": 6955 + }, + { + "epoch": 14.46985446985447, + "grad_norm": 0.0001003067081910558, + "learning_rate": 0.27814773831345996, + "loss": 0.2709, + "num_input_tokens_seen": 5308192, + "step": 6960 + }, + { + "epoch": 14.48024948024948, + "grad_norm": 0.00045547171612270176, + "learning_rate": 0.2781171126165902, + "loss": 0.2867, + "num_input_tokens_seen": 5311968, + "step": 6965 + }, + { + "epoch": 14.49064449064449, + "grad_norm": 0.0005999349523335695, + "learning_rate": 0.2780864671624517, + "loss": 0.2887, + "num_input_tokens_seen": 5315808, + "step": 6970 + }, + { + "epoch": 14.5010395010395, + "grad_norm": 0.0009601388592272997, + "learning_rate": 0.27805580195577034, + "loss": 0.2742, + "num_input_tokens_seen": 5319552, + "step": 6975 + }, + { + "epoch": 14.511434511434512, + "grad_norm": 0.0010883768554776907, + "learning_rate": 0.2780251170012751, + "loss": 0.2862, + "num_input_tokens_seen": 5323328, + "step": 6980 + }, + { + "epoch": 14.521829521829522, + "grad_norm": 0.0006590487901121378, + "learning_rate": 0.27799441230369787, + "loss": 0.2781, + "num_input_tokens_seen": 5327296, + "step": 6985 + }, + { + "epoch": 14.532224532224532, + "grad_norm": 0.001135352998971939, + "learning_rate": 0.27796368786777387, + "loss": 0.2862, + "num_input_tokens_seen": 5331168, + "step": 6990 + }, + { + "epoch": 14.542619542619542, + "grad_norm": 0.0002866137365344912, + "learning_rate": 0.277932943698241, + "loss": 0.2378, + "num_input_tokens_seen": 5335072, + "step": 6995 + }, + { + "epoch": 14.553014553014552, + "grad_norm": 0.00010941985237877816, + "learning_rate": 0.2779021797998406, + "loss": 0.297, + "num_input_tokens_seen": 5339008, + "step": 7000 + }, + { + "epoch": 14.553014553014552, + "eval_loss": 0.25118035078048706, + "eval_runtime": 13.4163, + "eval_samples_per_second": 63.803, + "eval_steps_per_second": 15.951, + "num_input_tokens_seen": 5339008, + "step": 7000 + }, + { + "epoch": 14.563409563409563, + "grad_norm": 0.00032486781128682196, + "learning_rate": 0.2778713961773167, + "loss": 0.2902, + "num_input_tokens_seen": 5343072, + "step": 7005 + }, + { + "epoch": 14.573804573804575, + "grad_norm": 0.0010244246805086732, + "learning_rate": 0.2778405928354166, + "loss": 0.2826, + "num_input_tokens_seen": 5346944, + "step": 7010 + }, + { + "epoch": 14.584199584199585, + "grad_norm": 0.0006725835264660418, + "learning_rate": 0.27780976977889055, + "loss": 0.2716, + "num_input_tokens_seen": 5350720, + "step": 7015 + }, + { + "epoch": 14.594594594594595, + "grad_norm": 0.001178294070996344, + "learning_rate": 0.27777892701249185, + "loss": 0.268, + "num_input_tokens_seen": 5354624, + "step": 7020 + }, + { + "epoch": 14.604989604989605, + "grad_norm": 9.888135537039489e-05, + "learning_rate": 0.2777480645409768, + "loss": 0.2664, + "num_input_tokens_seen": 5358432, + "step": 7025 + }, + { + "epoch": 14.615384615384615, + "grad_norm": 0.001524461666122079, + "learning_rate": 0.27771718236910486, + "loss": 0.2794, + "num_input_tokens_seen": 5362272, + "step": 7030 + }, + { + "epoch": 14.625779625779625, + "grad_norm": 0.0018248377600684762, + "learning_rate": 0.27768628050163835, + "loss": 0.285, + "num_input_tokens_seen": 5366080, + "step": 7035 + }, + { + "epoch": 14.636174636174637, + "grad_norm": 0.0005425760755315423, + "learning_rate": 0.2776553589433428, + "loss": 0.2868, + "num_input_tokens_seen": 5369952, + "step": 7040 + }, + { + "epoch": 14.646569646569647, + "grad_norm": 0.0002823318645823747, + "learning_rate": 0.27762441769898666, + "loss": 0.269, + "num_input_tokens_seen": 5373792, + "step": 7045 + }, + { + "epoch": 14.656964656964657, + "grad_norm": 0.0007993949111551046, + "learning_rate": 0.2775934567733415, + "loss": 0.2653, + "num_input_tokens_seen": 5377696, + "step": 7050 + }, + { + "epoch": 14.667359667359667, + "grad_norm": 0.0005036405054852366, + "learning_rate": 0.2775624761711819, + "loss": 0.2945, + "num_input_tokens_seen": 5381568, + "step": 7055 + }, + { + "epoch": 14.677754677754677, + "grad_norm": 0.0004559840017464012, + "learning_rate": 0.2775314758972854, + "loss": 0.2559, + "num_input_tokens_seen": 5385248, + "step": 7060 + }, + { + "epoch": 14.688149688149688, + "grad_norm": 0.00017250121163669974, + "learning_rate": 0.2775004559564327, + "loss": 0.2511, + "num_input_tokens_seen": 5389024, + "step": 7065 + }, + { + "epoch": 14.698544698544698, + "grad_norm": 0.0004886345122940838, + "learning_rate": 0.2774694163534073, + "loss": 0.2766, + "num_input_tokens_seen": 5392896, + "step": 7070 + }, + { + "epoch": 14.70893970893971, + "grad_norm": 0.00036456025554798543, + "learning_rate": 0.27743835709299614, + "loss": 0.2798, + "num_input_tokens_seen": 5396768, + "step": 7075 + }, + { + "epoch": 14.71933471933472, + "grad_norm": 0.0020755392033606768, + "learning_rate": 0.2774072781799888, + "loss": 0.2768, + "num_input_tokens_seen": 5400640, + "step": 7080 + }, + { + "epoch": 14.72972972972973, + "grad_norm": 0.0006276414496824145, + "learning_rate": 0.27737617961917804, + "loss": 0.282, + "num_input_tokens_seen": 5404448, + "step": 7085 + }, + { + "epoch": 14.74012474012474, + "grad_norm": 0.0009041255689226091, + "learning_rate": 0.27734506141535964, + "loss": 0.2608, + "num_input_tokens_seen": 5408224, + "step": 7090 + }, + { + "epoch": 14.75051975051975, + "grad_norm": 0.0009239327046088874, + "learning_rate": 0.2773139235733325, + "loss": 0.2736, + "num_input_tokens_seen": 5412064, + "step": 7095 + }, + { + "epoch": 14.76091476091476, + "grad_norm": 0.0006969145033508539, + "learning_rate": 0.2772827660978984, + "loss": 0.2683, + "num_input_tokens_seen": 5415808, + "step": 7100 + }, + { + "epoch": 14.771309771309772, + "grad_norm": 0.0007697590626776218, + "learning_rate": 0.27725158899386226, + "loss": 0.2809, + "num_input_tokens_seen": 5419552, + "step": 7105 + }, + { + "epoch": 14.781704781704782, + "grad_norm": 0.0008874837076291442, + "learning_rate": 0.27722039226603196, + "loss": 0.2784, + "num_input_tokens_seen": 5423296, + "step": 7110 + }, + { + "epoch": 14.792099792099792, + "grad_norm": 0.0005333812441676855, + "learning_rate": 0.2771891759192184, + "loss": 0.2919, + "num_input_tokens_seen": 5427008, + "step": 7115 + }, + { + "epoch": 14.802494802494802, + "grad_norm": 0.0011189348297193646, + "learning_rate": 0.2771579399582355, + "loss": 0.2493, + "num_input_tokens_seen": 5430784, + "step": 7120 + }, + { + "epoch": 14.812889812889813, + "grad_norm": 9.423021401744336e-05, + "learning_rate": 0.2771266843879004, + "loss": 0.3178, + "num_input_tokens_seen": 5434528, + "step": 7125 + }, + { + "epoch": 14.823284823284823, + "grad_norm": 0.0003147637180518359, + "learning_rate": 0.2770954092130329, + "loss": 0.2551, + "num_input_tokens_seen": 5438304, + "step": 7130 + }, + { + "epoch": 14.833679833679835, + "grad_norm": 0.000748441438190639, + "learning_rate": 0.27706411443845613, + "loss": 0.2665, + "num_input_tokens_seen": 5442080, + "step": 7135 + }, + { + "epoch": 14.844074844074845, + "grad_norm": 0.00019327978952787817, + "learning_rate": 0.27703280006899617, + "loss": 0.2796, + "num_input_tokens_seen": 5445952, + "step": 7140 + }, + { + "epoch": 14.854469854469855, + "grad_norm": 0.00251404638402164, + "learning_rate": 0.277001466109482, + "loss": 0.2741, + "num_input_tokens_seen": 5449728, + "step": 7145 + }, + { + "epoch": 14.864864864864865, + "grad_norm": 0.0012029935605823994, + "learning_rate": 0.2769701125647458, + "loss": 0.2416, + "num_input_tokens_seen": 5453440, + "step": 7150 + }, + { + "epoch": 14.875259875259875, + "grad_norm": 0.00038265660987235606, + "learning_rate": 0.27693873943962266, + "loss": 0.2996, + "num_input_tokens_seen": 5457440, + "step": 7155 + }, + { + "epoch": 14.885654885654885, + "grad_norm": 0.00014777277829125524, + "learning_rate": 0.2769073467389506, + "loss": 0.2727, + "num_input_tokens_seen": 5461344, + "step": 7160 + }, + { + "epoch": 14.896049896049895, + "grad_norm": 0.0017229628283530474, + "learning_rate": 0.2768759344675709, + "loss": 0.2927, + "num_input_tokens_seen": 5465280, + "step": 7165 + }, + { + "epoch": 14.906444906444907, + "grad_norm": 0.000883397413417697, + "learning_rate": 0.27684450263032767, + "loss": 0.2671, + "num_input_tokens_seen": 5468896, + "step": 7170 + }, + { + "epoch": 14.916839916839917, + "grad_norm": 0.0008071648189797997, + "learning_rate": 0.2768130512320682, + "loss": 0.2594, + "num_input_tokens_seen": 5472608, + "step": 7175 + }, + { + "epoch": 14.927234927234927, + "grad_norm": 0.0005227037472650409, + "learning_rate": 0.27678158027764244, + "loss": 0.2937, + "num_input_tokens_seen": 5476352, + "step": 7180 + }, + { + "epoch": 14.937629937629938, + "grad_norm": 0.00016350824444089085, + "learning_rate": 0.27675008977190385, + "loss": 0.2686, + "num_input_tokens_seen": 5480032, + "step": 7185 + }, + { + "epoch": 14.948024948024948, + "grad_norm": 0.00031125274836085737, + "learning_rate": 0.2767185797197086, + "loss": 0.2768, + "num_input_tokens_seen": 5483808, + "step": 7190 + }, + { + "epoch": 14.958419958419958, + "grad_norm": 0.00030609374516643584, + "learning_rate": 0.2766870501259159, + "loss": 0.2628, + "num_input_tokens_seen": 5487584, + "step": 7195 + }, + { + "epoch": 14.96881496881497, + "grad_norm": 0.00020283118647057563, + "learning_rate": 0.276655500995388, + "loss": 0.2724, + "num_input_tokens_seen": 5491424, + "step": 7200 + }, + { + "epoch": 14.96881496881497, + "eval_loss": 0.24888208508491516, + "eval_runtime": 13.4766, + "eval_samples_per_second": 63.517, + "eval_steps_per_second": 15.879, + "num_input_tokens_seen": 5491424, + "step": 7200 + }, + { + "epoch": 14.97920997920998, + "grad_norm": 0.000280176434898749, + "learning_rate": 0.27662393233299015, + "loss": 0.2685, + "num_input_tokens_seen": 5495296, + "step": 7205 + }, + { + "epoch": 14.98960498960499, + "grad_norm": 0.0021873991936445236, + "learning_rate": 0.27659234414359074, + "loss": 0.2794, + "num_input_tokens_seen": 5499232, + "step": 7210 + }, + { + "epoch": 15.0, + "grad_norm": 0.0002343564701732248, + "learning_rate": 0.27656073643206097, + "loss": 0.2657, + "num_input_tokens_seen": 5503048, + "step": 7215 + }, + { + "epoch": 15.01039501039501, + "grad_norm": 0.00028844503685832024, + "learning_rate": 0.27652910920327517, + "loss": 0.2634, + "num_input_tokens_seen": 5507048, + "step": 7220 + }, + { + "epoch": 15.02079002079002, + "grad_norm": 0.0009478189167566597, + "learning_rate": 0.2764974624621107, + "loss": 0.2657, + "num_input_tokens_seen": 5510952, + "step": 7225 + }, + { + "epoch": 15.03118503118503, + "grad_norm": 0.0002082430146401748, + "learning_rate": 0.2764657962134479, + "loss": 0.2751, + "num_input_tokens_seen": 5514792, + "step": 7230 + }, + { + "epoch": 15.041580041580042, + "grad_norm": 0.0013066637329757214, + "learning_rate": 0.27643411046217, + "loss": 0.2744, + "num_input_tokens_seen": 5518600, + "step": 7235 + }, + { + "epoch": 15.051975051975052, + "grad_norm": 0.0003306009457446635, + "learning_rate": 0.27640240521316334, + "loss": 0.2549, + "num_input_tokens_seen": 5522344, + "step": 7240 + }, + { + "epoch": 15.062370062370062, + "grad_norm": 0.0006251315353438258, + "learning_rate": 0.2763706804713174, + "loss": 0.2798, + "num_input_tokens_seen": 5526152, + "step": 7245 + }, + { + "epoch": 15.072765072765073, + "grad_norm": 0.0002411407040199265, + "learning_rate": 0.2763389362415245, + "loss": 0.2819, + "num_input_tokens_seen": 5529832, + "step": 7250 + }, + { + "epoch": 15.083160083160083, + "grad_norm": 0.000795442727394402, + "learning_rate": 0.27630717252867987, + "loss": 0.2973, + "num_input_tokens_seen": 5533576, + "step": 7255 + }, + { + "epoch": 15.093555093555093, + "grad_norm": 0.0003539549943525344, + "learning_rate": 0.276275389337682, + "loss": 0.273, + "num_input_tokens_seen": 5537512, + "step": 7260 + }, + { + "epoch": 15.103950103950105, + "grad_norm": 0.0010497792391106486, + "learning_rate": 0.2762435866734322, + "loss": 0.3076, + "num_input_tokens_seen": 5541256, + "step": 7265 + }, + { + "epoch": 15.114345114345115, + "grad_norm": 0.00047377869486808777, + "learning_rate": 0.27621176454083485, + "loss": 0.2641, + "num_input_tokens_seen": 5545160, + "step": 7270 + }, + { + "epoch": 15.124740124740125, + "grad_norm": 0.00047894002636894584, + "learning_rate": 0.2761799229447973, + "loss": 0.2676, + "num_input_tokens_seen": 5549032, + "step": 7275 + }, + { + "epoch": 15.135135135135135, + "grad_norm": 0.0009624660015106201, + "learning_rate": 0.27614806189023006, + "loss": 0.3008, + "num_input_tokens_seen": 5552936, + "step": 7280 + }, + { + "epoch": 15.145530145530145, + "grad_norm": 0.0003097022126894444, + "learning_rate": 0.27611618138204636, + "loss": 0.2847, + "num_input_tokens_seen": 5556744, + "step": 7285 + }, + { + "epoch": 15.155925155925155, + "grad_norm": 0.0007015501614660025, + "learning_rate": 0.2760842814251626, + "loss": 0.2779, + "num_input_tokens_seen": 5560424, + "step": 7290 + }, + { + "epoch": 15.166320166320165, + "grad_norm": 0.0013820198364555836, + "learning_rate": 0.2760523620244982, + "loss": 0.2677, + "num_input_tokens_seen": 5564296, + "step": 7295 + }, + { + "epoch": 15.176715176715177, + "grad_norm": 0.00037687172880396247, + "learning_rate": 0.27602042318497544, + "loss": 0.2634, + "num_input_tokens_seen": 5568264, + "step": 7300 + }, + { + "epoch": 15.187110187110187, + "grad_norm": 0.0014727680245414376, + "learning_rate": 0.2759884649115198, + "loss": 0.2779, + "num_input_tokens_seen": 5572008, + "step": 7305 + }, + { + "epoch": 15.197505197505198, + "grad_norm": 0.00024315298651345074, + "learning_rate": 0.2759564872090596, + "loss": 0.2565, + "num_input_tokens_seen": 5575880, + "step": 7310 + }, + { + "epoch": 15.207900207900208, + "grad_norm": 0.0002517875109333545, + "learning_rate": 0.2759244900825262, + "loss": 0.2737, + "num_input_tokens_seen": 5579592, + "step": 7315 + }, + { + "epoch": 15.218295218295218, + "grad_norm": 0.00039618040318600833, + "learning_rate": 0.2758924735368539, + "loss": 0.2653, + "num_input_tokens_seen": 5583304, + "step": 7320 + }, + { + "epoch": 15.228690228690228, + "grad_norm": 0.00029686916968785226, + "learning_rate": 0.27586043757698014, + "loss": 0.2701, + "num_input_tokens_seen": 5587208, + "step": 7325 + }, + { + "epoch": 15.23908523908524, + "grad_norm": 0.0010557174682617188, + "learning_rate": 0.27582838220784534, + "loss": 0.2777, + "num_input_tokens_seen": 5591016, + "step": 7330 + }, + { + "epoch": 15.24948024948025, + "grad_norm": 0.0010066907852888107, + "learning_rate": 0.27579630743439265, + "loss": 0.2736, + "num_input_tokens_seen": 5594824, + "step": 7335 + }, + { + "epoch": 15.25987525987526, + "grad_norm": 0.00014965262380428612, + "learning_rate": 0.2757642132615686, + "loss": 0.2507, + "num_input_tokens_seen": 5598600, + "step": 7340 + }, + { + "epoch": 15.27027027027027, + "grad_norm": 0.0001960218942258507, + "learning_rate": 0.2757320996943223, + "loss": 0.2588, + "num_input_tokens_seen": 5602280, + "step": 7345 + }, + { + "epoch": 15.28066528066528, + "grad_norm": 0.0004989554290659726, + "learning_rate": 0.2756999667376062, + "loss": 0.254, + "num_input_tokens_seen": 5606312, + "step": 7350 + }, + { + "epoch": 15.29106029106029, + "grad_norm": 0.0005768209230154753, + "learning_rate": 0.2756678143963756, + "loss": 0.2364, + "num_input_tokens_seen": 5610120, + "step": 7355 + }, + { + "epoch": 15.301455301455302, + "grad_norm": 0.00041111421887762845, + "learning_rate": 0.2756356426755888, + "loss": 0.2546, + "num_input_tokens_seen": 5614024, + "step": 7360 + }, + { + "epoch": 15.311850311850312, + "grad_norm": 0.0007249697227962315, + "learning_rate": 0.27560345158020705, + "loss": 0.2577, + "num_input_tokens_seen": 5617896, + "step": 7365 + }, + { + "epoch": 15.322245322245323, + "grad_norm": 0.0007824570639058948, + "learning_rate": 0.27557124111519465, + "loss": 0.2714, + "num_input_tokens_seen": 5621832, + "step": 7370 + }, + { + "epoch": 15.332640332640333, + "grad_norm": 0.00031998445047065616, + "learning_rate": 0.27553901128551883, + "loss": 0.2661, + "num_input_tokens_seen": 5625640, + "step": 7375 + }, + { + "epoch": 15.343035343035343, + "grad_norm": 0.0008927280432544649, + "learning_rate": 0.2755067620961498, + "loss": 0.2418, + "num_input_tokens_seen": 5629480, + "step": 7380 + }, + { + "epoch": 15.353430353430353, + "grad_norm": 0.00022247085871640593, + "learning_rate": 0.27547449355206094, + "loss": 0.2928, + "num_input_tokens_seen": 5633224, + "step": 7385 + }, + { + "epoch": 15.363825363825363, + "grad_norm": 0.0015979583840817213, + "learning_rate": 0.2754422056582283, + "loss": 0.2797, + "num_input_tokens_seen": 5636936, + "step": 7390 + }, + { + "epoch": 15.374220374220375, + "grad_norm": 0.00013116904301568866, + "learning_rate": 0.27540989841963115, + "loss": 0.273, + "num_input_tokens_seen": 5640712, + "step": 7395 + }, + { + "epoch": 15.384615384615385, + "grad_norm": 0.00010890357953030616, + "learning_rate": 0.27537757184125167, + "loss": 0.2623, + "num_input_tokens_seen": 5644520, + "step": 7400 + }, + { + "epoch": 15.384615384615385, + "eval_loss": 0.24839846789836884, + "eval_runtime": 13.4346, + "eval_samples_per_second": 63.716, + "eval_steps_per_second": 15.929, + "num_input_tokens_seen": 5644520, + "step": 7400 + }, + { + "epoch": 15.395010395010395, + "grad_norm": 0.00021131696121301502, + "learning_rate": 0.275345225928075, + "loss": 0.2648, + "num_input_tokens_seen": 5648264, + "step": 7405 + }, + { + "epoch": 15.405405405405405, + "grad_norm": 0.0008523308206349611, + "learning_rate": 0.2753128606850893, + "loss": 0.2736, + "num_input_tokens_seen": 5652072, + "step": 7410 + }, + { + "epoch": 15.415800415800415, + "grad_norm": 0.0011783391237258911, + "learning_rate": 0.2752804761172858, + "loss": 0.2763, + "num_input_tokens_seen": 5656008, + "step": 7415 + }, + { + "epoch": 15.426195426195425, + "grad_norm": 0.0001437998580513522, + "learning_rate": 0.27524807222965836, + "loss": 0.2707, + "num_input_tokens_seen": 5659848, + "step": 7420 + }, + { + "epoch": 15.436590436590437, + "grad_norm": 0.0004677558899857104, + "learning_rate": 0.27521564902720436, + "loss": 0.2402, + "num_input_tokens_seen": 5663592, + "step": 7425 + }, + { + "epoch": 15.446985446985448, + "grad_norm": 0.0001335811975877732, + "learning_rate": 0.2751832065149236, + "loss": 0.2776, + "num_input_tokens_seen": 5667272, + "step": 7430 + }, + { + "epoch": 15.457380457380458, + "grad_norm": 0.00037285161670297384, + "learning_rate": 0.2751507446978193, + "loss": 0.2617, + "num_input_tokens_seen": 5671208, + "step": 7435 + }, + { + "epoch": 15.467775467775468, + "grad_norm": 0.00011653076944639906, + "learning_rate": 0.2751182635808974, + "loss": 0.2621, + "num_input_tokens_seen": 5674984, + "step": 7440 + }, + { + "epoch": 15.478170478170478, + "grad_norm": 0.0005675626453012228, + "learning_rate": 0.27508576316916694, + "loss": 0.2913, + "num_input_tokens_seen": 5678856, + "step": 7445 + }, + { + "epoch": 15.488565488565488, + "grad_norm": 8.135367534123361e-05, + "learning_rate": 0.2750532434676399, + "loss": 0.2808, + "num_input_tokens_seen": 5682696, + "step": 7450 + }, + { + "epoch": 15.4989604989605, + "grad_norm": 0.0005800941144116223, + "learning_rate": 0.27502070448133115, + "loss": 0.2593, + "num_input_tokens_seen": 5686472, + "step": 7455 + }, + { + "epoch": 15.50935550935551, + "grad_norm": 0.0006041646702215075, + "learning_rate": 0.2749881462152587, + "loss": 0.2907, + "num_input_tokens_seen": 5690312, + "step": 7460 + }, + { + "epoch": 15.51975051975052, + "grad_norm": 0.0007312985835596919, + "learning_rate": 0.2749555686744434, + "loss": 0.2664, + "num_input_tokens_seen": 5694120, + "step": 7465 + }, + { + "epoch": 15.53014553014553, + "grad_norm": 0.00031196887721307576, + "learning_rate": 0.2749229718639091, + "loss": 0.2269, + "num_input_tokens_seen": 5697992, + "step": 7470 + }, + { + "epoch": 15.54054054054054, + "grad_norm": 0.0002736408787313849, + "learning_rate": 0.27489035578868265, + "loss": 0.2817, + "num_input_tokens_seen": 5701800, + "step": 7475 + }, + { + "epoch": 15.55093555093555, + "grad_norm": 0.0003088978410232812, + "learning_rate": 0.2748577204537939, + "loss": 0.2589, + "num_input_tokens_seen": 5705672, + "step": 7480 + }, + { + "epoch": 15.56133056133056, + "grad_norm": 0.0001431064010830596, + "learning_rate": 0.2748250658642756, + "loss": 0.2604, + "num_input_tokens_seen": 5709352, + "step": 7485 + }, + { + "epoch": 15.571725571725572, + "grad_norm": 0.00029914206243120134, + "learning_rate": 0.2747923920251634, + "loss": 0.2788, + "num_input_tokens_seen": 5713160, + "step": 7490 + }, + { + "epoch": 15.582120582120583, + "grad_norm": 0.00027012015925720334, + "learning_rate": 0.27475969894149627, + "loss": 0.2761, + "num_input_tokens_seen": 5716872, + "step": 7495 + }, + { + "epoch": 15.592515592515593, + "grad_norm": 0.0009495631093159318, + "learning_rate": 0.2747269866183156, + "loss": 0.2775, + "num_input_tokens_seen": 5720648, + "step": 7500 + }, + { + "epoch": 15.602910602910603, + "grad_norm": 0.0007115293992683291, + "learning_rate": 0.27469425506066625, + "loss": 0.2586, + "num_input_tokens_seen": 5724456, + "step": 7505 + }, + { + "epoch": 15.613305613305613, + "grad_norm": 0.00032747091609053314, + "learning_rate": 0.27466150427359576, + "loss": 0.2771, + "num_input_tokens_seen": 5728392, + "step": 7510 + }, + { + "epoch": 15.623700623700623, + "grad_norm": 0.00010101830412168056, + "learning_rate": 0.2746287342621547, + "loss": 0.2651, + "num_input_tokens_seen": 5732168, + "step": 7515 + }, + { + "epoch": 15.634095634095633, + "grad_norm": 0.0008337240433320403, + "learning_rate": 0.2745959450313966, + "loss": 0.2629, + "num_input_tokens_seen": 5735848, + "step": 7520 + }, + { + "epoch": 15.644490644490645, + "grad_norm": 0.0012070474913343787, + "learning_rate": 0.27456313658637804, + "loss": 0.277, + "num_input_tokens_seen": 5739720, + "step": 7525 + }, + { + "epoch": 15.654885654885655, + "grad_norm": 0.0004778236907441169, + "learning_rate": 0.27453030893215846, + "loss": 0.2817, + "num_input_tokens_seen": 5743368, + "step": 7530 + }, + { + "epoch": 15.665280665280665, + "grad_norm": 0.00017180699796881527, + "learning_rate": 0.2744974620738003, + "loss": 0.2627, + "num_input_tokens_seen": 5747176, + "step": 7535 + }, + { + "epoch": 15.675675675675675, + "grad_norm": 0.0009562495979480445, + "learning_rate": 0.27446459601636897, + "loss": 0.2576, + "num_input_tokens_seen": 5751144, + "step": 7540 + }, + { + "epoch": 15.686070686070686, + "grad_norm": 0.00031361792935058475, + "learning_rate": 0.2744317107649328, + "loss": 0.2596, + "num_input_tokens_seen": 5754952, + "step": 7545 + }, + { + "epoch": 15.696465696465696, + "grad_norm": 0.00020459649385884404, + "learning_rate": 0.2743988063245631, + "loss": 0.2636, + "num_input_tokens_seen": 5758792, + "step": 7550 + }, + { + "epoch": 15.706860706860708, + "grad_norm": 0.0005400121444836259, + "learning_rate": 0.2743658827003342, + "loss": 0.2946, + "num_input_tokens_seen": 5762600, + "step": 7555 + }, + { + "epoch": 15.717255717255718, + "grad_norm": 0.00036298486520536244, + "learning_rate": 0.27433293989732327, + "loss": 0.2715, + "num_input_tokens_seen": 5766376, + "step": 7560 + }, + { + "epoch": 15.727650727650728, + "grad_norm": 0.0008316165767610073, + "learning_rate": 0.27429997792061056, + "loss": 0.2731, + "num_input_tokens_seen": 5770056, + "step": 7565 + }, + { + "epoch": 15.738045738045738, + "grad_norm": 0.0005599592113867402, + "learning_rate": 0.27426699677527927, + "loss": 0.2803, + "num_input_tokens_seen": 5773992, + "step": 7570 + }, + { + "epoch": 15.748440748440748, + "grad_norm": 0.0001371019461657852, + "learning_rate": 0.2742339964664154, + "loss": 0.2791, + "num_input_tokens_seen": 5777736, + "step": 7575 + }, + { + "epoch": 15.758835758835758, + "grad_norm": 0.000719712523277849, + "learning_rate": 0.274200976999108, + "loss": 0.2826, + "num_input_tokens_seen": 5781608, + "step": 7580 + }, + { + "epoch": 15.76923076923077, + "grad_norm": 0.00038290329393930733, + "learning_rate": 0.27416793837844916, + "loss": 0.271, + "num_input_tokens_seen": 5785288, + "step": 7585 + }, + { + "epoch": 15.77962577962578, + "grad_norm": 0.0005798368365503848, + "learning_rate": 0.27413488060953384, + "loss": 0.2681, + "num_input_tokens_seen": 5789192, + "step": 7590 + }, + { + "epoch": 15.79002079002079, + "grad_norm": 0.0005446581053547561, + "learning_rate": 0.27410180369745996, + "loss": 0.2505, + "num_input_tokens_seen": 5792968, + "step": 7595 + }, + { + "epoch": 15.8004158004158, + "grad_norm": 0.0006107335793785751, + "learning_rate": 0.27406870764732844, + "loss": 0.2063, + "num_input_tokens_seen": 5796744, + "step": 7600 + }, + { + "epoch": 15.8004158004158, + "eval_loss": 0.2708362638950348, + "eval_runtime": 13.4023, + "eval_samples_per_second": 63.87, + "eval_steps_per_second": 15.967, + "num_input_tokens_seen": 5796744, + "step": 7600 + }, + { + "epoch": 15.81081081081081, + "grad_norm": 0.0011588734341785312, + "learning_rate": 0.27403559246424297, + "loss": 0.289, + "num_input_tokens_seen": 5800552, + "step": 7605 + }, + { + "epoch": 15.82120582120582, + "grad_norm": 0.00027321293600834906, + "learning_rate": 0.2740024581533105, + "loss": 0.2921, + "num_input_tokens_seen": 5804360, + "step": 7610 + }, + { + "epoch": 15.83160083160083, + "grad_norm": 0.0006256555789150298, + "learning_rate": 0.2739693047196406, + "loss": 0.2967, + "num_input_tokens_seen": 5808328, + "step": 7615 + }, + { + "epoch": 15.841995841995843, + "grad_norm": 0.00046540325274690986, + "learning_rate": 0.27393613216834606, + "loss": 0.2725, + "num_input_tokens_seen": 5812136, + "step": 7620 + }, + { + "epoch": 15.852390852390853, + "grad_norm": 0.002030816627666354, + "learning_rate": 0.2739029405045424, + "loss": 0.2905, + "num_input_tokens_seen": 5815912, + "step": 7625 + }, + { + "epoch": 15.862785862785863, + "grad_norm": 0.00039951325743459165, + "learning_rate": 0.2738697297333483, + "loss": 0.2606, + "num_input_tokens_seen": 5819624, + "step": 7630 + }, + { + "epoch": 15.873180873180873, + "grad_norm": 0.00013634399510920048, + "learning_rate": 0.2738364998598852, + "loss": 0.2707, + "num_input_tokens_seen": 5823400, + "step": 7635 + }, + { + "epoch": 15.883575883575883, + "grad_norm": 0.0010050374548882246, + "learning_rate": 0.27380325088927765, + "loss": 0.2715, + "num_input_tokens_seen": 5827272, + "step": 7640 + }, + { + "epoch": 15.893970893970893, + "grad_norm": 0.0004820208705496043, + "learning_rate": 0.27376998282665294, + "loss": 0.2729, + "num_input_tokens_seen": 5831080, + "step": 7645 + }, + { + "epoch": 15.904365904365905, + "grad_norm": 0.0012906744377687573, + "learning_rate": 0.27373669567714154, + "loss": 0.2669, + "num_input_tokens_seen": 5834760, + "step": 7650 + }, + { + "epoch": 15.914760914760915, + "grad_norm": 0.0008492125198245049, + "learning_rate": 0.27370338944587663, + "loss": 0.2603, + "num_input_tokens_seen": 5838472, + "step": 7655 + }, + { + "epoch": 15.925155925155925, + "grad_norm": 0.0007136305794119835, + "learning_rate": 0.27367006413799455, + "loss": 0.2649, + "num_input_tokens_seen": 5842248, + "step": 7660 + }, + { + "epoch": 15.935550935550935, + "grad_norm": 0.0005543690058402717, + "learning_rate": 0.2736367197586345, + "loss": 0.263, + "num_input_tokens_seen": 5846056, + "step": 7665 + }, + { + "epoch": 15.945945945945946, + "grad_norm": 0.0001712001539999619, + "learning_rate": 0.2736033563129385, + "loss": 0.2767, + "num_input_tokens_seen": 5849704, + "step": 7670 + }, + { + "epoch": 15.956340956340956, + "grad_norm": 0.0004004684742540121, + "learning_rate": 0.27356997380605164, + "loss": 0.2759, + "num_input_tokens_seen": 5853512, + "step": 7675 + }, + { + "epoch": 15.966735966735968, + "grad_norm": 0.00021644837397616357, + "learning_rate": 0.27353657224312194, + "loss": 0.2459, + "num_input_tokens_seen": 5857544, + "step": 7680 + }, + { + "epoch": 15.977130977130978, + "grad_norm": 0.00010654184006853029, + "learning_rate": 0.2735031516293004, + "loss": 0.2886, + "num_input_tokens_seen": 5861384, + "step": 7685 + }, + { + "epoch": 15.987525987525988, + "grad_norm": 0.00025882787303999066, + "learning_rate": 0.2734697119697408, + "loss": 0.2721, + "num_input_tokens_seen": 5865128, + "step": 7690 + }, + { + "epoch": 15.997920997920998, + "grad_norm": 0.0003930949023924768, + "learning_rate": 0.27343625326959997, + "loss": 0.2756, + "num_input_tokens_seen": 5869000, + "step": 7695 + }, + { + "epoch": 16.008316008316008, + "grad_norm": 0.0003782971471082419, + "learning_rate": 0.27340277553403775, + "loss": 0.2788, + "num_input_tokens_seen": 5872928, + "step": 7700 + }, + { + "epoch": 16.01871101871102, + "grad_norm": 7.335251575568691e-05, + "learning_rate": 0.2733692787682167, + "loss": 0.2612, + "num_input_tokens_seen": 5877024, + "step": 7705 + }, + { + "epoch": 16.02910602910603, + "grad_norm": 0.0006023618625476956, + "learning_rate": 0.27333576297730255, + "loss": 0.276, + "num_input_tokens_seen": 5880928, + "step": 7710 + }, + { + "epoch": 16.03950103950104, + "grad_norm": 0.0004253872612025589, + "learning_rate": 0.2733022281664638, + "loss": 0.2591, + "num_input_tokens_seen": 5884704, + "step": 7715 + }, + { + "epoch": 16.04989604989605, + "grad_norm": 0.0008588171331211925, + "learning_rate": 0.273268674340872, + "loss": 0.2762, + "num_input_tokens_seen": 5888544, + "step": 7720 + }, + { + "epoch": 16.06029106029106, + "grad_norm": 0.0009492579847574234, + "learning_rate": 0.27323510150570146, + "loss": 0.285, + "num_input_tokens_seen": 5892352, + "step": 7725 + }, + { + "epoch": 16.070686070686072, + "grad_norm": 0.0006792630301788449, + "learning_rate": 0.27320150966612966, + "loss": 0.2829, + "num_input_tokens_seen": 5896128, + "step": 7730 + }, + { + "epoch": 16.08108108108108, + "grad_norm": 0.0004948316491208971, + "learning_rate": 0.2731678988273368, + "loss": 0.2592, + "num_input_tokens_seen": 5900064, + "step": 7735 + }, + { + "epoch": 16.091476091476093, + "grad_norm": 0.0008550684433430433, + "learning_rate": 0.27313426899450605, + "loss": 0.3345, + "num_input_tokens_seen": 5903712, + "step": 7740 + }, + { + "epoch": 16.1018711018711, + "grad_norm": 0.00038054073229432106, + "learning_rate": 0.27310062017282366, + "loss": 0.2934, + "num_input_tokens_seen": 5907552, + "step": 7745 + }, + { + "epoch": 16.112266112266113, + "grad_norm": 0.0004169708117842674, + "learning_rate": 0.2730669523674787, + "loss": 0.2648, + "num_input_tokens_seen": 5911264, + "step": 7750 + }, + { + "epoch": 16.12266112266112, + "grad_norm": 0.0009244192042388022, + "learning_rate": 0.2730332655836631, + "loss": 0.2928, + "num_input_tokens_seen": 5915008, + "step": 7755 + }, + { + "epoch": 16.133056133056133, + "grad_norm": 0.0001721863227430731, + "learning_rate": 0.2729995598265718, + "loss": 0.2555, + "num_input_tokens_seen": 5918816, + "step": 7760 + }, + { + "epoch": 16.143451143451145, + "grad_norm": 0.0008631807286292315, + "learning_rate": 0.2729658351014027, + "loss": 0.2653, + "num_input_tokens_seen": 5922560, + "step": 7765 + }, + { + "epoch": 16.153846153846153, + "grad_norm": 0.0011688689701259136, + "learning_rate": 0.27293209141335656, + "loss": 0.272, + "num_input_tokens_seen": 5926304, + "step": 7770 + }, + { + "epoch": 16.164241164241165, + "grad_norm": 7.41495969123207e-05, + "learning_rate": 0.27289832876763703, + "loss": 0.2706, + "num_input_tokens_seen": 5930176, + "step": 7775 + }, + { + "epoch": 16.174636174636174, + "grad_norm": 0.000559632433578372, + "learning_rate": 0.27286454716945074, + "loss": 0.2639, + "num_input_tokens_seen": 5934080, + "step": 7780 + }, + { + "epoch": 16.185031185031185, + "grad_norm": 0.00014519080286845565, + "learning_rate": 0.27283074662400725, + "loss": 0.274, + "num_input_tokens_seen": 5938080, + "step": 7785 + }, + { + "epoch": 16.195426195426194, + "grad_norm": 0.0015915262047201395, + "learning_rate": 0.2727969271365191, + "loss": 0.2936, + "num_input_tokens_seen": 5941792, + "step": 7790 + }, + { + "epoch": 16.205821205821206, + "grad_norm": 0.00047141005052253604, + "learning_rate": 0.2727630887122016, + "loss": 0.2788, + "num_input_tokens_seen": 5945600, + "step": 7795 + }, + { + "epoch": 16.216216216216218, + "grad_norm": 0.0008207892533391714, + "learning_rate": 0.27272923135627314, + "loss": 0.276, + "num_input_tokens_seen": 5949536, + "step": 7800 + }, + { + "epoch": 16.216216216216218, + "eval_loss": 0.24984216690063477, + "eval_runtime": 13.4177, + "eval_samples_per_second": 63.796, + "eval_steps_per_second": 15.949, + "num_input_tokens_seen": 5949536, + "step": 7800 + }, + { + "epoch": 16.226611226611226, + "grad_norm": 0.00035605349694378674, + "learning_rate": 0.2726953550739548, + "loss": 0.2676, + "num_input_tokens_seen": 5953504, + "step": 7805 + }, + { + "epoch": 16.237006237006238, + "grad_norm": 0.0010749581269919872, + "learning_rate": 0.27266145987047086, + "loss": 0.2728, + "num_input_tokens_seen": 5957216, + "step": 7810 + }, + { + "epoch": 16.247401247401246, + "grad_norm": 0.0012606856180354953, + "learning_rate": 0.27262754575104836, + "loss": 0.2844, + "num_input_tokens_seen": 5961088, + "step": 7815 + }, + { + "epoch": 16.257796257796258, + "grad_norm": 0.00040339259430766106, + "learning_rate": 0.27259361272091726, + "loss": 0.2683, + "num_input_tokens_seen": 5964896, + "step": 7820 + }, + { + "epoch": 16.26819126819127, + "grad_norm": 0.00016065733507275581, + "learning_rate": 0.27255966078531046, + "loss": 0.3041, + "num_input_tokens_seen": 5968768, + "step": 7825 + }, + { + "epoch": 16.27858627858628, + "grad_norm": 0.0006202685763128102, + "learning_rate": 0.2725256899494638, + "loss": 0.2618, + "num_input_tokens_seen": 5972800, + "step": 7830 + }, + { + "epoch": 16.28898128898129, + "grad_norm": 0.000727097736671567, + "learning_rate": 0.272491700218616, + "loss": 0.2975, + "num_input_tokens_seen": 5976608, + "step": 7835 + }, + { + "epoch": 16.2993762993763, + "grad_norm": 0.0005357848131097853, + "learning_rate": 0.27245769159800876, + "loss": 0.2763, + "num_input_tokens_seen": 5980480, + "step": 7840 + }, + { + "epoch": 16.30977130977131, + "grad_norm": 0.0002505619777366519, + "learning_rate": 0.2724236640928865, + "loss": 0.3001, + "num_input_tokens_seen": 5984128, + "step": 7845 + }, + { + "epoch": 16.32016632016632, + "grad_norm": 0.00027863041032105684, + "learning_rate": 0.27238961770849673, + "loss": 0.2463, + "num_input_tokens_seen": 5988064, + "step": 7850 + }, + { + "epoch": 16.33056133056133, + "grad_norm": 4.527588316705078e-05, + "learning_rate": 0.27235555245008997, + "loss": 0.2369, + "num_input_tokens_seen": 5991808, + "step": 7855 + }, + { + "epoch": 16.340956340956343, + "grad_norm": 0.00029998397803865373, + "learning_rate": 0.2723214683229193, + "loss": 0.2976, + "num_input_tokens_seen": 5995616, + "step": 7860 + }, + { + "epoch": 16.35135135135135, + "grad_norm": 0.001439696061424911, + "learning_rate": 0.27228736533224107, + "loss": 0.2644, + "num_input_tokens_seen": 5999488, + "step": 7865 + }, + { + "epoch": 16.361746361746363, + "grad_norm": 0.0008670384413562715, + "learning_rate": 0.27225324348331437, + "loss": 0.2881, + "num_input_tokens_seen": 6003232, + "step": 7870 + }, + { + "epoch": 16.37214137214137, + "grad_norm": 0.0005674163694493473, + "learning_rate": 0.27221910278140116, + "loss": 0.2788, + "num_input_tokens_seen": 6007104, + "step": 7875 + }, + { + "epoch": 16.382536382536383, + "grad_norm": 0.0002761151408776641, + "learning_rate": 0.2721849432317664, + "loss": 0.2599, + "num_input_tokens_seen": 6010880, + "step": 7880 + }, + { + "epoch": 16.39293139293139, + "grad_norm": 0.0007668761536478996, + "learning_rate": 0.2721507648396779, + "loss": 0.277, + "num_input_tokens_seen": 6014560, + "step": 7885 + }, + { + "epoch": 16.403326403326403, + "grad_norm": 0.00029539118986576796, + "learning_rate": 0.27211656761040653, + "loss": 0.2843, + "num_input_tokens_seen": 6018272, + "step": 7890 + }, + { + "epoch": 16.413721413721415, + "grad_norm": 0.0005167144699953496, + "learning_rate": 0.2720823515492257, + "loss": 0.2809, + "num_input_tokens_seen": 6022176, + "step": 7895 + }, + { + "epoch": 16.424116424116423, + "grad_norm": 0.0008906522998586297, + "learning_rate": 0.27204811666141215, + "loss": 0.2531, + "num_input_tokens_seen": 6025984, + "step": 7900 + }, + { + "epoch": 16.434511434511435, + "grad_norm": 0.0008986930479295552, + "learning_rate": 0.2720138629522452, + "loss": 0.2458, + "num_input_tokens_seen": 6029952, + "step": 7905 + }, + { + "epoch": 16.444906444906444, + "grad_norm": 0.00016851077089086175, + "learning_rate": 0.2719795904270073, + "loss": 0.2899, + "num_input_tokens_seen": 6033792, + "step": 7910 + }, + { + "epoch": 16.455301455301456, + "grad_norm": 0.0010905377566814423, + "learning_rate": 0.2719452990909837, + "loss": 0.2609, + "num_input_tokens_seen": 6037664, + "step": 7915 + }, + { + "epoch": 16.465696465696467, + "grad_norm": 0.0002057001693174243, + "learning_rate": 0.2719109889494625, + "loss": 0.2727, + "num_input_tokens_seen": 6041632, + "step": 7920 + }, + { + "epoch": 16.476091476091476, + "grad_norm": 0.0011404691031202674, + "learning_rate": 0.27187666000773475, + "loss": 0.2702, + "num_input_tokens_seen": 6045408, + "step": 7925 + }, + { + "epoch": 16.486486486486488, + "grad_norm": 0.0003962080227211118, + "learning_rate": 0.2718423122710944, + "loss": 0.2657, + "num_input_tokens_seen": 6049152, + "step": 7930 + }, + { + "epoch": 16.496881496881496, + "grad_norm": 0.00016637657245155424, + "learning_rate": 0.2718079457448384, + "loss": 0.2666, + "num_input_tokens_seen": 6053024, + "step": 7935 + }, + { + "epoch": 16.507276507276508, + "grad_norm": 0.0007608078303746879, + "learning_rate": 0.27177356043426637, + "loss": 0.2789, + "num_input_tokens_seen": 6056800, + "step": 7940 + }, + { + "epoch": 16.517671517671516, + "grad_norm": 0.001341905095614493, + "learning_rate": 0.27173915634468104, + "loss": 0.2705, + "num_input_tokens_seen": 6060672, + "step": 7945 + }, + { + "epoch": 16.528066528066528, + "grad_norm": 0.0006301902467384934, + "learning_rate": 0.27170473348138796, + "loss": 0.2423, + "num_input_tokens_seen": 6064480, + "step": 7950 + }, + { + "epoch": 16.53846153846154, + "grad_norm": 0.0005650501698255539, + "learning_rate": 0.27167029184969554, + "loss": 0.2648, + "num_input_tokens_seen": 6068160, + "step": 7955 + }, + { + "epoch": 16.54885654885655, + "grad_norm": 0.00035487956483848393, + "learning_rate": 0.27163583145491504, + "loss": 0.2455, + "num_input_tokens_seen": 6071968, + "step": 7960 + }, + { + "epoch": 16.55925155925156, + "grad_norm": 0.0006121479091234505, + "learning_rate": 0.2716013523023608, + "loss": 0.2575, + "num_input_tokens_seen": 6075712, + "step": 7965 + }, + { + "epoch": 16.56964656964657, + "grad_norm": 0.0007161162793636322, + "learning_rate": 0.27156685439734995, + "loss": 0.2725, + "num_input_tokens_seen": 6079520, + "step": 7970 + }, + { + "epoch": 16.58004158004158, + "grad_norm": 0.0010961029911413789, + "learning_rate": 0.2715323377452024, + "loss": 0.2676, + "num_input_tokens_seen": 6083424, + "step": 7975 + }, + { + "epoch": 16.59043659043659, + "grad_norm": 0.0007766846683807671, + "learning_rate": 0.2714978023512411, + "loss": 0.286, + "num_input_tokens_seen": 6087264, + "step": 7980 + }, + { + "epoch": 16.6008316008316, + "grad_norm": 0.000907827285118401, + "learning_rate": 0.2714632482207918, + "loss": 0.2839, + "num_input_tokens_seen": 6091008, + "step": 7985 + }, + { + "epoch": 16.611226611226613, + "grad_norm": 0.0006716377683915198, + "learning_rate": 0.2714286753591833, + "loss": 0.2888, + "num_input_tokens_seen": 6094816, + "step": 7990 + }, + { + "epoch": 16.62162162162162, + "grad_norm": 0.0005515064112842083, + "learning_rate": 0.27139408377174706, + "loss": 0.2532, + "num_input_tokens_seen": 6098624, + "step": 7995 + }, + { + "epoch": 16.632016632016633, + "grad_norm": 0.0005116141401231289, + "learning_rate": 0.27135947346381756, + "loss": 0.3006, + "num_input_tokens_seen": 6102304, + "step": 8000 + }, + { + "epoch": 16.632016632016633, + "eval_loss": 0.2607586085796356, + "eval_runtime": 13.4437, + "eval_samples_per_second": 63.673, + "eval_steps_per_second": 15.918, + "num_input_tokens_seen": 6102304, + "step": 8000 + }, + { + "epoch": 16.64241164241164, + "grad_norm": 0.0005792885785922408, + "learning_rate": 0.2713248444407322, + "loss": 0.2745, + "num_input_tokens_seen": 6106176, + "step": 8005 + }, + { + "epoch": 16.652806652806653, + "grad_norm": 0.0004844397772103548, + "learning_rate": 0.27129019670783106, + "loss": 0.2792, + "num_input_tokens_seen": 6110048, + "step": 8010 + }, + { + "epoch": 16.66320166320166, + "grad_norm": 0.0007052401197142899, + "learning_rate": 0.27125553027045746, + "loss": 0.2722, + "num_input_tokens_seen": 6113888, + "step": 8015 + }, + { + "epoch": 16.673596673596673, + "grad_norm": 0.00034177390625700355, + "learning_rate": 0.2712208451339572, + "loss": 0.2939, + "num_input_tokens_seen": 6117632, + "step": 8020 + }, + { + "epoch": 16.683991683991685, + "grad_norm": 0.00014036981156095862, + "learning_rate": 0.27118614130367935, + "loss": 0.2338, + "num_input_tokens_seen": 6121376, + "step": 8025 + }, + { + "epoch": 16.694386694386694, + "grad_norm": 0.00019832489488180727, + "learning_rate": 0.2711514187849756, + "loss": 0.2479, + "num_input_tokens_seen": 6124992, + "step": 8030 + }, + { + "epoch": 16.704781704781706, + "grad_norm": 0.0009102554759010673, + "learning_rate": 0.27111667758320057, + "loss": 0.2694, + "num_input_tokens_seen": 6128736, + "step": 8035 + }, + { + "epoch": 16.715176715176714, + "grad_norm": 0.00021431792993098497, + "learning_rate": 0.27108191770371176, + "loss": 0.2581, + "num_input_tokens_seen": 6132608, + "step": 8040 + }, + { + "epoch": 16.725571725571726, + "grad_norm": 0.000998622621409595, + "learning_rate": 0.2710471391518697, + "loss": 0.2702, + "num_input_tokens_seen": 6136448, + "step": 8045 + }, + { + "epoch": 16.735966735966738, + "grad_norm": 0.00026419892674311996, + "learning_rate": 0.2710123419330375, + "loss": 0.264, + "num_input_tokens_seen": 6140448, + "step": 8050 + }, + { + "epoch": 16.746361746361746, + "grad_norm": 0.0005204705521464348, + "learning_rate": 0.2709775260525816, + "loss": 0.2498, + "num_input_tokens_seen": 6144128, + "step": 8055 + }, + { + "epoch": 16.756756756756758, + "grad_norm": 0.0010125788394361734, + "learning_rate": 0.27094269151587075, + "loss": 0.2903, + "num_input_tokens_seen": 6148000, + "step": 8060 + }, + { + "epoch": 16.767151767151766, + "grad_norm": 0.0008944572764448822, + "learning_rate": 0.27090783832827703, + "loss": 0.2695, + "num_input_tokens_seen": 6151840, + "step": 8065 + }, + { + "epoch": 16.777546777546778, + "grad_norm": 8.207056816900149e-05, + "learning_rate": 0.2708729664951753, + "loss": 0.2824, + "num_input_tokens_seen": 6155392, + "step": 8070 + }, + { + "epoch": 16.787941787941786, + "grad_norm": 0.0005768160335719585, + "learning_rate": 0.27083807602194304, + "loss": 0.2624, + "num_input_tokens_seen": 6159328, + "step": 8075 + }, + { + "epoch": 16.7983367983368, + "grad_norm": 0.0002121137804351747, + "learning_rate": 0.270803166913961, + "loss": 0.2763, + "num_input_tokens_seen": 6163136, + "step": 8080 + }, + { + "epoch": 16.80873180873181, + "grad_norm": 0.0006614199955947697, + "learning_rate": 0.27076823917661247, + "loss": 0.263, + "num_input_tokens_seen": 6167008, + "step": 8085 + }, + { + "epoch": 16.81912681912682, + "grad_norm": 0.0002814335748553276, + "learning_rate": 0.2707332928152838, + "loss": 0.268, + "num_input_tokens_seen": 6170816, + "step": 8090 + }, + { + "epoch": 16.82952182952183, + "grad_norm": 0.00060169497737661, + "learning_rate": 0.2706983278353641, + "loss": 0.27, + "num_input_tokens_seen": 6174656, + "step": 8095 + }, + { + "epoch": 16.83991683991684, + "grad_norm": 5.6001288612606004e-05, + "learning_rate": 0.27066334424224553, + "loss": 0.2714, + "num_input_tokens_seen": 6178432, + "step": 8100 + }, + { + "epoch": 16.85031185031185, + "grad_norm": 0.000641416700091213, + "learning_rate": 0.27062834204132297, + "loss": 0.2545, + "num_input_tokens_seen": 6182080, + "step": 8105 + }, + { + "epoch": 16.86070686070686, + "grad_norm": 0.0007159411325119436, + "learning_rate": 0.27059332123799407, + "loss": 0.269, + "num_input_tokens_seen": 6185984, + "step": 8110 + }, + { + "epoch": 16.87110187110187, + "grad_norm": 0.0010314378887414932, + "learning_rate": 0.27055828183765956, + "loss": 0.253, + "num_input_tokens_seen": 6189760, + "step": 8115 + }, + { + "epoch": 16.881496881496883, + "grad_norm": 6.857199332443997e-05, + "learning_rate": 0.270523223845723, + "loss": 0.2312, + "num_input_tokens_seen": 6193440, + "step": 8120 + }, + { + "epoch": 16.89189189189189, + "grad_norm": 0.00027284023235552013, + "learning_rate": 0.2704881472675907, + "loss": 0.2688, + "num_input_tokens_seen": 6197120, + "step": 8125 + }, + { + "epoch": 16.902286902286903, + "grad_norm": 0.001017877133563161, + "learning_rate": 0.270453052108672, + "loss": 0.2583, + "num_input_tokens_seen": 6201120, + "step": 8130 + }, + { + "epoch": 16.91268191268191, + "grad_norm": 0.0009476360864937305, + "learning_rate": 0.2704179383743789, + "loss": 0.2787, + "num_input_tokens_seen": 6204960, + "step": 8135 + }, + { + "epoch": 16.923076923076923, + "grad_norm": 0.00039831383037380874, + "learning_rate": 0.27038280607012644, + "loss": 0.2585, + "num_input_tokens_seen": 6208768, + "step": 8140 + }, + { + "epoch": 16.933471933471935, + "grad_norm": 0.0005661526811309159, + "learning_rate": 0.27034765520133247, + "loss": 0.2789, + "num_input_tokens_seen": 6212544, + "step": 8145 + }, + { + "epoch": 16.943866943866944, + "grad_norm": 0.00018347745935898274, + "learning_rate": 0.2703124857734177, + "loss": 0.2604, + "num_input_tokens_seen": 6216320, + "step": 8150 + }, + { + "epoch": 16.954261954261955, + "grad_norm": 0.0001623460848350078, + "learning_rate": 0.27027729779180565, + "loss": 0.2784, + "num_input_tokens_seen": 6220064, + "step": 8155 + }, + { + "epoch": 16.964656964656964, + "grad_norm": 0.00019272784993518144, + "learning_rate": 0.27024209126192283, + "loss": 0.2705, + "num_input_tokens_seen": 6223808, + "step": 8160 + }, + { + "epoch": 16.975051975051976, + "grad_norm": 0.001076703192666173, + "learning_rate": 0.2702068661891984, + "loss": 0.2757, + "num_input_tokens_seen": 6227712, + "step": 8165 + }, + { + "epoch": 16.985446985446984, + "grad_norm": 0.0003786480228882283, + "learning_rate": 0.2701716225790647, + "loss": 0.2784, + "num_input_tokens_seen": 6231488, + "step": 8170 + }, + { + "epoch": 16.995841995841996, + "grad_norm": 0.0003053348627872765, + "learning_rate": 0.27013636043695655, + "loss": 0.2735, + "num_input_tokens_seen": 6235296, + "step": 8175 + }, + { + "epoch": 17.006237006237008, + "grad_norm": 8.648289076518267e-05, + "learning_rate": 0.27010107976831194, + "loss": 0.2439, + "num_input_tokens_seen": 6238992, + "step": 8180 + }, + { + "epoch": 17.016632016632016, + "grad_norm": 0.00018706229457166046, + "learning_rate": 0.2700657805785715, + "loss": 0.2835, + "num_input_tokens_seen": 6242928, + "step": 8185 + }, + { + "epoch": 17.027027027027028, + "grad_norm": 0.00018215205636806786, + "learning_rate": 0.2700304628731789, + "loss": 0.2769, + "num_input_tokens_seen": 6246672, + "step": 8190 + }, + { + "epoch": 17.037422037422036, + "grad_norm": 0.00028125691460445523, + "learning_rate": 0.26999512665758046, + "loss": 0.284, + "num_input_tokens_seen": 6250544, + "step": 8195 + }, + { + "epoch": 17.04781704781705, + "grad_norm": 0.000305652356473729, + "learning_rate": 0.2699597719372256, + "loss": 0.2718, + "num_input_tokens_seen": 6254288, + "step": 8200 + }, + { + "epoch": 17.04781704781705, + "eval_loss": 0.2478809356689453, + "eval_runtime": 13.4489, + "eval_samples_per_second": 63.649, + "eval_steps_per_second": 15.912, + "num_input_tokens_seen": 6254288, + "step": 8200 + }, + { + "epoch": 17.058212058212057, + "grad_norm": 0.00012966342910658568, + "learning_rate": 0.26992439871756635, + "loss": 0.2274, + "num_input_tokens_seen": 6258128, + "step": 8205 + }, + { + "epoch": 17.06860706860707, + "grad_norm": 6.038767969585024e-05, + "learning_rate": 0.2698890070040578, + "loss": 0.2561, + "num_input_tokens_seen": 6262000, + "step": 8210 + }, + { + "epoch": 17.07900207900208, + "grad_norm": 0.0008076019585132599, + "learning_rate": 0.2698535968021577, + "loss": 0.3054, + "num_input_tokens_seen": 6265840, + "step": 8215 + }, + { + "epoch": 17.08939708939709, + "grad_norm": 0.0015476435655727983, + "learning_rate": 0.26981816811732684, + "loss": 0.2983, + "num_input_tokens_seen": 6269744, + "step": 8220 + }, + { + "epoch": 17.0997920997921, + "grad_norm": 0.0005021239048801363, + "learning_rate": 0.26978272095502875, + "loss": 0.2605, + "num_input_tokens_seen": 6273552, + "step": 8225 + }, + { + "epoch": 17.11018711018711, + "grad_norm": 0.0007540836813859642, + "learning_rate": 0.26974725532072974, + "loss": 0.3453, + "num_input_tokens_seen": 6277296, + "step": 8230 + }, + { + "epoch": 17.12058212058212, + "grad_norm": 0.0003067753277719021, + "learning_rate": 0.26971177121989914, + "loss": 0.2624, + "num_input_tokens_seen": 6281168, + "step": 8235 + }, + { + "epoch": 17.13097713097713, + "grad_norm": 0.00023701327154412866, + "learning_rate": 0.2696762686580091, + "loss": 0.2911, + "num_input_tokens_seen": 6285072, + "step": 8240 + }, + { + "epoch": 17.14137214137214, + "grad_norm": 0.0005202029133215547, + "learning_rate": 0.26964074764053436, + "loss": 0.2678, + "num_input_tokens_seen": 6288944, + "step": 8245 + }, + { + "epoch": 17.151767151767153, + "grad_norm": 0.00014841160736978054, + "learning_rate": 0.2696052081729529, + "loss": 0.2911, + "num_input_tokens_seen": 6292816, + "step": 8250 + }, + { + "epoch": 17.16216216216216, + "grad_norm": 0.0014726329827681184, + "learning_rate": 0.2695696502607453, + "loss": 0.2773, + "num_input_tokens_seen": 6296560, + "step": 8255 + }, + { + "epoch": 17.172557172557173, + "grad_norm": 0.00014644402835983783, + "learning_rate": 0.26953407390939504, + "loss": 0.2748, + "num_input_tokens_seen": 6300496, + "step": 8260 + }, + { + "epoch": 17.18295218295218, + "grad_norm": 0.00020524821593426168, + "learning_rate": 0.26949847912438835, + "loss": 0.278, + "num_input_tokens_seen": 6304368, + "step": 8265 + }, + { + "epoch": 17.193347193347194, + "grad_norm": 0.0004535742336884141, + "learning_rate": 0.26946286591121454, + "loss": 0.2807, + "num_input_tokens_seen": 6308176, + "step": 8270 + }, + { + "epoch": 17.203742203742205, + "grad_norm": 0.00041877449257299304, + "learning_rate": 0.2694272342753655, + "loss": 0.2715, + "num_input_tokens_seen": 6311984, + "step": 8275 + }, + { + "epoch": 17.214137214137214, + "grad_norm": 0.00035122406552545726, + "learning_rate": 0.26939158422233617, + "loss": 0.2698, + "num_input_tokens_seen": 6315888, + "step": 8280 + }, + { + "epoch": 17.224532224532226, + "grad_norm": 0.06246982514858246, + "learning_rate": 0.26935591575762413, + "loss": 1.5278, + "num_input_tokens_seen": 6319536, + "step": 8285 + }, + { + "epoch": 17.234927234927234, + "grad_norm": 0.001877827220596373, + "learning_rate": 0.26932022888672996, + "loss": 0.268, + "num_input_tokens_seen": 6323440, + "step": 8290 + }, + { + "epoch": 17.245322245322246, + "grad_norm": 0.000290247262455523, + "learning_rate": 0.26928452361515703, + "loss": 0.2952, + "num_input_tokens_seen": 6327344, + "step": 8295 + }, + { + "epoch": 17.255717255717254, + "grad_norm": 0.0005606574704870582, + "learning_rate": 0.26924879994841155, + "loss": 0.2703, + "num_input_tokens_seen": 6331120, + "step": 8300 + }, + { + "epoch": 17.266112266112266, + "grad_norm": 0.0005547350156120956, + "learning_rate": 0.2692130578920025, + "loss": 0.2909, + "num_input_tokens_seen": 6334832, + "step": 8305 + }, + { + "epoch": 17.276507276507278, + "grad_norm": 0.002565674716606736, + "learning_rate": 0.26917729745144187, + "loss": 0.2818, + "num_input_tokens_seen": 6338544, + "step": 8310 + }, + { + "epoch": 17.286902286902286, + "grad_norm": 0.0018962292233482003, + "learning_rate": 0.2691415186322443, + "loss": 0.2926, + "num_input_tokens_seen": 6342512, + "step": 8315 + }, + { + "epoch": 17.2972972972973, + "grad_norm": 0.0005084316362626851, + "learning_rate": 0.2691057214399273, + "loss": 0.2885, + "num_input_tokens_seen": 6346448, + "step": 8320 + }, + { + "epoch": 17.307692307692307, + "grad_norm": 0.0011925476137548685, + "learning_rate": 0.2690699058800113, + "loss": 0.2757, + "num_input_tokens_seen": 6350352, + "step": 8325 + }, + { + "epoch": 17.31808731808732, + "grad_norm": 0.0004931745934300125, + "learning_rate": 0.2690340719580194, + "loss": 0.2636, + "num_input_tokens_seen": 6354032, + "step": 8330 + }, + { + "epoch": 17.328482328482327, + "grad_norm": 0.00015508159413002431, + "learning_rate": 0.2689982196794778, + "loss": 0.2568, + "num_input_tokens_seen": 6357872, + "step": 8335 + }, + { + "epoch": 17.33887733887734, + "grad_norm": 0.00019498851906973869, + "learning_rate": 0.2689623490499153, + "loss": 0.2772, + "num_input_tokens_seen": 6361648, + "step": 8340 + }, + { + "epoch": 17.34927234927235, + "grad_norm": 0.00045442607370205224, + "learning_rate": 0.2689264600748636, + "loss": 0.2547, + "num_input_tokens_seen": 6365616, + "step": 8345 + }, + { + "epoch": 17.35966735966736, + "grad_norm": 0.0021386456210166216, + "learning_rate": 0.26889055275985724, + "loss": 0.2911, + "num_input_tokens_seen": 6369360, + "step": 8350 + }, + { + "epoch": 17.37006237006237, + "grad_norm": 0.0002435635105939582, + "learning_rate": 0.2688546271104335, + "loss": 0.2772, + "num_input_tokens_seen": 6373264, + "step": 8355 + }, + { + "epoch": 17.38045738045738, + "grad_norm": 0.0006914423429407179, + "learning_rate": 0.26881868313213275, + "loss": 0.2872, + "num_input_tokens_seen": 6376976, + "step": 8360 + }, + { + "epoch": 17.39085239085239, + "grad_norm": 0.0010851302649825811, + "learning_rate": 0.2687827208304978, + "loss": 0.2756, + "num_input_tokens_seen": 6380816, + "step": 8365 + }, + { + "epoch": 17.401247401247403, + "grad_norm": 0.0008882470428943634, + "learning_rate": 0.26874674021107464, + "loss": 0.2485, + "num_input_tokens_seen": 6384624, + "step": 8370 + }, + { + "epoch": 17.41164241164241, + "grad_norm": 0.0010572298197075725, + "learning_rate": 0.2687107412794118, + "loss": 0.3055, + "num_input_tokens_seen": 6388528, + "step": 8375 + }, + { + "epoch": 17.422037422037423, + "grad_norm": 0.0014390433207154274, + "learning_rate": 0.26867472404106096, + "loss": 0.2798, + "num_input_tokens_seen": 6392400, + "step": 8380 + }, + { + "epoch": 17.43243243243243, + "grad_norm": 0.000769271980971098, + "learning_rate": 0.26863868850157624, + "loss": 0.2627, + "num_input_tokens_seen": 6396144, + "step": 8385 + }, + { + "epoch": 17.442827442827443, + "grad_norm": 0.001266435720026493, + "learning_rate": 0.26860263466651485, + "loss": 0.2194, + "num_input_tokens_seen": 6399984, + "step": 8390 + }, + { + "epoch": 17.453222453222452, + "grad_norm": 0.00041078811045736074, + "learning_rate": 0.26856656254143674, + "loss": 0.2418, + "num_input_tokens_seen": 6403696, + "step": 8395 + }, + { + "epoch": 17.463617463617464, + "grad_norm": 0.00017173706146422774, + "learning_rate": 0.2685304721319047, + "loss": 0.2789, + "num_input_tokens_seen": 6407504, + "step": 8400 + }, + { + "epoch": 17.463617463617464, + "eval_loss": 0.2646408677101135, + "eval_runtime": 13.421, + "eval_samples_per_second": 63.781, + "eval_steps_per_second": 15.945, + "num_input_tokens_seen": 6407504, + "step": 8400 + }, + { + "epoch": 17.474012474012476, + "grad_norm": 0.0004560579836834222, + "learning_rate": 0.2684943634434843, + "loss": 0.2693, + "num_input_tokens_seen": 6411440, + "step": 8405 + }, + { + "epoch": 17.484407484407484, + "grad_norm": 0.0002050929906545207, + "learning_rate": 0.268458236481744, + "loss": 0.2705, + "num_input_tokens_seen": 6415152, + "step": 8410 + }, + { + "epoch": 17.494802494802496, + "grad_norm": 0.0006892663077451289, + "learning_rate": 0.2684220912522549, + "loss": 0.2904, + "num_input_tokens_seen": 6418864, + "step": 8415 + }, + { + "epoch": 17.505197505197504, + "grad_norm": 0.0005261699552647769, + "learning_rate": 0.2683859277605913, + "loss": 0.2858, + "num_input_tokens_seen": 6422768, + "step": 8420 + }, + { + "epoch": 17.515592515592516, + "grad_norm": 0.0010613807244226336, + "learning_rate": 0.2683497460123298, + "loss": 0.3047, + "num_input_tokens_seen": 6426608, + "step": 8425 + }, + { + "epoch": 17.525987525987524, + "grad_norm": 0.0001666166353970766, + "learning_rate": 0.26831354601305013, + "loss": 0.2852, + "num_input_tokens_seen": 6430416, + "step": 8430 + }, + { + "epoch": 17.536382536382536, + "grad_norm": 0.00035167779424227774, + "learning_rate": 0.26827732776833496, + "loss": 0.2776, + "num_input_tokens_seen": 6434224, + "step": 8435 + }, + { + "epoch": 17.546777546777548, + "grad_norm": 0.00012222674558870494, + "learning_rate": 0.26824109128376944, + "loss": 0.2765, + "num_input_tokens_seen": 6437840, + "step": 8440 + }, + { + "epoch": 17.557172557172557, + "grad_norm": 0.001379015389829874, + "learning_rate": 0.2682048365649417, + "loss": 0.2897, + "num_input_tokens_seen": 6441680, + "step": 8445 + }, + { + "epoch": 17.56756756756757, + "grad_norm": 0.001555951894260943, + "learning_rate": 0.2681685636174428, + "loss": 0.2629, + "num_input_tokens_seen": 6445584, + "step": 8450 + }, + { + "epoch": 17.577962577962577, + "grad_norm": 0.0002624621265567839, + "learning_rate": 0.2681322724468663, + "loss": 0.2609, + "num_input_tokens_seen": 6449328, + "step": 8455 + }, + { + "epoch": 17.58835758835759, + "grad_norm": 0.00029405258828774095, + "learning_rate": 0.2680959630588089, + "loss": 0.2491, + "num_input_tokens_seen": 6453008, + "step": 8460 + }, + { + "epoch": 17.598752598752597, + "grad_norm": 0.0002858947846107185, + "learning_rate": 0.26805963545886985, + "loss": 0.2783, + "num_input_tokens_seen": 6456784, + "step": 8465 + }, + { + "epoch": 17.60914760914761, + "grad_norm": 0.000558042956981808, + "learning_rate": 0.26802328965265143, + "loss": 0.2625, + "num_input_tokens_seen": 6460560, + "step": 8470 + }, + { + "epoch": 17.61954261954262, + "grad_norm": 0.00012233134475536644, + "learning_rate": 0.26798692564575854, + "loss": 0.2762, + "num_input_tokens_seen": 6464432, + "step": 8475 + }, + { + "epoch": 17.62993762993763, + "grad_norm": 0.0005754560115747154, + "learning_rate": 0.26795054344379904, + "loss": 0.2805, + "num_input_tokens_seen": 6468208, + "step": 8480 + }, + { + "epoch": 17.64033264033264, + "grad_norm": 0.0006691869348287582, + "learning_rate": 0.2679141430523835, + "loss": 0.2718, + "num_input_tokens_seen": 6471952, + "step": 8485 + }, + { + "epoch": 17.65072765072765, + "grad_norm": 0.0010407933732494712, + "learning_rate": 0.2678777244771252, + "loss": 0.276, + "num_input_tokens_seen": 6475984, + "step": 8490 + }, + { + "epoch": 17.66112266112266, + "grad_norm": 0.0005314350128173828, + "learning_rate": 0.2678412877236405, + "loss": 0.2679, + "num_input_tokens_seen": 6479760, + "step": 8495 + }, + { + "epoch": 17.671517671517673, + "grad_norm": 0.002024328801780939, + "learning_rate": 0.2678048327975484, + "loss": 0.2955, + "num_input_tokens_seen": 6483600, + "step": 8500 + }, + { + "epoch": 17.68191268191268, + "grad_norm": 0.0006551137194037437, + "learning_rate": 0.2677683597044706, + "loss": 0.2871, + "num_input_tokens_seen": 6487408, + "step": 8505 + }, + { + "epoch": 17.692307692307693, + "grad_norm": 0.000241446090512909, + "learning_rate": 0.2677318684500318, + "loss": 0.2678, + "num_input_tokens_seen": 6490960, + "step": 8510 + }, + { + "epoch": 17.7027027027027, + "grad_norm": 0.00017587537877261639, + "learning_rate": 0.2676953590398593, + "loss": 0.2662, + "num_input_tokens_seen": 6494864, + "step": 8515 + }, + { + "epoch": 17.713097713097714, + "grad_norm": 0.0007013955037109554, + "learning_rate": 0.2676588314795834, + "loss": 0.2684, + "num_input_tokens_seen": 6498896, + "step": 8520 + }, + { + "epoch": 17.723492723492722, + "grad_norm": 0.00043093739077448845, + "learning_rate": 0.26762228577483715, + "loss": 0.2562, + "num_input_tokens_seen": 6502640, + "step": 8525 + }, + { + "epoch": 17.733887733887734, + "grad_norm": 0.00014598824782297015, + "learning_rate": 0.2675857219312563, + "loss": 0.2764, + "num_input_tokens_seen": 6506448, + "step": 8530 + }, + { + "epoch": 17.744282744282746, + "grad_norm": 0.0002772712614387274, + "learning_rate": 0.2675491399544794, + "loss": 0.2812, + "num_input_tokens_seen": 6510224, + "step": 8535 + }, + { + "epoch": 17.754677754677754, + "grad_norm": 0.0004319855070207268, + "learning_rate": 0.2675125398501479, + "loss": 0.2915, + "num_input_tokens_seen": 6513936, + "step": 8540 + }, + { + "epoch": 17.765072765072766, + "grad_norm": 0.0006322558037936687, + "learning_rate": 0.26747592162390604, + "loss": 0.2231, + "num_input_tokens_seen": 6517712, + "step": 8545 + }, + { + "epoch": 17.775467775467774, + "grad_norm": 0.0010780047159641981, + "learning_rate": 0.26743928528140076, + "loss": 0.2958, + "num_input_tokens_seen": 6521392, + "step": 8550 + }, + { + "epoch": 17.785862785862786, + "grad_norm": 0.0008319044718518853, + "learning_rate": 0.26740263082828186, + "loss": 0.3161, + "num_input_tokens_seen": 6525200, + "step": 8555 + }, + { + "epoch": 17.796257796257795, + "grad_norm": 0.0001987960422411561, + "learning_rate": 0.2673659582702019, + "loss": 0.2785, + "num_input_tokens_seen": 6529136, + "step": 8560 + }, + { + "epoch": 17.806652806652806, + "grad_norm": 0.0006443793536163867, + "learning_rate": 0.2673292676128163, + "loss": 0.2685, + "num_input_tokens_seen": 6533072, + "step": 8565 + }, + { + "epoch": 17.81704781704782, + "grad_norm": 0.0005758479237556458, + "learning_rate": 0.2672925588617831, + "loss": 0.2659, + "num_input_tokens_seen": 6536848, + "step": 8570 + }, + { + "epoch": 17.827442827442827, + "grad_norm": 0.0006438523414544761, + "learning_rate": 0.2672558320227634, + "loss": 0.2873, + "num_input_tokens_seen": 6540688, + "step": 8575 + }, + { + "epoch": 17.83783783783784, + "grad_norm": 0.0018779652891680598, + "learning_rate": 0.2672190871014209, + "loss": 0.2756, + "num_input_tokens_seen": 6544464, + "step": 8580 + }, + { + "epoch": 17.848232848232847, + "grad_norm": 0.0008292693528346717, + "learning_rate": 0.267182324103422, + "loss": 0.2708, + "num_input_tokens_seen": 6548400, + "step": 8585 + }, + { + "epoch": 17.85862785862786, + "grad_norm": 0.00012196957686683163, + "learning_rate": 0.2671455430344362, + "loss": 0.297, + "num_input_tokens_seen": 6552208, + "step": 8590 + }, + { + "epoch": 17.86902286902287, + "grad_norm": 0.0006085454951971769, + "learning_rate": 0.2671087439001355, + "loss": 0.2732, + "num_input_tokens_seen": 6555984, + "step": 8595 + }, + { + "epoch": 17.87941787941788, + "grad_norm": 0.00011699511378537863, + "learning_rate": 0.2670719267061948, + "loss": 0.2784, + "num_input_tokens_seen": 6559760, + "step": 8600 + }, + { + "epoch": 17.87941787941788, + "eval_loss": 0.25499433279037476, + "eval_runtime": 13.4542, + "eval_samples_per_second": 63.623, + "eval_steps_per_second": 15.906, + "num_input_tokens_seen": 6559760, + "step": 8600 + }, + { + "epoch": 17.88981288981289, + "grad_norm": 0.0009677985799498856, + "learning_rate": 0.2670350914582918, + "loss": 0.2762, + "num_input_tokens_seen": 6563472, + "step": 8605 + }, + { + "epoch": 17.9002079002079, + "grad_norm": 0.00014068270684219897, + "learning_rate": 0.26699823816210694, + "loss": 0.2666, + "num_input_tokens_seen": 6567216, + "step": 8610 + }, + { + "epoch": 17.91060291060291, + "grad_norm": 0.0009792582131922245, + "learning_rate": 0.26696136682332344, + "loss": 0.2785, + "num_input_tokens_seen": 6571024, + "step": 8615 + }, + { + "epoch": 17.92099792099792, + "grad_norm": 0.0001228635519510135, + "learning_rate": 0.2669244774476274, + "loss": 0.2752, + "num_input_tokens_seen": 6574832, + "step": 8620 + }, + { + "epoch": 17.93139293139293, + "grad_norm": 0.00035585102159529924, + "learning_rate": 0.2668875700407075, + "loss": 0.262, + "num_input_tokens_seen": 6578576, + "step": 8625 + }, + { + "epoch": 17.941787941787943, + "grad_norm": 0.0007766150520183146, + "learning_rate": 0.26685064460825547, + "loss": 0.2501, + "num_input_tokens_seen": 6582512, + "step": 8630 + }, + { + "epoch": 17.95218295218295, + "grad_norm": 0.0016311418730765581, + "learning_rate": 0.26681370115596553, + "loss": 0.2986, + "num_input_tokens_seen": 6586224, + "step": 8635 + }, + { + "epoch": 17.962577962577964, + "grad_norm": 0.0008403054671362042, + "learning_rate": 0.26677673968953497, + "loss": 0.2684, + "num_input_tokens_seen": 6590160, + "step": 8640 + }, + { + "epoch": 17.972972972972972, + "grad_norm": 0.00011512546916492283, + "learning_rate": 0.2667397602146636, + "loss": 0.2552, + "num_input_tokens_seen": 6593936, + "step": 8645 + }, + { + "epoch": 17.983367983367984, + "grad_norm": 5.480998879647814e-05, + "learning_rate": 0.2667027627370542, + "loss": 0.2682, + "num_input_tokens_seen": 6597776, + "step": 8650 + }, + { + "epoch": 17.993762993762992, + "grad_norm": 0.00033503398299217224, + "learning_rate": 0.26666574726241216, + "loss": 0.2073, + "num_input_tokens_seen": 6601584, + "step": 8655 + }, + { + "epoch": 18.004158004158004, + "grad_norm": 0.0008311098208650947, + "learning_rate": 0.2666287137964458, + "loss": 0.3243, + "num_input_tokens_seen": 6605376, + "step": 8660 + }, + { + "epoch": 18.014553014553016, + "grad_norm": 0.0022813421674072742, + "learning_rate": 0.26659166234486614, + "loss": 0.2788, + "num_input_tokens_seen": 6609216, + "step": 8665 + }, + { + "epoch": 18.024948024948024, + "grad_norm": 0.00010725743777584285, + "learning_rate": 0.2665545929133869, + "loss": 0.2594, + "num_input_tokens_seen": 6613024, + "step": 8670 + }, + { + "epoch": 18.035343035343036, + "grad_norm": 0.0017440845258533955, + "learning_rate": 0.2665175055077248, + "loss": 0.2974, + "num_input_tokens_seen": 6616672, + "step": 8675 + }, + { + "epoch": 18.045738045738045, + "grad_norm": 0.00010363061301177368, + "learning_rate": 0.2664804001335991, + "loss": 0.2815, + "num_input_tokens_seen": 6620576, + "step": 8680 + }, + { + "epoch": 18.056133056133056, + "grad_norm": 0.0005967674660496414, + "learning_rate": 0.26644327679673185, + "loss": 0.2956, + "num_input_tokens_seen": 6624416, + "step": 8685 + }, + { + "epoch": 18.066528066528065, + "grad_norm": 0.0006394773372448981, + "learning_rate": 0.26640613550284803, + "loss": 0.2895, + "num_input_tokens_seen": 6628352, + "step": 8690 + }, + { + "epoch": 18.076923076923077, + "grad_norm": 0.0001469774724682793, + "learning_rate": 0.26636897625767525, + "loss": 0.2603, + "num_input_tokens_seen": 6632192, + "step": 8695 + }, + { + "epoch": 18.08731808731809, + "grad_norm": 0.00020608387421816587, + "learning_rate": 0.266331799066944, + "loss": 0.2621, + "num_input_tokens_seen": 6635968, + "step": 8700 + }, + { + "epoch": 18.097713097713097, + "grad_norm": 0.0005856282077729702, + "learning_rate": 0.2662946039363874, + "loss": 0.2861, + "num_input_tokens_seen": 6639616, + "step": 8705 + }, + { + "epoch": 18.10810810810811, + "grad_norm": 0.000237450614804402, + "learning_rate": 0.2662573908717414, + "loss": 0.2684, + "num_input_tokens_seen": 6643392, + "step": 8710 + }, + { + "epoch": 18.118503118503117, + "grad_norm": 0.00040499100578017533, + "learning_rate": 0.2662201598787447, + "loss": 0.2808, + "num_input_tokens_seen": 6647264, + "step": 8715 + }, + { + "epoch": 18.12889812889813, + "grad_norm": 0.0005211139796301723, + "learning_rate": 0.2661829109631389, + "loss": 0.2748, + "num_input_tokens_seen": 6651008, + "step": 8720 + }, + { + "epoch": 18.13929313929314, + "grad_norm": 0.0001843898353399709, + "learning_rate": 0.26614564413066816, + "loss": 0.2776, + "num_input_tokens_seen": 6654944, + "step": 8725 + }, + { + "epoch": 18.14968814968815, + "grad_norm": 0.0003061479073949158, + "learning_rate": 0.2661083593870795, + "loss": 0.2611, + "num_input_tokens_seen": 6658816, + "step": 8730 + }, + { + "epoch": 18.16008316008316, + "grad_norm": 0.0004203488351777196, + "learning_rate": 0.26607105673812276, + "loss": 0.2641, + "num_input_tokens_seen": 6662528, + "step": 8735 + }, + { + "epoch": 18.17047817047817, + "grad_norm": 0.0005340935895219445, + "learning_rate": 0.2660337361895504, + "loss": 0.2724, + "num_input_tokens_seen": 6666336, + "step": 8740 + }, + { + "epoch": 18.18087318087318, + "grad_norm": 0.0003870767541229725, + "learning_rate": 0.26599639774711775, + "loss": 0.263, + "num_input_tokens_seen": 6670272, + "step": 8745 + }, + { + "epoch": 18.19126819126819, + "grad_norm": 0.0005318223848007619, + "learning_rate": 0.2659590414165829, + "loss": 0.2776, + "num_input_tokens_seen": 6674112, + "step": 8750 + }, + { + "epoch": 18.2016632016632, + "grad_norm": 0.000512822880409658, + "learning_rate": 0.2659216672037066, + "loss": 0.2741, + "num_input_tokens_seen": 6677952, + "step": 8755 + }, + { + "epoch": 18.212058212058214, + "grad_norm": 0.0001758085418259725, + "learning_rate": 0.26588427511425244, + "loss": 0.2342, + "num_input_tokens_seen": 6681728, + "step": 8760 + }, + { + "epoch": 18.222453222453222, + "grad_norm": 0.0004862227360717952, + "learning_rate": 0.26584686515398676, + "loss": 0.2872, + "num_input_tokens_seen": 6685504, + "step": 8765 + }, + { + "epoch": 18.232848232848234, + "grad_norm": 0.001122916815802455, + "learning_rate": 0.2658094373286787, + "loss": 0.282, + "num_input_tokens_seen": 6689280, + "step": 8770 + }, + { + "epoch": 18.243243243243242, + "grad_norm": 0.001084034563973546, + "learning_rate": 0.2657719916441, + "loss": 0.2795, + "num_input_tokens_seen": 6693088, + "step": 8775 + }, + { + "epoch": 18.253638253638254, + "grad_norm": 0.0008276728331111372, + "learning_rate": 0.2657345281060253, + "loss": 0.2644, + "num_input_tokens_seen": 6696736, + "step": 8780 + }, + { + "epoch": 18.264033264033262, + "grad_norm": 0.000359440193278715, + "learning_rate": 0.26569704672023203, + "loss": 0.2474, + "num_input_tokens_seen": 6700512, + "step": 8785 + }, + { + "epoch": 18.274428274428274, + "grad_norm": 0.0006516285357065499, + "learning_rate": 0.26565954749250015, + "loss": 0.3119, + "num_input_tokens_seen": 6704288, + "step": 8790 + }, + { + "epoch": 18.284823284823286, + "grad_norm": 0.00048808593419380486, + "learning_rate": 0.2656220304286126, + "loss": 0.2913, + "num_input_tokens_seen": 6708064, + "step": 8795 + }, + { + "epoch": 18.295218295218294, + "grad_norm": 5.839101504534483e-05, + "learning_rate": 0.265584495534355, + "loss": 0.2703, + "num_input_tokens_seen": 6711968, + "step": 8800 + }, + { + "epoch": 18.295218295218294, + "eval_loss": 0.24927042424678802, + "eval_runtime": 13.4681, + "eval_samples_per_second": 63.558, + "eval_steps_per_second": 15.889, + "num_input_tokens_seen": 6711968, + "step": 8800 + }, + { + "epoch": 18.305613305613306, + "grad_norm": 0.00012995293946005404, + "learning_rate": 0.2655469428155156, + "loss": 0.2737, + "num_input_tokens_seen": 6715776, + "step": 8805 + }, + { + "epoch": 18.316008316008315, + "grad_norm": 0.0002193444815929979, + "learning_rate": 0.2655093722778856, + "loss": 0.2988, + "num_input_tokens_seen": 6719616, + "step": 8810 + }, + { + "epoch": 18.326403326403327, + "grad_norm": 0.0019608838483691216, + "learning_rate": 0.2654717839272588, + "loss": 0.2965, + "num_input_tokens_seen": 6723360, + "step": 8815 + }, + { + "epoch": 18.33679833679834, + "grad_norm": 7.834819552954286e-05, + "learning_rate": 0.2654341777694318, + "loss": 0.2791, + "num_input_tokens_seen": 6727136, + "step": 8820 + }, + { + "epoch": 18.347193347193347, + "grad_norm": 0.0009709492442198098, + "learning_rate": 0.265396553810204, + "loss": 0.2687, + "num_input_tokens_seen": 6730944, + "step": 8825 + }, + { + "epoch": 18.35758835758836, + "grad_norm": 0.000904196291230619, + "learning_rate": 0.26535891205537737, + "loss": 0.3192, + "num_input_tokens_seen": 6734880, + "step": 8830 + }, + { + "epoch": 18.367983367983367, + "grad_norm": 9.113308624364436e-05, + "learning_rate": 0.26532125251075683, + "loss": 0.2812, + "num_input_tokens_seen": 6738688, + "step": 8835 + }, + { + "epoch": 18.37837837837838, + "grad_norm": 0.00037280612741596997, + "learning_rate": 0.26528357518214996, + "loss": 0.2732, + "num_input_tokens_seen": 6742592, + "step": 8840 + }, + { + "epoch": 18.388773388773387, + "grad_norm": 0.0012490348890423775, + "learning_rate": 0.26524588007536704, + "loss": 0.2933, + "num_input_tokens_seen": 6746432, + "step": 8845 + }, + { + "epoch": 18.3991683991684, + "grad_norm": 0.0004500504001043737, + "learning_rate": 0.26520816719622115, + "loss": 0.2797, + "num_input_tokens_seen": 6750240, + "step": 8850 + }, + { + "epoch": 18.40956340956341, + "grad_norm": 0.0007879060576669872, + "learning_rate": 0.2651704365505281, + "loss": 0.2724, + "num_input_tokens_seen": 6754016, + "step": 8855 + }, + { + "epoch": 18.41995841995842, + "grad_norm": 0.0003343670687172562, + "learning_rate": 0.26513268814410634, + "loss": 0.264, + "num_input_tokens_seen": 6757856, + "step": 8860 + }, + { + "epoch": 18.43035343035343, + "grad_norm": 0.00012655530008487403, + "learning_rate": 0.2650949219827773, + "loss": 0.2844, + "num_input_tokens_seen": 6761728, + "step": 8865 + }, + { + "epoch": 18.44074844074844, + "grad_norm": 0.00011438181536504999, + "learning_rate": 0.26505713807236486, + "loss": 0.2695, + "num_input_tokens_seen": 6765600, + "step": 8870 + }, + { + "epoch": 18.45114345114345, + "grad_norm": 0.0012365286238491535, + "learning_rate": 0.26501933641869585, + "loss": 0.2434, + "num_input_tokens_seen": 6769344, + "step": 8875 + }, + { + "epoch": 18.46153846153846, + "grad_norm": 0.00010640593245625496, + "learning_rate": 0.26498151702759976, + "loss": 0.2694, + "num_input_tokens_seen": 6773184, + "step": 8880 + }, + { + "epoch": 18.471933471933472, + "grad_norm": 0.00030520043219439685, + "learning_rate": 0.2649436799049088, + "loss": 0.257, + "num_input_tokens_seen": 6777024, + "step": 8885 + }, + { + "epoch": 18.482328482328484, + "grad_norm": 0.0005573865491896868, + "learning_rate": 0.2649058250564579, + "loss": 0.2832, + "num_input_tokens_seen": 6780832, + "step": 8890 + }, + { + "epoch": 18.492723492723492, + "grad_norm": 0.0004710531502496451, + "learning_rate": 0.26486795248808476, + "loss": 0.2699, + "num_input_tokens_seen": 6784576, + "step": 8895 + }, + { + "epoch": 18.503118503118504, + "grad_norm": 0.00010646675218595192, + "learning_rate": 0.2648300622056298, + "loss": 0.2869, + "num_input_tokens_seen": 6788384, + "step": 8900 + }, + { + "epoch": 18.513513513513512, + "grad_norm": 0.0007506481488235295, + "learning_rate": 0.2647921542149363, + "loss": 0.2755, + "num_input_tokens_seen": 6792128, + "step": 8905 + }, + { + "epoch": 18.523908523908524, + "grad_norm": 0.0007890819106251001, + "learning_rate": 0.26475422852185, + "loss": 0.2538, + "num_input_tokens_seen": 6796128, + "step": 8910 + }, + { + "epoch": 18.534303534303533, + "grad_norm": 0.0008609460201114416, + "learning_rate": 0.2647162851322196, + "loss": 0.2943, + "num_input_tokens_seen": 6799840, + "step": 8915 + }, + { + "epoch": 18.544698544698544, + "grad_norm": 0.0003641176735982299, + "learning_rate": 0.2646783240518964, + "loss": 0.2921, + "num_input_tokens_seen": 6803520, + "step": 8920 + }, + { + "epoch": 18.555093555093556, + "grad_norm": 0.000554008933249861, + "learning_rate": 0.26464034528673447, + "loss": 0.2832, + "num_input_tokens_seen": 6807520, + "step": 8925 + }, + { + "epoch": 18.565488565488565, + "grad_norm": 0.00026493993937037885, + "learning_rate": 0.26460234884259065, + "loss": 0.2728, + "num_input_tokens_seen": 6811328, + "step": 8930 + }, + { + "epoch": 18.575883575883577, + "grad_norm": 0.0005672450643032789, + "learning_rate": 0.2645643347253245, + "loss": 0.2624, + "num_input_tokens_seen": 6815008, + "step": 8935 + }, + { + "epoch": 18.586278586278585, + "grad_norm": 0.0002318425540579483, + "learning_rate": 0.2645263029407982, + "loss": 0.2755, + "num_input_tokens_seen": 6818848, + "step": 8940 + }, + { + "epoch": 18.596673596673597, + "grad_norm": 0.0004432741552591324, + "learning_rate": 0.2644882534948767, + "loss": 0.2792, + "num_input_tokens_seen": 6822688, + "step": 8945 + }, + { + "epoch": 18.60706860706861, + "grad_norm": 0.00048561059520579875, + "learning_rate": 0.2644501863934278, + "loss": 0.2747, + "num_input_tokens_seen": 6826688, + "step": 8950 + }, + { + "epoch": 18.617463617463617, + "grad_norm": 0.000606888032052666, + "learning_rate": 0.26441210164232193, + "loss": 0.3011, + "num_input_tokens_seen": 6830528, + "step": 8955 + }, + { + "epoch": 18.62785862785863, + "grad_norm": 0.0005163914756849408, + "learning_rate": 0.26437399924743216, + "loss": 0.2517, + "num_input_tokens_seen": 6834240, + "step": 8960 + }, + { + "epoch": 18.638253638253637, + "grad_norm": 6.785208097426221e-05, + "learning_rate": 0.26433587921463436, + "loss": 0.2817, + "num_input_tokens_seen": 6838048, + "step": 8965 + }, + { + "epoch": 18.64864864864865, + "grad_norm": 0.0003469529328867793, + "learning_rate": 0.2642977415498072, + "loss": 0.2594, + "num_input_tokens_seen": 6841888, + "step": 8970 + }, + { + "epoch": 18.659043659043657, + "grad_norm": 0.0003062535251956433, + "learning_rate": 0.26425958625883195, + "loss": 0.2825, + "num_input_tokens_seen": 6845600, + "step": 8975 + }, + { + "epoch": 18.66943866943867, + "grad_norm": 0.0006538403104059398, + "learning_rate": 0.2642214133475926, + "loss": 0.2804, + "num_input_tokens_seen": 6849440, + "step": 8980 + }, + { + "epoch": 18.67983367983368, + "grad_norm": 0.0005928523023612797, + "learning_rate": 0.26418322282197587, + "loss": 0.2745, + "num_input_tokens_seen": 6853312, + "step": 8985 + }, + { + "epoch": 18.69022869022869, + "grad_norm": 0.0002444878628011793, + "learning_rate": 0.2641450146878714, + "loss": 0.27, + "num_input_tokens_seen": 6857152, + "step": 8990 + }, + { + "epoch": 18.7006237006237, + "grad_norm": 0.00040578245534561574, + "learning_rate": 0.26410678895117107, + "loss": 0.2709, + "num_input_tokens_seen": 6860992, + "step": 8995 + }, + { + "epoch": 18.71101871101871, + "grad_norm": 0.0005521972198039293, + "learning_rate": 0.26406854561777, + "loss": 0.2681, + "num_input_tokens_seen": 6864736, + "step": 9000 + }, + { + "epoch": 18.71101871101871, + "eval_loss": 0.2508901059627533, + "eval_runtime": 13.4323, + "eval_samples_per_second": 63.727, + "eval_steps_per_second": 15.932, + "num_input_tokens_seen": 6864736, + "step": 9000 + }, + { + "epoch": 18.72141372141372, + "grad_norm": 0.00047054438618943095, + "learning_rate": 0.26403028469356576, + "loss": 0.2499, + "num_input_tokens_seen": 6868448, + "step": 9005 + }, + { + "epoch": 18.731808731808734, + "grad_norm": 0.0003133145801257342, + "learning_rate": 0.2639920061844585, + "loss": 0.2431, + "num_input_tokens_seen": 6872256, + "step": 9010 + }, + { + "epoch": 18.742203742203742, + "grad_norm": 0.00085276341997087, + "learning_rate": 0.2639537100963515, + "loss": 0.2633, + "num_input_tokens_seen": 6876000, + "step": 9015 + }, + { + "epoch": 18.752598752598754, + "grad_norm": 0.0002503904397599399, + "learning_rate": 0.26391539643515033, + "loss": 0.2895, + "num_input_tokens_seen": 6879744, + "step": 9020 + }, + { + "epoch": 18.762993762993762, + "grad_norm": 0.0019540635403245687, + "learning_rate": 0.26387706520676346, + "loss": 0.2955, + "num_input_tokens_seen": 6883616, + "step": 9025 + }, + { + "epoch": 18.773388773388774, + "grad_norm": 0.0002149842621292919, + "learning_rate": 0.26383871641710205, + "loss": 0.2712, + "num_input_tokens_seen": 6887392, + "step": 9030 + }, + { + "epoch": 18.783783783783782, + "grad_norm": 0.0004719561547972262, + "learning_rate": 0.26380035007208, + "loss": 0.2685, + "num_input_tokens_seen": 6891232, + "step": 9035 + }, + { + "epoch": 18.794178794178794, + "grad_norm": 0.00020663491159211844, + "learning_rate": 0.26376196617761394, + "loss": 0.2877, + "num_input_tokens_seen": 6895136, + "step": 9040 + }, + { + "epoch": 18.804573804573806, + "grad_norm": 0.0012713199248537421, + "learning_rate": 0.263723564739623, + "loss": 0.2961, + "num_input_tokens_seen": 6899040, + "step": 9045 + }, + { + "epoch": 18.814968814968815, + "grad_norm": 0.0002601715677883476, + "learning_rate": 0.2636851457640293, + "loss": 0.2586, + "num_input_tokens_seen": 6902784, + "step": 9050 + }, + { + "epoch": 18.825363825363826, + "grad_norm": 0.00018605576769914478, + "learning_rate": 0.26364670925675737, + "loss": 0.2886, + "num_input_tokens_seen": 6906496, + "step": 9055 + }, + { + "epoch": 18.835758835758835, + "grad_norm": 0.0005511005292646587, + "learning_rate": 0.2636082552237347, + "loss": 0.2451, + "num_input_tokens_seen": 6910464, + "step": 9060 + }, + { + "epoch": 18.846153846153847, + "grad_norm": 0.00011461966641945764, + "learning_rate": 0.26356978367089146, + "loss": 0.2569, + "num_input_tokens_seen": 6914368, + "step": 9065 + }, + { + "epoch": 18.856548856548855, + "grad_norm": 0.0006976470467634499, + "learning_rate": 0.26353129460416036, + "loss": 0.2446, + "num_input_tokens_seen": 6918176, + "step": 9070 + }, + { + "epoch": 18.866943866943867, + "grad_norm": 0.00028620948432944715, + "learning_rate": 0.2634927880294769, + "loss": 0.2619, + "num_input_tokens_seen": 6921920, + "step": 9075 + }, + { + "epoch": 18.87733887733888, + "grad_norm": 0.0001477905607316643, + "learning_rate": 0.26345426395277927, + "loss": 0.2735, + "num_input_tokens_seen": 6925728, + "step": 9080 + }, + { + "epoch": 18.887733887733887, + "grad_norm": 0.00023691196111030877, + "learning_rate": 0.2634157223800084, + "loss": 0.2552, + "num_input_tokens_seen": 6929632, + "step": 9085 + }, + { + "epoch": 18.8981288981289, + "grad_norm": 0.0005094673251733184, + "learning_rate": 0.26337716331710787, + "loss": 0.2477, + "num_input_tokens_seen": 6933472, + "step": 9090 + }, + { + "epoch": 18.908523908523907, + "grad_norm": 0.0001158994491561316, + "learning_rate": 0.2633385867700239, + "loss": 0.2712, + "num_input_tokens_seen": 6937248, + "step": 9095 + }, + { + "epoch": 18.91891891891892, + "grad_norm": 0.000534303777385503, + "learning_rate": 0.2632999927447056, + "loss": 0.2652, + "num_input_tokens_seen": 6940992, + "step": 9100 + }, + { + "epoch": 18.929313929313928, + "grad_norm": 0.00021672413276974112, + "learning_rate": 0.2632613812471046, + "loss": 0.273, + "num_input_tokens_seen": 6944768, + "step": 9105 + }, + { + "epoch": 18.93970893970894, + "grad_norm": 0.00037548094405792654, + "learning_rate": 0.2632227522831753, + "loss": 0.2566, + "num_input_tokens_seen": 6948608, + "step": 9110 + }, + { + "epoch": 18.95010395010395, + "grad_norm": 0.00047370526590384543, + "learning_rate": 0.26318410585887475, + "loss": 0.2544, + "num_input_tokens_seen": 6952576, + "step": 9115 + }, + { + "epoch": 18.96049896049896, + "grad_norm": 7.9651981650386e-05, + "learning_rate": 0.2631454419801627, + "loss": 0.2246, + "num_input_tokens_seen": 6956352, + "step": 9120 + }, + { + "epoch": 18.97089397089397, + "grad_norm": 0.0004938665078952909, + "learning_rate": 0.2631067606530016, + "loss": 0.2546, + "num_input_tokens_seen": 6960288, + "step": 9125 + }, + { + "epoch": 18.98128898128898, + "grad_norm": 0.0008581780130043626, + "learning_rate": 0.2630680618833567, + "loss": 0.2666, + "num_input_tokens_seen": 6964032, + "step": 9130 + }, + { + "epoch": 18.991683991683992, + "grad_norm": 0.0009217361221089959, + "learning_rate": 0.26302934567719566, + "loss": 0.2866, + "num_input_tokens_seen": 6967712, + "step": 9135 + }, + { + "epoch": 19.002079002079004, + "grad_norm": 0.00028948872932232916, + "learning_rate": 0.2629906120404892, + "loss": 0.2791, + "num_input_tokens_seen": 6971504, + "step": 9140 + }, + { + "epoch": 19.012474012474012, + "grad_norm": 0.000636327953543514, + "learning_rate": 0.26295186097921036, + "loss": 0.2738, + "num_input_tokens_seen": 6975248, + "step": 9145 + }, + { + "epoch": 19.022869022869024, + "grad_norm": 0.0005886437138542533, + "learning_rate": 0.2629130924993351, + "loss": 0.2748, + "num_input_tokens_seen": 6978960, + "step": 9150 + }, + { + "epoch": 19.033264033264032, + "grad_norm": 0.0005098135443404317, + "learning_rate": 0.2628743066068421, + "loss": 0.2498, + "num_input_tokens_seen": 6982768, + "step": 9155 + }, + { + "epoch": 19.043659043659044, + "grad_norm": 0.000199946211068891, + "learning_rate": 0.26283550330771244, + "loss": 0.2905, + "num_input_tokens_seen": 6986608, + "step": 9160 + }, + { + "epoch": 19.054054054054053, + "grad_norm": 0.001007182989269495, + "learning_rate": 0.2627966826079303, + "loss": 0.2544, + "num_input_tokens_seen": 6990320, + "step": 9165 + }, + { + "epoch": 19.064449064449065, + "grad_norm": 0.00046881879097782075, + "learning_rate": 0.26275784451348216, + "loss": 0.2618, + "num_input_tokens_seen": 6994128, + "step": 9170 + }, + { + "epoch": 19.074844074844076, + "grad_norm": 0.00011254283162998036, + "learning_rate": 0.2627189890303574, + "loss": 0.2752, + "num_input_tokens_seen": 6997840, + "step": 9175 + }, + { + "epoch": 19.085239085239085, + "grad_norm": 0.0006154330330900848, + "learning_rate": 0.262680116164548, + "loss": 0.255, + "num_input_tokens_seen": 7001712, + "step": 9180 + }, + { + "epoch": 19.095634095634097, + "grad_norm": 0.0003279992379248142, + "learning_rate": 0.2626412259220487, + "loss": 0.2617, + "num_input_tokens_seen": 7005680, + "step": 9185 + }, + { + "epoch": 19.106029106029105, + "grad_norm": 0.0008878106600604951, + "learning_rate": 0.2626023183088568, + "loss": 0.2342, + "num_input_tokens_seen": 7009520, + "step": 9190 + }, + { + "epoch": 19.116424116424117, + "grad_norm": 0.0005408648867160082, + "learning_rate": 0.26256339333097234, + "loss": 0.2626, + "num_input_tokens_seen": 7013232, + "step": 9195 + }, + { + "epoch": 19.126819126819125, + "grad_norm": 0.00024181387561839074, + "learning_rate": 0.2625244509943981, + "loss": 0.286, + "num_input_tokens_seen": 7016944, + "step": 9200 + }, + { + "epoch": 19.126819126819125, + "eval_loss": 0.261119544506073, + "eval_runtime": 13.4245, + "eval_samples_per_second": 63.764, + "eval_steps_per_second": 15.941, + "num_input_tokens_seen": 7016944, + "step": 9200 + }, + { + "epoch": 19.137214137214137, + "grad_norm": 0.0011105149751529098, + "learning_rate": 0.2624854913051395, + "loss": 0.2593, + "num_input_tokens_seen": 7020912, + "step": 9205 + }, + { + "epoch": 19.14760914760915, + "grad_norm": 0.00030829329625703394, + "learning_rate": 0.26244651426920446, + "loss": 0.2514, + "num_input_tokens_seen": 7024656, + "step": 9210 + }, + { + "epoch": 19.158004158004157, + "grad_norm": 0.0009486337658017874, + "learning_rate": 0.26240751989260386, + "loss": 0.2905, + "num_input_tokens_seen": 7028496, + "step": 9215 + }, + { + "epoch": 19.16839916839917, + "grad_norm": 0.00012095565762137994, + "learning_rate": 0.2623685081813511, + "loss": 0.2648, + "num_input_tokens_seen": 7032368, + "step": 9220 + }, + { + "epoch": 19.178794178794178, + "grad_norm": 0.00048102933214977384, + "learning_rate": 0.2623294791414623, + "loss": 0.2724, + "num_input_tokens_seen": 7036016, + "step": 9225 + }, + { + "epoch": 19.18918918918919, + "grad_norm": 0.0005834060139022768, + "learning_rate": 0.26229043277895614, + "loss": 0.276, + "num_input_tokens_seen": 7039696, + "step": 9230 + }, + { + "epoch": 19.1995841995842, + "grad_norm": 0.0004141333338338882, + "learning_rate": 0.2622513690998542, + "loss": 0.2648, + "num_input_tokens_seen": 7043472, + "step": 9235 + }, + { + "epoch": 19.20997920997921, + "grad_norm": 0.0008386893314309418, + "learning_rate": 0.26221228811018044, + "loss": 0.2842, + "num_input_tokens_seen": 7047280, + "step": 9240 + }, + { + "epoch": 19.22037422037422, + "grad_norm": 0.0005181029555387795, + "learning_rate": 0.2621731898159617, + "loss": 0.2813, + "num_input_tokens_seen": 7051120, + "step": 9245 + }, + { + "epoch": 19.23076923076923, + "grad_norm": 0.00014839498908258975, + "learning_rate": 0.26213407422322743, + "loss": 0.2604, + "num_input_tokens_seen": 7054928, + "step": 9250 + }, + { + "epoch": 19.241164241164242, + "grad_norm": 0.0004908874398097396, + "learning_rate": 0.2620949413380098, + "loss": 0.2759, + "num_input_tokens_seen": 7058608, + "step": 9255 + }, + { + "epoch": 19.25155925155925, + "grad_norm": 0.00044023379450663924, + "learning_rate": 0.26205579116634353, + "loss": 0.2544, + "num_input_tokens_seen": 7062640, + "step": 9260 + }, + { + "epoch": 19.261954261954262, + "grad_norm": 0.0010700172279030085, + "learning_rate": 0.26201662371426604, + "loss": 0.2648, + "num_input_tokens_seen": 7066256, + "step": 9265 + }, + { + "epoch": 19.272349272349274, + "grad_norm": 0.00025858485605567694, + "learning_rate": 0.2619774389878175, + "loss": 0.2869, + "num_input_tokens_seen": 7069872, + "step": 9270 + }, + { + "epoch": 19.282744282744282, + "grad_norm": 0.00018654775340110064, + "learning_rate": 0.2619382369930407, + "loss": 0.2719, + "num_input_tokens_seen": 7073808, + "step": 9275 + }, + { + "epoch": 19.293139293139294, + "grad_norm": 0.00035718854633159935, + "learning_rate": 0.261899017735981, + "loss": 0.2781, + "num_input_tokens_seen": 7077712, + "step": 9280 + }, + { + "epoch": 19.303534303534303, + "grad_norm": 0.00031909672543406487, + "learning_rate": 0.2618597812226866, + "loss": 0.2708, + "num_input_tokens_seen": 7081424, + "step": 9285 + }, + { + "epoch": 19.313929313929314, + "grad_norm": 0.00016799983859527856, + "learning_rate": 0.2618205274592082, + "loss": 0.249, + "num_input_tokens_seen": 7085264, + "step": 9290 + }, + { + "epoch": 19.324324324324323, + "grad_norm": 0.0005648421938531101, + "learning_rate": 0.2617812564515992, + "loss": 0.2411, + "num_input_tokens_seen": 7089136, + "step": 9295 + }, + { + "epoch": 19.334719334719335, + "grad_norm": 0.00031264216522686183, + "learning_rate": 0.2617419682059158, + "loss": 0.2935, + "num_input_tokens_seen": 7093040, + "step": 9300 + }, + { + "epoch": 19.345114345114347, + "grad_norm": 0.0014313278952613473, + "learning_rate": 0.26170266272821663, + "loss": 0.2583, + "num_input_tokens_seen": 7096848, + "step": 9305 + }, + { + "epoch": 19.355509355509355, + "grad_norm": 0.0004390967369545251, + "learning_rate": 0.26166334002456315, + "loss": 0.2752, + "num_input_tokens_seen": 7100592, + "step": 9310 + }, + { + "epoch": 19.365904365904367, + "grad_norm": 0.00035188955371268094, + "learning_rate": 0.2616240001010194, + "loss": 0.2691, + "num_input_tokens_seen": 7104368, + "step": 9315 + }, + { + "epoch": 19.376299376299375, + "grad_norm": 0.00023063593835104257, + "learning_rate": 0.26158464296365197, + "loss": 0.2851, + "num_input_tokens_seen": 7108176, + "step": 9320 + }, + { + "epoch": 19.386694386694387, + "grad_norm": 8.219943265430629e-05, + "learning_rate": 0.2615452686185304, + "loss": 0.2777, + "num_input_tokens_seen": 7112080, + "step": 9325 + }, + { + "epoch": 19.397089397089395, + "grad_norm": 9.776716615306213e-05, + "learning_rate": 0.26150587707172673, + "loss": 0.2817, + "num_input_tokens_seen": 7115952, + "step": 9330 + }, + { + "epoch": 19.407484407484407, + "grad_norm": 0.00042810491868294775, + "learning_rate": 0.2614664683293154, + "loss": 0.2818, + "num_input_tokens_seen": 7119856, + "step": 9335 + }, + { + "epoch": 19.41787941787942, + "grad_norm": 0.0006230822764337063, + "learning_rate": 0.26142704239737397, + "loss": 0.2647, + "num_input_tokens_seen": 7123792, + "step": 9340 + }, + { + "epoch": 19.428274428274428, + "grad_norm": 0.00037900966708548367, + "learning_rate": 0.26138759928198235, + "loss": 0.2594, + "num_input_tokens_seen": 7127568, + "step": 9345 + }, + { + "epoch": 19.43866943866944, + "grad_norm": 0.0005252771079540253, + "learning_rate": 0.26134813898922304, + "loss": 0.2884, + "num_input_tokens_seen": 7131344, + "step": 9350 + }, + { + "epoch": 19.449064449064448, + "grad_norm": 0.001312321750447154, + "learning_rate": 0.26130866152518145, + "loss": 0.2711, + "num_input_tokens_seen": 7135344, + "step": 9355 + }, + { + "epoch": 19.45945945945946, + "grad_norm": 0.0001656927925068885, + "learning_rate": 0.2612691668959455, + "loss": 0.2614, + "num_input_tokens_seen": 7139024, + "step": 9360 + }, + { + "epoch": 19.46985446985447, + "grad_norm": 0.00027781364042311907, + "learning_rate": 0.2612296551076057, + "loss": 0.2944, + "num_input_tokens_seen": 7142864, + "step": 9365 + }, + { + "epoch": 19.48024948024948, + "grad_norm": 0.0007671174826100469, + "learning_rate": 0.26119012616625525, + "loss": 0.2647, + "num_input_tokens_seen": 7146480, + "step": 9370 + }, + { + "epoch": 19.490644490644492, + "grad_norm": 8.600906585343182e-05, + "learning_rate": 0.26115058007799, + "loss": 0.2617, + "num_input_tokens_seen": 7150288, + "step": 9375 + }, + { + "epoch": 19.5010395010395, + "grad_norm": 0.0005850433371961117, + "learning_rate": 0.26111101684890864, + "loss": 0.2728, + "num_input_tokens_seen": 7154192, + "step": 9380 + }, + { + "epoch": 19.511434511434512, + "grad_norm": 0.00041858365875668824, + "learning_rate": 0.26107143648511205, + "loss": 0.2783, + "num_input_tokens_seen": 7158032, + "step": 9385 + }, + { + "epoch": 19.52182952182952, + "grad_norm": 0.00014683879271615297, + "learning_rate": 0.2610318389927042, + "loss": 0.2644, + "num_input_tokens_seen": 7162032, + "step": 9390 + }, + { + "epoch": 19.532224532224532, + "grad_norm": 0.0009800773113965988, + "learning_rate": 0.26099222437779146, + "loss": 0.2617, + "num_input_tokens_seen": 7165744, + "step": 9395 + }, + { + "epoch": 19.542619542619544, + "grad_norm": 0.0006388739566318691, + "learning_rate": 0.26095259264648285, + "loss": 0.2682, + "num_input_tokens_seen": 7169456, + "step": 9400 + }, + { + "epoch": 19.542619542619544, + "eval_loss": 0.2484639585018158, + "eval_runtime": 13.5127, + "eval_samples_per_second": 63.348, + "eval_steps_per_second": 15.837, + "num_input_tokens_seen": 7169456, + "step": 9400 + }, + { + "epoch": 19.553014553014552, + "grad_norm": 0.00011533289944054559, + "learning_rate": 0.2609129438048902, + "loss": 0.2732, + "num_input_tokens_seen": 7173200, + "step": 9405 + }, + { + "epoch": 19.563409563409564, + "grad_norm": 0.0007667131721973419, + "learning_rate": 0.2608732778591278, + "loss": 0.2647, + "num_input_tokens_seen": 7177072, + "step": 9410 + }, + { + "epoch": 19.573804573804573, + "grad_norm": 0.0008894977509044111, + "learning_rate": 0.2608335948153126, + "loss": 0.2836, + "num_input_tokens_seen": 7180784, + "step": 9415 + }, + { + "epoch": 19.584199584199585, + "grad_norm": 0.0003119780740235001, + "learning_rate": 0.26079389467956426, + "loss": 0.2856, + "num_input_tokens_seen": 7184656, + "step": 9420 + }, + { + "epoch": 19.594594594594593, + "grad_norm": 0.0007998279179446399, + "learning_rate": 0.26075417745800505, + "loss": 0.2803, + "num_input_tokens_seen": 7188528, + "step": 9425 + }, + { + "epoch": 19.604989604989605, + "grad_norm": 0.0005445594433695078, + "learning_rate": 0.26071444315675985, + "loss": 0.28, + "num_input_tokens_seen": 7192528, + "step": 9430 + }, + { + "epoch": 19.615384615384617, + "grad_norm": 0.00012255639012437314, + "learning_rate": 0.2606746917819562, + "loss": 0.2746, + "num_input_tokens_seen": 7196336, + "step": 9435 + }, + { + "epoch": 19.625779625779625, + "grad_norm": 0.0008745240047574043, + "learning_rate": 0.2606349233397242, + "loss": 0.2558, + "num_input_tokens_seen": 7200240, + "step": 9440 + }, + { + "epoch": 19.636174636174637, + "grad_norm": 0.00013615828356705606, + "learning_rate": 0.26059513783619676, + "loss": 0.2483, + "num_input_tokens_seen": 7203984, + "step": 9445 + }, + { + "epoch": 19.646569646569645, + "grad_norm": 0.0005787672125734389, + "learning_rate": 0.26055533527750924, + "loss": 0.316, + "num_input_tokens_seen": 7207792, + "step": 9450 + }, + { + "epoch": 19.656964656964657, + "grad_norm": 6.797454261686653e-05, + "learning_rate": 0.26051551566979964, + "loss": 0.2696, + "num_input_tokens_seen": 7211504, + "step": 9455 + }, + { + "epoch": 19.66735966735967, + "grad_norm": 0.0009281965903937817, + "learning_rate": 0.26047567901920876, + "loss": 0.2733, + "num_input_tokens_seen": 7215408, + "step": 9460 + }, + { + "epoch": 19.677754677754677, + "grad_norm": 0.00013899813347961754, + "learning_rate": 0.2604358253318798, + "loss": 0.2603, + "num_input_tokens_seen": 7219312, + "step": 9465 + }, + { + "epoch": 19.68814968814969, + "grad_norm": 0.00020733648852910846, + "learning_rate": 0.26039595461395876, + "loss": 0.2515, + "num_input_tokens_seen": 7223088, + "step": 9470 + }, + { + "epoch": 19.698544698544698, + "grad_norm": 6.337475497275591e-05, + "learning_rate": 0.26035606687159424, + "loss": 0.2115, + "num_input_tokens_seen": 7226896, + "step": 9475 + }, + { + "epoch": 19.70893970893971, + "grad_norm": 0.00015888236521277577, + "learning_rate": 0.26031616211093733, + "loss": 0.2753, + "num_input_tokens_seen": 7230608, + "step": 9480 + }, + { + "epoch": 19.719334719334718, + "grad_norm": 0.00015440689458046108, + "learning_rate": 0.26027624033814195, + "loss": 0.2827, + "num_input_tokens_seen": 7234544, + "step": 9485 + }, + { + "epoch": 19.72972972972973, + "grad_norm": 0.0013223260175436735, + "learning_rate": 0.2602363015593645, + "loss": 0.3068, + "num_input_tokens_seen": 7238256, + "step": 9490 + }, + { + "epoch": 19.74012474012474, + "grad_norm": 0.0003211788134649396, + "learning_rate": 0.26019634578076395, + "loss": 0.2836, + "num_input_tokens_seen": 7242192, + "step": 9495 + }, + { + "epoch": 19.75051975051975, + "grad_norm": 0.0006311276811175048, + "learning_rate": 0.26015637300850214, + "loss": 0.2358, + "num_input_tokens_seen": 7245968, + "step": 9500 + }, + { + "epoch": 19.760914760914762, + "grad_norm": 0.0005032388726249337, + "learning_rate": 0.26011638324874325, + "loss": 0.2884, + "num_input_tokens_seen": 7249872, + "step": 9505 + }, + { + "epoch": 19.77130977130977, + "grad_norm": 0.0005025675054639578, + "learning_rate": 0.2600763765076543, + "loss": 0.2675, + "num_input_tokens_seen": 7253680, + "step": 9510 + }, + { + "epoch": 19.781704781704782, + "grad_norm": 0.03965955972671509, + "learning_rate": 0.2600363527914048, + "loss": 0.4439, + "num_input_tokens_seen": 7257424, + "step": 9515 + }, + { + "epoch": 19.79209979209979, + "grad_norm": 0.005118821747601032, + "learning_rate": 0.25999631210616686, + "loss": 0.2639, + "num_input_tokens_seen": 7261232, + "step": 9520 + }, + { + "epoch": 19.802494802494802, + "grad_norm": 0.00065595842897892, + "learning_rate": 0.25995625445811527, + "loss": 0.2578, + "num_input_tokens_seen": 7265040, + "step": 9525 + }, + { + "epoch": 19.812889812889814, + "grad_norm": 0.0015273848548531532, + "learning_rate": 0.2599161798534275, + "loss": 0.2256, + "num_input_tokens_seen": 7268784, + "step": 9530 + }, + { + "epoch": 19.823284823284823, + "grad_norm": 0.0006779061513952911, + "learning_rate": 0.25987608829828346, + "loss": 0.2673, + "num_input_tokens_seen": 7272656, + "step": 9535 + }, + { + "epoch": 19.833679833679835, + "grad_norm": 0.0011744765797629952, + "learning_rate": 0.25983597979886586, + "loss": 0.2496, + "num_input_tokens_seen": 7276464, + "step": 9540 + }, + { + "epoch": 19.844074844074843, + "grad_norm": 0.0007340550655499101, + "learning_rate": 0.2597958543613599, + "loss": 0.2774, + "num_input_tokens_seen": 7280240, + "step": 9545 + }, + { + "epoch": 19.854469854469855, + "grad_norm": 0.000658744596876204, + "learning_rate": 0.25975571199195335, + "loss": 0.2739, + "num_input_tokens_seen": 7283952, + "step": 9550 + }, + { + "epoch": 19.864864864864863, + "grad_norm": 0.0026952181942760944, + "learning_rate": 0.25971555269683677, + "loss": 0.2896, + "num_input_tokens_seen": 7287728, + "step": 9555 + }, + { + "epoch": 19.875259875259875, + "grad_norm": 0.04383690282702446, + "learning_rate": 0.25967537648220324, + "loss": 0.6043, + "num_input_tokens_seen": 7291856, + "step": 9560 + }, + { + "epoch": 19.885654885654887, + "grad_norm": 0.0022812476381659508, + "learning_rate": 0.2596351833542483, + "loss": 0.3452, + "num_input_tokens_seen": 7295600, + "step": 9565 + }, + { + "epoch": 19.896049896049895, + "grad_norm": 0.001968629891052842, + "learning_rate": 0.25959497331917036, + "loss": 0.3131, + "num_input_tokens_seen": 7299472, + "step": 9570 + }, + { + "epoch": 19.906444906444907, + "grad_norm": 0.000367986096534878, + "learning_rate": 0.2595547463831703, + "loss": 0.2942, + "num_input_tokens_seen": 7303344, + "step": 9575 + }, + { + "epoch": 19.916839916839916, + "grad_norm": 0.0011965053854510188, + "learning_rate": 0.25951450255245156, + "loss": 0.2876, + "num_input_tokens_seen": 7307152, + "step": 9580 + }, + { + "epoch": 19.927234927234927, + "grad_norm": 0.0003532941045705229, + "learning_rate": 0.2594742418332203, + "loss": 0.2766, + "num_input_tokens_seen": 7311056, + "step": 9585 + }, + { + "epoch": 19.93762993762994, + "grad_norm": 0.0007409484242089093, + "learning_rate": 0.2594339642316852, + "loss": 0.2898, + "num_input_tokens_seen": 7315088, + "step": 9590 + }, + { + "epoch": 19.948024948024948, + "grad_norm": 0.0010140087688341737, + "learning_rate": 0.2593936697540576, + "loss": 0.2852, + "num_input_tokens_seen": 7319024, + "step": 9595 + }, + { + "epoch": 19.95841995841996, + "grad_norm": 0.001010441337712109, + "learning_rate": 0.2593533584065514, + "loss": 0.2833, + "num_input_tokens_seen": 7322736, + "step": 9600 + }, + { + "epoch": 19.95841995841996, + "eval_loss": 0.2534381151199341, + "eval_runtime": 13.4488, + "eval_samples_per_second": 63.649, + "eval_steps_per_second": 15.912, + "num_input_tokens_seen": 7322736, + "step": 9600 + }, + { + "epoch": 19.968814968814968, + "grad_norm": 0.0007372833206318319, + "learning_rate": 0.2593130301953831, + "loss": 0.2804, + "num_input_tokens_seen": 7326448, + "step": 9605 + }, + { + "epoch": 19.97920997920998, + "grad_norm": 0.0008373813470825553, + "learning_rate": 0.2592726851267718, + "loss": 0.2646, + "num_input_tokens_seen": 7330320, + "step": 9610 + }, + { + "epoch": 19.989604989604988, + "grad_norm": 0.00010396361176390201, + "learning_rate": 0.2592323232069393, + "loss": 0.2772, + "num_input_tokens_seen": 7334192, + "step": 9615 + }, + { + "epoch": 20.0, + "grad_norm": 0.0002812615130096674, + "learning_rate": 0.25919194444210986, + "loss": 0.2622, + "num_input_tokens_seen": 7337920, + "step": 9620 + }, + { + "epoch": 20.010395010395012, + "grad_norm": 0.00020592918735928833, + "learning_rate": 0.2591515488385103, + "loss": 0.2429, + "num_input_tokens_seen": 7341696, + "step": 9625 + }, + { + "epoch": 20.02079002079002, + "grad_norm": 0.0008334507001563907, + "learning_rate": 0.2591111364023704, + "loss": 0.3063, + "num_input_tokens_seen": 7345504, + "step": 9630 + }, + { + "epoch": 20.031185031185032, + "grad_norm": 0.00033460030681453645, + "learning_rate": 0.259070707139922, + "loss": 0.2648, + "num_input_tokens_seen": 7349216, + "step": 9635 + }, + { + "epoch": 20.04158004158004, + "grad_norm": 0.00011835694749606773, + "learning_rate": 0.25903026105739985, + "loss": 0.2671, + "num_input_tokens_seen": 7352960, + "step": 9640 + }, + { + "epoch": 20.051975051975052, + "grad_norm": 0.0007295488030649722, + "learning_rate": 0.2589897981610413, + "loss": 0.2547, + "num_input_tokens_seen": 7356672, + "step": 9645 + }, + { + "epoch": 20.06237006237006, + "grad_norm": 0.0009349461179226637, + "learning_rate": 0.2589493184570863, + "loss": 0.2629, + "num_input_tokens_seen": 7360608, + "step": 9650 + }, + { + "epoch": 20.072765072765073, + "grad_norm": 0.0007338064024224877, + "learning_rate": 0.25890882195177717, + "loss": 0.2556, + "num_input_tokens_seen": 7364512, + "step": 9655 + }, + { + "epoch": 20.083160083160084, + "grad_norm": 0.00016088932170532644, + "learning_rate": 0.25886830865135907, + "loss": 0.2756, + "num_input_tokens_seen": 7368128, + "step": 9660 + }, + { + "epoch": 20.093555093555093, + "grad_norm": 0.0022302772849798203, + "learning_rate": 0.25882777856207967, + "loss": 0.2782, + "num_input_tokens_seen": 7371872, + "step": 9665 + }, + { + "epoch": 20.103950103950105, + "grad_norm": 0.0006010463112033904, + "learning_rate": 0.2587872316901892, + "loss": 0.266, + "num_input_tokens_seen": 7375776, + "step": 9670 + }, + { + "epoch": 20.114345114345113, + "grad_norm": 0.00026487052673473954, + "learning_rate": 0.25874666804194046, + "loss": 0.2673, + "num_input_tokens_seen": 7379552, + "step": 9675 + }, + { + "epoch": 20.124740124740125, + "grad_norm": 0.0009183789370581508, + "learning_rate": 0.258706087623589, + "loss": 0.2788, + "num_input_tokens_seen": 7383328, + "step": 9680 + }, + { + "epoch": 20.135135135135137, + "grad_norm": 0.0006324736168608069, + "learning_rate": 0.25866549044139264, + "loss": 0.2627, + "num_input_tokens_seen": 7387040, + "step": 9685 + }, + { + "epoch": 20.145530145530145, + "grad_norm": 0.0010920120403170586, + "learning_rate": 0.25862487650161214, + "loss": 0.2796, + "num_input_tokens_seen": 7390912, + "step": 9690 + }, + { + "epoch": 20.155925155925157, + "grad_norm": 0.0003527570515871048, + "learning_rate": 0.2585842458105106, + "loss": 0.2667, + "num_input_tokens_seen": 7394720, + "step": 9695 + }, + { + "epoch": 20.166320166320165, + "grad_norm": 0.0004457251343410462, + "learning_rate": 0.2585435983743538, + "loss": 0.2679, + "num_input_tokens_seen": 7398432, + "step": 9700 + }, + { + "epoch": 20.176715176715177, + "grad_norm": 0.00019624625565484166, + "learning_rate": 0.2585029341994101, + "loss": 0.2579, + "num_input_tokens_seen": 7402176, + "step": 9705 + }, + { + "epoch": 20.187110187110186, + "grad_norm": 0.0003071998944506049, + "learning_rate": 0.2584622532919504, + "loss": 0.2727, + "num_input_tokens_seen": 7406016, + "step": 9710 + }, + { + "epoch": 20.197505197505198, + "grad_norm": 0.0002194322441937402, + "learning_rate": 0.2584215556582482, + "loss": 0.2724, + "num_input_tokens_seen": 7409760, + "step": 9715 + }, + { + "epoch": 20.20790020790021, + "grad_norm": 9.876451804302633e-05, + "learning_rate": 0.25838084130457967, + "loss": 0.2872, + "num_input_tokens_seen": 7413568, + "step": 9720 + }, + { + "epoch": 20.218295218295218, + "grad_norm": 0.0002197983703808859, + "learning_rate": 0.2583401102372234, + "loss": 0.2761, + "num_input_tokens_seen": 7417440, + "step": 9725 + }, + { + "epoch": 20.22869022869023, + "grad_norm": 0.00025207240832969546, + "learning_rate": 0.2582993624624606, + "loss": 0.2601, + "num_input_tokens_seen": 7421184, + "step": 9730 + }, + { + "epoch": 20.239085239085238, + "grad_norm": 0.0005256514414213598, + "learning_rate": 0.25825859798657513, + "loss": 0.2699, + "num_input_tokens_seen": 7425024, + "step": 9735 + }, + { + "epoch": 20.24948024948025, + "grad_norm": 0.00022195992642082274, + "learning_rate": 0.25821781681585343, + "loss": 0.2753, + "num_input_tokens_seen": 7428896, + "step": 9740 + }, + { + "epoch": 20.25987525987526, + "grad_norm": 0.0005951013299636543, + "learning_rate": 0.2581770189565844, + "loss": 0.2621, + "num_input_tokens_seen": 7432736, + "step": 9745 + }, + { + "epoch": 20.27027027027027, + "grad_norm": 0.0007667667814530432, + "learning_rate": 0.25813620441505963, + "loss": 0.2847, + "num_input_tokens_seen": 7436576, + "step": 9750 + }, + { + "epoch": 20.280665280665282, + "grad_norm": 0.0010342283640056849, + "learning_rate": 0.2580953731975732, + "loss": 0.2742, + "num_input_tokens_seen": 7440416, + "step": 9755 + }, + { + "epoch": 20.29106029106029, + "grad_norm": 0.0021008229814469814, + "learning_rate": 0.2580545253104218, + "loss": 0.2748, + "num_input_tokens_seen": 7444512, + "step": 9760 + }, + { + "epoch": 20.301455301455302, + "grad_norm": 0.0007322630845010281, + "learning_rate": 0.2580136607599047, + "loss": 0.2675, + "num_input_tokens_seen": 7448384, + "step": 9765 + }, + { + "epoch": 20.31185031185031, + "grad_norm": 0.00025445473147556186, + "learning_rate": 0.2579727795523238, + "loss": 0.2903, + "num_input_tokens_seen": 7452256, + "step": 9770 + }, + { + "epoch": 20.322245322245323, + "grad_norm": 0.00046907784417271614, + "learning_rate": 0.25793188169398334, + "loss": 0.2735, + "num_input_tokens_seen": 7456192, + "step": 9775 + }, + { + "epoch": 20.33264033264033, + "grad_norm": 0.0007322622113861144, + "learning_rate": 0.25789096719119037, + "loss": 0.2819, + "num_input_tokens_seen": 7460000, + "step": 9780 + }, + { + "epoch": 20.343035343035343, + "grad_norm": 0.0002992659283336252, + "learning_rate": 0.2578500360502544, + "loss": 0.2549, + "num_input_tokens_seen": 7463712, + "step": 9785 + }, + { + "epoch": 20.353430353430355, + "grad_norm": 0.000893943477421999, + "learning_rate": 0.2578090882774876, + "loss": 0.2948, + "num_input_tokens_seen": 7467488, + "step": 9790 + }, + { + "epoch": 20.363825363825363, + "grad_norm": 0.0009516054415144026, + "learning_rate": 0.25776812387920456, + "loss": 0.2806, + "num_input_tokens_seen": 7471168, + "step": 9795 + }, + { + "epoch": 20.374220374220375, + "grad_norm": 0.00033791764872148633, + "learning_rate": 0.2577271428617225, + "loss": 0.2695, + "num_input_tokens_seen": 7474848, + "step": 9800 + }, + { + "epoch": 20.374220374220375, + "eval_loss": 0.25159597396850586, + "eval_runtime": 13.4199, + "eval_samples_per_second": 63.786, + "eval_steps_per_second": 15.946, + "num_input_tokens_seen": 7474848, + "step": 9800 + }, + { + "epoch": 20.384615384615383, + "grad_norm": 8.797948248684406e-05, + "learning_rate": 0.25768614523136124, + "loss": 0.2517, + "num_input_tokens_seen": 7478560, + "step": 9805 + }, + { + "epoch": 20.395010395010395, + "grad_norm": 0.0013992168242111802, + "learning_rate": 0.25764513099444314, + "loss": 0.2617, + "num_input_tokens_seen": 7482400, + "step": 9810 + }, + { + "epoch": 20.405405405405407, + "grad_norm": 0.0007539973012171686, + "learning_rate": 0.25760410015729307, + "loss": 0.2801, + "num_input_tokens_seen": 7486272, + "step": 9815 + }, + { + "epoch": 20.415800415800415, + "grad_norm": 0.0001553182810312137, + "learning_rate": 0.2575630527262385, + "loss": 0.2691, + "num_input_tokens_seen": 7490144, + "step": 9820 + }, + { + "epoch": 20.426195426195427, + "grad_norm": 0.0006334925419650972, + "learning_rate": 0.25752198870760945, + "loss": 0.264, + "num_input_tokens_seen": 7493856, + "step": 9825 + }, + { + "epoch": 20.436590436590436, + "grad_norm": 0.00036544946487993, + "learning_rate": 0.2574809081077386, + "loss": 0.2717, + "num_input_tokens_seen": 7497632, + "step": 9830 + }, + { + "epoch": 20.446985446985448, + "grad_norm": 0.0003197744663339108, + "learning_rate": 0.257439810932961, + "loss": 0.2719, + "num_input_tokens_seen": 7501408, + "step": 9835 + }, + { + "epoch": 20.457380457380456, + "grad_norm": 0.0008498377283103764, + "learning_rate": 0.2573986971896144, + "loss": 0.2618, + "num_input_tokens_seen": 7505344, + "step": 9840 + }, + { + "epoch": 20.467775467775468, + "grad_norm": 0.00033469058689661324, + "learning_rate": 0.257357566884039, + "loss": 0.2736, + "num_input_tokens_seen": 7509376, + "step": 9845 + }, + { + "epoch": 20.47817047817048, + "grad_norm": 0.00034822215093299747, + "learning_rate": 0.25731642002257765, + "loss": 0.2763, + "num_input_tokens_seen": 7513248, + "step": 9850 + }, + { + "epoch": 20.488565488565488, + "grad_norm": 0.0014490610919892788, + "learning_rate": 0.25727525661157574, + "loss": 0.2894, + "num_input_tokens_seen": 7517184, + "step": 9855 + }, + { + "epoch": 20.4989604989605, + "grad_norm": 0.0005810630973428488, + "learning_rate": 0.2572340766573811, + "loss": 0.2576, + "num_input_tokens_seen": 7521056, + "step": 9860 + }, + { + "epoch": 20.509355509355508, + "grad_norm": 0.0001826606021495536, + "learning_rate": 0.25719288016634434, + "loss": 0.2497, + "num_input_tokens_seen": 7524960, + "step": 9865 + }, + { + "epoch": 20.51975051975052, + "grad_norm": 0.0007309364737011492, + "learning_rate": 0.25715166714481835, + "loss": 0.2728, + "num_input_tokens_seen": 7528704, + "step": 9870 + }, + { + "epoch": 20.53014553014553, + "grad_norm": 0.00028674094937741756, + "learning_rate": 0.2571104375991587, + "loss": 0.2209, + "num_input_tokens_seen": 7532576, + "step": 9875 + }, + { + "epoch": 20.54054054054054, + "grad_norm": 0.00024698779452592134, + "learning_rate": 0.2570691915357236, + "loss": 0.2693, + "num_input_tokens_seen": 7536416, + "step": 9880 + }, + { + "epoch": 20.550935550935552, + "grad_norm": 0.00039102367009036243, + "learning_rate": 0.2570279289608736, + "loss": 0.272, + "num_input_tokens_seen": 7540096, + "step": 9885 + }, + { + "epoch": 20.56133056133056, + "grad_norm": 0.00012924833572469652, + "learning_rate": 0.256986649880972, + "loss": 0.2837, + "num_input_tokens_seen": 7543872, + "step": 9890 + }, + { + "epoch": 20.571725571725572, + "grad_norm": 0.0007486614631488919, + "learning_rate": 0.25694535430238447, + "loss": 0.2788, + "num_input_tokens_seen": 7547680, + "step": 9895 + }, + { + "epoch": 20.58212058212058, + "grad_norm": 0.0007160227396525443, + "learning_rate": 0.25690404223147933, + "loss": 0.2522, + "num_input_tokens_seen": 7551424, + "step": 9900 + }, + { + "epoch": 20.592515592515593, + "grad_norm": 0.001348499208688736, + "learning_rate": 0.2568627136746275, + "loss": 0.2796, + "num_input_tokens_seen": 7555232, + "step": 9905 + }, + { + "epoch": 20.602910602910605, + "grad_norm": 0.0009232627926394343, + "learning_rate": 0.25682136863820226, + "loss": 0.2903, + "num_input_tokens_seen": 7559008, + "step": 9910 + }, + { + "epoch": 20.613305613305613, + "grad_norm": 0.0003160377382300794, + "learning_rate": 0.25678000712857957, + "loss": 0.2805, + "num_input_tokens_seen": 7562848, + "step": 9915 + }, + { + "epoch": 20.623700623700625, + "grad_norm": 0.000509598758071661, + "learning_rate": 0.2567386291521379, + "loss": 0.2229, + "num_input_tokens_seen": 7566624, + "step": 9920 + }, + { + "epoch": 20.634095634095633, + "grad_norm": 0.00032215958344750106, + "learning_rate": 0.2566972347152583, + "loss": 0.2889, + "num_input_tokens_seen": 7570464, + "step": 9925 + }, + { + "epoch": 20.644490644490645, + "grad_norm": 0.0004113535978831351, + "learning_rate": 0.2566558238243242, + "loss": 0.3072, + "num_input_tokens_seen": 7574400, + "step": 9930 + }, + { + "epoch": 20.654885654885653, + "grad_norm": 0.00024217707687057555, + "learning_rate": 0.25661439648572176, + "loss": 0.2576, + "num_input_tokens_seen": 7578176, + "step": 9935 + }, + { + "epoch": 20.665280665280665, + "grad_norm": 0.0017230167286470532, + "learning_rate": 0.25657295270583963, + "loss": 0.2495, + "num_input_tokens_seen": 7581888, + "step": 9940 + }, + { + "epoch": 20.675675675675677, + "grad_norm": 0.0003539897734299302, + "learning_rate": 0.25653149249106894, + "loss": 0.2644, + "num_input_tokens_seen": 7585760, + "step": 9945 + }, + { + "epoch": 20.686070686070686, + "grad_norm": 0.000334468757500872, + "learning_rate": 0.25649001584780323, + "loss": 0.2343, + "num_input_tokens_seen": 7589568, + "step": 9950 + }, + { + "epoch": 20.696465696465697, + "grad_norm": 0.0006285231211222708, + "learning_rate": 0.2564485227824389, + "loss": 0.2659, + "num_input_tokens_seen": 7593440, + "step": 9955 + }, + { + "epoch": 20.706860706860706, + "grad_norm": 0.0005435169441625476, + "learning_rate": 0.25640701330137466, + "loss": 0.273, + "num_input_tokens_seen": 7597312, + "step": 9960 + }, + { + "epoch": 20.717255717255718, + "grad_norm": 0.0008654603152535856, + "learning_rate": 0.2563654874110117, + "loss": 0.2667, + "num_input_tokens_seen": 7601152, + "step": 9965 + }, + { + "epoch": 20.727650727650726, + "grad_norm": 0.00017654300609137863, + "learning_rate": 0.256323945117754, + "loss": 0.2629, + "num_input_tokens_seen": 7604832, + "step": 9970 + }, + { + "epoch": 20.738045738045738, + "grad_norm": 0.0002609273069538176, + "learning_rate": 0.2562823864280078, + "loss": 0.2307, + "num_input_tokens_seen": 7608544, + "step": 9975 + }, + { + "epoch": 20.74844074844075, + "grad_norm": 0.00043342800927348435, + "learning_rate": 0.25624081134818194, + "loss": 0.2535, + "num_input_tokens_seen": 7612320, + "step": 9980 + }, + { + "epoch": 20.758835758835758, + "grad_norm": 0.00020509760361164808, + "learning_rate": 0.2561992198846879, + "loss": 0.2662, + "num_input_tokens_seen": 7616096, + "step": 9985 + }, + { + "epoch": 20.76923076923077, + "grad_norm": 0.0001900003699120134, + "learning_rate": 0.25615761204393955, + "loss": 0.2493, + "num_input_tokens_seen": 7619872, + "step": 9990 + }, + { + "epoch": 20.77962577962578, + "grad_norm": 0.000363758736057207, + "learning_rate": 0.2561159878323534, + "loss": 0.2442, + "num_input_tokens_seen": 7623552, + "step": 9995 + }, + { + "epoch": 20.79002079002079, + "grad_norm": 0.0004418054304551333, + "learning_rate": 0.2560743472563483, + "loss": 0.2552, + "num_input_tokens_seen": 7627360, + "step": 10000 + }, + { + "epoch": 20.79002079002079, + "eval_loss": 0.24894678592681885, + "eval_runtime": 13.4271, + "eval_samples_per_second": 63.751, + "eval_steps_per_second": 15.938, + "num_input_tokens_seen": 7627360, + "step": 10000 + }, + { + "epoch": 20.8004158004158, + "grad_norm": 0.0006407376495189965, + "learning_rate": 0.25603269032234593, + "loss": 0.2771, + "num_input_tokens_seen": 7631232, + "step": 10005 + }, + { + "epoch": 20.81081081081081, + "grad_norm": 0.0007491199648939073, + "learning_rate": 0.2559910170367702, + "loss": 0.2777, + "num_input_tokens_seen": 7635072, + "step": 10010 + }, + { + "epoch": 20.821205821205822, + "grad_norm": 0.0012122910702601075, + "learning_rate": 0.2559493274060477, + "loss": 0.2741, + "num_input_tokens_seen": 7638848, + "step": 10015 + }, + { + "epoch": 20.83160083160083, + "grad_norm": 0.0010487971594557166, + "learning_rate": 0.2559076214366074, + "loss": 0.274, + "num_input_tokens_seen": 7642752, + "step": 10020 + }, + { + "epoch": 20.841995841995843, + "grad_norm": 0.00025001520407386124, + "learning_rate": 0.25586589913488106, + "loss": 0.2621, + "num_input_tokens_seen": 7646528, + "step": 10025 + }, + { + "epoch": 20.85239085239085, + "grad_norm": 0.001138955820351839, + "learning_rate": 0.2558241605073026, + "loss": 0.2703, + "num_input_tokens_seen": 7650368, + "step": 10030 + }, + { + "epoch": 20.862785862785863, + "grad_norm": 0.00033714191522449255, + "learning_rate": 0.25578240556030873, + "loss": 0.2699, + "num_input_tokens_seen": 7654272, + "step": 10035 + }, + { + "epoch": 20.873180873180875, + "grad_norm": 0.0009100576862692833, + "learning_rate": 0.2557406343003386, + "loss": 0.2823, + "num_input_tokens_seen": 7658048, + "step": 10040 + }, + { + "epoch": 20.883575883575883, + "grad_norm": 0.0009798434330150485, + "learning_rate": 0.25569884673383375, + "loss": 0.2898, + "num_input_tokens_seen": 7661984, + "step": 10045 + }, + { + "epoch": 20.893970893970895, + "grad_norm": 0.00022918320610187948, + "learning_rate": 0.25565704286723856, + "loss": 0.2781, + "num_input_tokens_seen": 7665728, + "step": 10050 + }, + { + "epoch": 20.904365904365903, + "grad_norm": 0.0010545308468863368, + "learning_rate": 0.25561522270699955, + "loss": 0.2716, + "num_input_tokens_seen": 7669536, + "step": 10055 + }, + { + "epoch": 20.914760914760915, + "grad_norm": 0.0003725242568179965, + "learning_rate": 0.25557338625956594, + "loss": 0.2825, + "num_input_tokens_seen": 7673344, + "step": 10060 + }, + { + "epoch": 20.925155925155924, + "grad_norm": 0.0005269638495519757, + "learning_rate": 0.25553153353138947, + "loss": 0.2602, + "num_input_tokens_seen": 7677312, + "step": 10065 + }, + { + "epoch": 20.935550935550935, + "grad_norm": 0.0002566588227637112, + "learning_rate": 0.2554896645289243, + "loss": 0.276, + "num_input_tokens_seen": 7681024, + "step": 10070 + }, + { + "epoch": 20.945945945945947, + "grad_norm": 0.0005664086784236133, + "learning_rate": 0.2554477792586272, + "loss": 0.2685, + "num_input_tokens_seen": 7684960, + "step": 10075 + }, + { + "epoch": 20.956340956340956, + "grad_norm": 0.000668698747176677, + "learning_rate": 0.25540587772695744, + "loss": 0.2683, + "num_input_tokens_seen": 7688736, + "step": 10080 + }, + { + "epoch": 20.966735966735968, + "grad_norm": 0.00025064373039640486, + "learning_rate": 0.2553639599403767, + "loss": 0.2804, + "num_input_tokens_seen": 7692704, + "step": 10085 + }, + { + "epoch": 20.977130977130976, + "grad_norm": 0.00024317228235304356, + "learning_rate": 0.2553220259053493, + "loss": 0.2681, + "num_input_tokens_seen": 7696416, + "step": 10090 + }, + { + "epoch": 20.987525987525988, + "grad_norm": 0.0006572874845005572, + "learning_rate": 0.2552800756283419, + "loss": 0.2749, + "num_input_tokens_seen": 7700256, + "step": 10095 + }, + { + "epoch": 20.997920997921, + "grad_norm": 9.075023262994364e-05, + "learning_rate": 0.25523810911582373, + "loss": 0.2799, + "num_input_tokens_seen": 7704096, + "step": 10100 + }, + { + "epoch": 21.008316008316008, + "grad_norm": 0.0001262922160094604, + "learning_rate": 0.25519612637426675, + "loss": 0.2768, + "num_input_tokens_seen": 7707792, + "step": 10105 + }, + { + "epoch": 21.01871101871102, + "grad_norm": 0.0014157682890072465, + "learning_rate": 0.25515412741014504, + "loss": 0.2899, + "num_input_tokens_seen": 7711440, + "step": 10110 + }, + { + "epoch": 21.02910602910603, + "grad_norm": 0.0004617314843926579, + "learning_rate": 0.2551121122299355, + "loss": 0.2706, + "num_input_tokens_seen": 7715472, + "step": 10115 + }, + { + "epoch": 21.03950103950104, + "grad_norm": 0.0007789027877151966, + "learning_rate": 0.2550700808401173, + "loss": 0.2766, + "num_input_tokens_seen": 7719376, + "step": 10120 + }, + { + "epoch": 21.04989604989605, + "grad_norm": 0.00013219054380897433, + "learning_rate": 0.2550280332471722, + "loss": 0.2775, + "num_input_tokens_seen": 7723152, + "step": 10125 + }, + { + "epoch": 21.06029106029106, + "grad_norm": 0.0005019409582018852, + "learning_rate": 0.2549859694575845, + "loss": 0.2737, + "num_input_tokens_seen": 7726960, + "step": 10130 + }, + { + "epoch": 21.070686070686072, + "grad_norm": 0.0005093032377772033, + "learning_rate": 0.254943889477841, + "loss": 0.2408, + "num_input_tokens_seen": 7730704, + "step": 10135 + }, + { + "epoch": 21.08108108108108, + "grad_norm": 0.0010023302165791392, + "learning_rate": 0.25490179331443097, + "loss": 0.2966, + "num_input_tokens_seen": 7734512, + "step": 10140 + }, + { + "epoch": 21.091476091476093, + "grad_norm": 0.0007999064400792122, + "learning_rate": 0.25485968097384615, + "loss": 0.3068, + "num_input_tokens_seen": 7738288, + "step": 10145 + }, + { + "epoch": 21.1018711018711, + "grad_norm": 0.0008813382592052221, + "learning_rate": 0.25481755246258075, + "loss": 0.2654, + "num_input_tokens_seen": 7742032, + "step": 10150 + }, + { + "epoch": 21.112266112266113, + "grad_norm": 0.0009257479105144739, + "learning_rate": 0.2547754077871315, + "loss": 0.261, + "num_input_tokens_seen": 7745936, + "step": 10155 + }, + { + "epoch": 21.12266112266112, + "grad_norm": 0.000368928856914863, + "learning_rate": 0.25473324695399774, + "loss": 0.2379, + "num_input_tokens_seen": 7749712, + "step": 10160 + }, + { + "epoch": 21.133056133056133, + "grad_norm": 0.0005834809853695333, + "learning_rate": 0.25469106996968105, + "loss": 0.2869, + "num_input_tokens_seen": 7753616, + "step": 10165 + }, + { + "epoch": 21.143451143451145, + "grad_norm": 0.0004701579746324569, + "learning_rate": 0.2546488768406858, + "loss": 0.2059, + "num_input_tokens_seen": 7757392, + "step": 10170 + }, + { + "epoch": 21.153846153846153, + "grad_norm": 0.0005686631775461137, + "learning_rate": 0.25460666757351863, + "loss": 0.2578, + "num_input_tokens_seen": 7761168, + "step": 10175 + }, + { + "epoch": 21.164241164241165, + "grad_norm": 0.00017975068476516753, + "learning_rate": 0.25456444217468877, + "loss": 0.2558, + "num_input_tokens_seen": 7765008, + "step": 10180 + }, + { + "epoch": 21.174636174636174, + "grad_norm": 0.0004805494681932032, + "learning_rate": 0.25452220065070785, + "loss": 0.2751, + "num_input_tokens_seen": 7768784, + "step": 10185 + }, + { + "epoch": 21.185031185031185, + "grad_norm": 0.00020410734578035772, + "learning_rate": 0.2544799430080901, + "loss": 0.2718, + "num_input_tokens_seen": 7772496, + "step": 10190 + }, + { + "epoch": 21.195426195426194, + "grad_norm": 0.00018421626009512693, + "learning_rate": 0.2544376692533522, + "loss": 0.2871, + "num_input_tokens_seen": 7776208, + "step": 10195 + }, + { + "epoch": 21.205821205821206, + "grad_norm": 0.0006492980755865574, + "learning_rate": 0.2543953793930132, + "loss": 0.2899, + "num_input_tokens_seen": 7779952, + "step": 10200 + }, + { + "epoch": 21.205821205821206, + "eval_loss": 0.28288534283638, + "eval_runtime": 13.409, + "eval_samples_per_second": 63.838, + "eval_steps_per_second": 15.959, + "num_input_tokens_seen": 7779952, + "step": 10200 + }, + { + "epoch": 21.216216216216218, + "grad_norm": 0.0006508665392175317, + "learning_rate": 0.2543530734335948, + "loss": 0.273, + "num_input_tokens_seen": 7783856, + "step": 10205 + }, + { + "epoch": 21.226611226611226, + "grad_norm": 0.0001785898202797398, + "learning_rate": 0.2543107513816211, + "loss": 0.2718, + "num_input_tokens_seen": 7787728, + "step": 10210 + }, + { + "epoch": 21.237006237006238, + "grad_norm": 6.725174171151593e-05, + "learning_rate": 0.25426841324361865, + "loss": 0.2359, + "num_input_tokens_seen": 7791664, + "step": 10215 + }, + { + "epoch": 21.247401247401246, + "grad_norm": 0.00042148795910179615, + "learning_rate": 0.2542260590261166, + "loss": 0.2726, + "num_input_tokens_seen": 7795472, + "step": 10220 + }, + { + "epoch": 21.257796257796258, + "grad_norm": 0.00023086188593879342, + "learning_rate": 0.2541836887356465, + "loss": 0.273, + "num_input_tokens_seen": 7799280, + "step": 10225 + }, + { + "epoch": 21.26819126819127, + "grad_norm": 0.00020786735694855452, + "learning_rate": 0.2541413023787423, + "loss": 0.2751, + "num_input_tokens_seen": 7802992, + "step": 10230 + }, + { + "epoch": 21.27858627858628, + "grad_norm": 0.0010975359473377466, + "learning_rate": 0.2540988999619405, + "loss": 0.2831, + "num_input_tokens_seen": 7806832, + "step": 10235 + }, + { + "epoch": 21.28898128898129, + "grad_norm": 0.00010360888700233772, + "learning_rate": 0.25405648149178023, + "loss": 0.2701, + "num_input_tokens_seen": 7810544, + "step": 10240 + }, + { + "epoch": 21.2993762993763, + "grad_norm": 0.00020738496095873415, + "learning_rate": 0.2540140469748028, + "loss": 0.268, + "num_input_tokens_seen": 7814192, + "step": 10245 + }, + { + "epoch": 21.30977130977131, + "grad_norm": 0.0003965446085203439, + "learning_rate": 0.25397159641755224, + "loss": 0.2814, + "num_input_tokens_seen": 7818064, + "step": 10250 + }, + { + "epoch": 21.32016632016632, + "grad_norm": 0.00012908197822980583, + "learning_rate": 0.2539291298265749, + "loss": 0.2769, + "num_input_tokens_seen": 7821904, + "step": 10255 + }, + { + "epoch": 21.33056133056133, + "grad_norm": 0.0003077381697949022, + "learning_rate": 0.2538866472084197, + "loss": 0.2785, + "num_input_tokens_seen": 7825616, + "step": 10260 + }, + { + "epoch": 21.340956340956343, + "grad_norm": 0.000493488390929997, + "learning_rate": 0.25384414856963794, + "loss": 0.2562, + "num_input_tokens_seen": 7829360, + "step": 10265 + }, + { + "epoch": 21.35135135135135, + "grad_norm": 0.00016070547280833125, + "learning_rate": 0.25380163391678356, + "loss": 0.2738, + "num_input_tokens_seen": 7833200, + "step": 10270 + }, + { + "epoch": 21.361746361746363, + "grad_norm": 0.0006622174405492842, + "learning_rate": 0.2537591032564127, + "loss": 0.2475, + "num_input_tokens_seen": 7837040, + "step": 10275 + }, + { + "epoch": 21.37214137214137, + "grad_norm": 0.0009966210927814245, + "learning_rate": 0.25371655659508424, + "loss": 0.2871, + "num_input_tokens_seen": 7840720, + "step": 10280 + }, + { + "epoch": 21.382536382536383, + "grad_norm": 0.0007196349324658513, + "learning_rate": 0.25367399393935935, + "loss": 0.2812, + "num_input_tokens_seen": 7844496, + "step": 10285 + }, + { + "epoch": 21.39293139293139, + "grad_norm": 0.0010075728641822934, + "learning_rate": 0.25363141529580174, + "loss": 0.2871, + "num_input_tokens_seen": 7848624, + "step": 10290 + }, + { + "epoch": 21.403326403326403, + "grad_norm": 0.0006293815677054226, + "learning_rate": 0.2535888206709776, + "loss": 0.2491, + "num_input_tokens_seen": 7852304, + "step": 10295 + }, + { + "epoch": 21.413721413721415, + "grad_norm": 0.000508844677824527, + "learning_rate": 0.2535462100714555, + "loss": 0.2878, + "num_input_tokens_seen": 7856208, + "step": 10300 + }, + { + "epoch": 21.424116424116423, + "grad_norm": 0.000340888713253662, + "learning_rate": 0.2535035835038066, + "loss": 0.2555, + "num_input_tokens_seen": 7859856, + "step": 10305 + }, + { + "epoch": 21.434511434511435, + "grad_norm": 0.00020129710901528597, + "learning_rate": 0.2534609409746044, + "loss": 0.2759, + "num_input_tokens_seen": 7863664, + "step": 10310 + }, + { + "epoch": 21.444906444906444, + "grad_norm": 0.0003353781357873231, + "learning_rate": 0.253418282490425, + "loss": 0.2764, + "num_input_tokens_seen": 7867440, + "step": 10315 + }, + { + "epoch": 21.455301455301456, + "grad_norm": 0.0006508993683382869, + "learning_rate": 0.2533756080578467, + "loss": 0.2649, + "num_input_tokens_seen": 7871280, + "step": 10320 + }, + { + "epoch": 21.465696465696467, + "grad_norm": 0.00023910297022666782, + "learning_rate": 0.25333291768345056, + "loss": 0.2258, + "num_input_tokens_seen": 7875344, + "step": 10325 + }, + { + "epoch": 21.476091476091476, + "grad_norm": 5.987012991681695e-05, + "learning_rate": 0.25329021137381996, + "loss": 0.2548, + "num_input_tokens_seen": 7879312, + "step": 10330 + }, + { + "epoch": 21.486486486486488, + "grad_norm": 0.0001869567931862548, + "learning_rate": 0.25324748913554074, + "loss": 0.2932, + "num_input_tokens_seen": 7883152, + "step": 10335 + }, + { + "epoch": 21.496881496881496, + "grad_norm": 0.000450716121122241, + "learning_rate": 0.2532047509752013, + "loss": 0.2614, + "num_input_tokens_seen": 7887056, + "step": 10340 + }, + { + "epoch": 21.507276507276508, + "grad_norm": 0.0002124433813150972, + "learning_rate": 0.25316199689939217, + "loss": 0.2742, + "num_input_tokens_seen": 7891056, + "step": 10345 + }, + { + "epoch": 21.517671517671516, + "grad_norm": 0.0001818554155761376, + "learning_rate": 0.2531192269147068, + "loss": 0.2602, + "num_input_tokens_seen": 7894928, + "step": 10350 + }, + { + "epoch": 21.528066528066528, + "grad_norm": 0.00017955267685465515, + "learning_rate": 0.2530764410277407, + "loss": 0.2612, + "num_input_tokens_seen": 7898768, + "step": 10355 + }, + { + "epoch": 21.53846153846154, + "grad_norm": 0.0003220679936930537, + "learning_rate": 0.25303363924509203, + "loss": 0.2595, + "num_input_tokens_seen": 7902576, + "step": 10360 + }, + { + "epoch": 21.54885654885655, + "grad_norm": 0.0006659325445070863, + "learning_rate": 0.25299082157336145, + "loss": 0.2495, + "num_input_tokens_seen": 7906480, + "step": 10365 + }, + { + "epoch": 21.55925155925156, + "grad_norm": 0.00011034584167646244, + "learning_rate": 0.2529479880191519, + "loss": 0.273, + "num_input_tokens_seen": 7910320, + "step": 10370 + }, + { + "epoch": 21.56964656964657, + "grad_norm": 0.00046427754568867385, + "learning_rate": 0.2529051385890689, + "loss": 0.28, + "num_input_tokens_seen": 7914000, + "step": 10375 + }, + { + "epoch": 21.58004158004158, + "grad_norm": 0.0007128691649995744, + "learning_rate": 0.2528622732897203, + "loss": 0.2708, + "num_input_tokens_seen": 7917840, + "step": 10380 + }, + { + "epoch": 21.59043659043659, + "grad_norm": 0.00014065760478843004, + "learning_rate": 0.25281939212771654, + "loss": 0.2712, + "num_input_tokens_seen": 7921648, + "step": 10385 + }, + { + "epoch": 21.6008316008316, + "grad_norm": 7.154179911594838e-05, + "learning_rate": 0.2527764951096704, + "loss": 0.2709, + "num_input_tokens_seen": 7925456, + "step": 10390 + }, + { + "epoch": 21.611226611226613, + "grad_norm": 0.0003487508511170745, + "learning_rate": 0.2527335822421971, + "loss": 0.2784, + "num_input_tokens_seen": 7929168, + "step": 10395 + }, + { + "epoch": 21.62162162162162, + "grad_norm": 0.0002465861034579575, + "learning_rate": 0.25269065353191444, + "loss": 0.2748, + "num_input_tokens_seen": 7932848, + "step": 10400 + }, + { + "epoch": 21.62162162162162, + "eval_loss": 0.2521704435348511, + "eval_runtime": 13.4291, + "eval_samples_per_second": 63.742, + "eval_steps_per_second": 15.936, + "num_input_tokens_seen": 7932848, + "step": 10400 + }, + { + "epoch": 21.632016632016633, + "grad_norm": 0.0005314700538292527, + "learning_rate": 0.2526477089854425, + "loss": 0.2804, + "num_input_tokens_seen": 7936816, + "step": 10405 + }, + { + "epoch": 21.64241164241164, + "grad_norm": 0.00019947731925640255, + "learning_rate": 0.25260474860940385, + "loss": 0.2572, + "num_input_tokens_seen": 7940656, + "step": 10410 + }, + { + "epoch": 21.652806652806653, + "grad_norm": 0.0004532934108283371, + "learning_rate": 0.2525617724104236, + "loss": 0.295, + "num_input_tokens_seen": 7944592, + "step": 10415 + }, + { + "epoch": 21.66320166320166, + "grad_norm": 0.0006383038125932217, + "learning_rate": 0.25251878039512915, + "loss": 0.2543, + "num_input_tokens_seen": 7948432, + "step": 10420 + }, + { + "epoch": 21.673596673596673, + "grad_norm": 0.0005806440021842718, + "learning_rate": 0.25247577257015047, + "loss": 0.2594, + "num_input_tokens_seen": 7952112, + "step": 10425 + }, + { + "epoch": 21.683991683991685, + "grad_norm": 0.00012706445704679936, + "learning_rate": 0.2524327489421198, + "loss": 0.2646, + "num_input_tokens_seen": 7955920, + "step": 10430 + }, + { + "epoch": 21.694386694386694, + "grad_norm": 0.0002419205120531842, + "learning_rate": 0.25238970951767203, + "loss": 0.252, + "num_input_tokens_seen": 7959664, + "step": 10435 + }, + { + "epoch": 21.704781704781706, + "grad_norm": 0.00037064144271425903, + "learning_rate": 0.25234665430344433, + "loss": 0.2793, + "num_input_tokens_seen": 7963472, + "step": 10440 + }, + { + "epoch": 21.715176715176714, + "grad_norm": 0.0004733344539999962, + "learning_rate": 0.2523035833060764, + "loss": 0.2702, + "num_input_tokens_seen": 7967184, + "step": 10445 + }, + { + "epoch": 21.725571725571726, + "grad_norm": 0.00035818744800053537, + "learning_rate": 0.2522604965322103, + "loss": 0.274, + "num_input_tokens_seen": 7971152, + "step": 10450 + }, + { + "epoch": 21.735966735966738, + "grad_norm": 0.0007519974024035037, + "learning_rate": 0.25221739398849047, + "loss": 0.2924, + "num_input_tokens_seen": 7974896, + "step": 10455 + }, + { + "epoch": 21.746361746361746, + "grad_norm": 0.001055193948559463, + "learning_rate": 0.252174275681564, + "loss": 0.279, + "num_input_tokens_seen": 7978640, + "step": 10460 + }, + { + "epoch": 21.756756756756758, + "grad_norm": 0.0007141061942093074, + "learning_rate": 0.2521311416180802, + "loss": 0.2514, + "num_input_tokens_seen": 7982288, + "step": 10465 + }, + { + "epoch": 21.767151767151766, + "grad_norm": 0.0005833365139551461, + "learning_rate": 0.25208799180469094, + "loss": 0.2757, + "num_input_tokens_seen": 7986128, + "step": 10470 + }, + { + "epoch": 21.777546777546778, + "grad_norm": 0.000380914454581216, + "learning_rate": 0.2520448262480504, + "loss": 0.2948, + "num_input_tokens_seen": 7989968, + "step": 10475 + }, + { + "epoch": 21.787941787941786, + "grad_norm": 0.0007166217546910048, + "learning_rate": 0.25200164495481525, + "loss": 0.2682, + "num_input_tokens_seen": 7993776, + "step": 10480 + }, + { + "epoch": 21.7983367983368, + "grad_norm": 0.00014241099415812641, + "learning_rate": 0.25195844793164474, + "loss": 0.2781, + "num_input_tokens_seen": 7997648, + "step": 10485 + }, + { + "epoch": 21.80873180873181, + "grad_norm": 8.604626054875553e-05, + "learning_rate": 0.2519152351852001, + "loss": 0.2536, + "num_input_tokens_seen": 8001232, + "step": 10490 + }, + { + "epoch": 21.81912681912682, + "grad_norm": 0.0009744454291649163, + "learning_rate": 0.25187200672214555, + "loss": 0.2761, + "num_input_tokens_seen": 8005136, + "step": 10495 + }, + { + "epoch": 21.82952182952183, + "grad_norm": 0.0005549754132516682, + "learning_rate": 0.2518287625491473, + "loss": 0.2852, + "num_input_tokens_seen": 8008912, + "step": 10500 + }, + { + "epoch": 21.83991683991684, + "grad_norm": 0.0008176462142728269, + "learning_rate": 0.25178550267287425, + "loss": 0.2812, + "num_input_tokens_seen": 8012816, + "step": 10505 + }, + { + "epoch": 21.85031185031185, + "grad_norm": 0.00022608082508668303, + "learning_rate": 0.2517422270999976, + "loss": 0.2662, + "num_input_tokens_seen": 8016752, + "step": 10510 + }, + { + "epoch": 21.86070686070686, + "grad_norm": 0.0005152887897565961, + "learning_rate": 0.2516989358371909, + "loss": 0.2868, + "num_input_tokens_seen": 8020656, + "step": 10515 + }, + { + "epoch": 21.87110187110187, + "grad_norm": 0.000178591872099787, + "learning_rate": 0.25165562889113025, + "loss": 0.2719, + "num_input_tokens_seen": 8024304, + "step": 10520 + }, + { + "epoch": 21.881496881496883, + "grad_norm": 0.0011281645856797695, + "learning_rate": 0.2516123062684942, + "loss": 0.2791, + "num_input_tokens_seen": 8028208, + "step": 10525 + }, + { + "epoch": 21.89189189189189, + "grad_norm": 9.733244951348752e-05, + "learning_rate": 0.25156896797596356, + "loss": 0.2685, + "num_input_tokens_seen": 8032112, + "step": 10530 + }, + { + "epoch": 21.902286902286903, + "grad_norm": 0.00015054376854095608, + "learning_rate": 0.2515256140202216, + "loss": 0.2576, + "num_input_tokens_seen": 8035824, + "step": 10535 + }, + { + "epoch": 21.91268191268191, + "grad_norm": 0.000295458099571988, + "learning_rate": 0.25148224440795425, + "loss": 0.2895, + "num_input_tokens_seen": 8039696, + "step": 10540 + }, + { + "epoch": 21.923076923076923, + "grad_norm": 0.0014142782893031836, + "learning_rate": 0.2514388591458494, + "loss": 0.2574, + "num_input_tokens_seen": 8043440, + "step": 10545 + }, + { + "epoch": 21.933471933471935, + "grad_norm": 0.0007092607556842268, + "learning_rate": 0.2513954582405977, + "loss": 0.2599, + "num_input_tokens_seen": 8047248, + "step": 10550 + }, + { + "epoch": 21.943866943866944, + "grad_norm": 0.00017317199672106653, + "learning_rate": 0.2513520416988922, + "loss": 0.2618, + "num_input_tokens_seen": 8051152, + "step": 10555 + }, + { + "epoch": 21.954261954261955, + "grad_norm": 0.0002898134116549045, + "learning_rate": 0.2513086095274281, + "loss": 0.2821, + "num_input_tokens_seen": 8054992, + "step": 10560 + }, + { + "epoch": 21.964656964656964, + "grad_norm": 0.0005926968879066408, + "learning_rate": 0.25126516173290336, + "loss": 0.2752, + "num_input_tokens_seen": 8058832, + "step": 10565 + }, + { + "epoch": 21.975051975051976, + "grad_norm": 9.112121188081801e-05, + "learning_rate": 0.2512216983220181, + "loss": 0.229, + "num_input_tokens_seen": 8062608, + "step": 10570 + }, + { + "epoch": 21.985446985446984, + "grad_norm": 0.0007004551007412374, + "learning_rate": 0.25117821930147494, + "loss": 0.2907, + "num_input_tokens_seen": 8066352, + "step": 10575 + }, + { + "epoch": 21.995841995841996, + "grad_norm": 8.978394907899201e-05, + "learning_rate": 0.2511347246779788, + "loss": 0.2694, + "num_input_tokens_seen": 8070096, + "step": 10580 + }, + { + "epoch": 22.006237006237008, + "grad_norm": 0.0003931598912458867, + "learning_rate": 0.25109121445823723, + "loss": 0.2648, + "num_input_tokens_seen": 8073960, + "step": 10585 + }, + { + "epoch": 22.016632016632016, + "grad_norm": 0.0006780650583095849, + "learning_rate": 0.25104768864896004, + "loss": 0.2781, + "num_input_tokens_seen": 8077736, + "step": 10590 + }, + { + "epoch": 22.027027027027028, + "grad_norm": 0.0006108471425250173, + "learning_rate": 0.2510041472568594, + "loss": 0.2637, + "num_input_tokens_seen": 8081416, + "step": 10595 + }, + { + "epoch": 22.037422037422036, + "grad_norm": 0.00024073536042124033, + "learning_rate": 0.25096059028864987, + "loss": 0.2799, + "num_input_tokens_seen": 8085448, + "step": 10600 + }, + { + "epoch": 22.037422037422036, + "eval_loss": 0.2572762966156006, + "eval_runtime": 13.4248, + "eval_samples_per_second": 63.763, + "eval_steps_per_second": 15.941, + "num_input_tokens_seen": 8085448, + "step": 10600 + }, + { + "epoch": 22.04781704781705, + "grad_norm": 0.0017821715446189046, + "learning_rate": 0.25091701775104863, + "loss": 0.2837, + "num_input_tokens_seen": 8089320, + "step": 10605 + }, + { + "epoch": 22.058212058212057, + "grad_norm": 0.0003902337048202753, + "learning_rate": 0.250873429650775, + "loss": 0.2756, + "num_input_tokens_seen": 8092904, + "step": 10610 + }, + { + "epoch": 22.06860706860707, + "grad_norm": 0.0001144780617323704, + "learning_rate": 0.25082982599455095, + "loss": 0.2529, + "num_input_tokens_seen": 8096680, + "step": 10615 + }, + { + "epoch": 22.07900207900208, + "grad_norm": 0.00048730819253250957, + "learning_rate": 0.2507862067891006, + "loss": 0.2651, + "num_input_tokens_seen": 8100488, + "step": 10620 + }, + { + "epoch": 22.08939708939709, + "grad_norm": 9.896356641547754e-05, + "learning_rate": 0.25074257204115064, + "loss": 0.286, + "num_input_tokens_seen": 8104456, + "step": 10625 + }, + { + "epoch": 22.0997920997921, + "grad_norm": 0.0005414117476902902, + "learning_rate": 0.25069892175742997, + "loss": 0.2675, + "num_input_tokens_seen": 8108168, + "step": 10630 + }, + { + "epoch": 22.11018711018711, + "grad_norm": 0.0002482913841959089, + "learning_rate": 0.25065525594467014, + "loss": 0.2635, + "num_input_tokens_seen": 8111912, + "step": 10635 + }, + { + "epoch": 22.12058212058212, + "grad_norm": 0.00011848725989693776, + "learning_rate": 0.2506115746096049, + "loss": 0.2229, + "num_input_tokens_seen": 8115848, + "step": 10640 + }, + { + "epoch": 22.13097713097713, + "grad_norm": 0.0006396223325282335, + "learning_rate": 0.25056787775897055, + "loss": 0.3073, + "num_input_tokens_seen": 8119592, + "step": 10645 + }, + { + "epoch": 22.14137214137214, + "grad_norm": 0.00043715719948522747, + "learning_rate": 0.2505241653995056, + "loss": 0.2772, + "num_input_tokens_seen": 8123336, + "step": 10650 + }, + { + "epoch": 22.151767151767153, + "grad_norm": 0.0002397741045570001, + "learning_rate": 0.25048043753795113, + "loss": 0.2833, + "num_input_tokens_seen": 8127016, + "step": 10655 + }, + { + "epoch": 22.16216216216216, + "grad_norm": 0.0006872433004900813, + "learning_rate": 0.2504366941810504, + "loss": 0.2638, + "num_input_tokens_seen": 8130824, + "step": 10660 + }, + { + "epoch": 22.172557172557173, + "grad_norm": 0.0006273009348660707, + "learning_rate": 0.2503929353355493, + "loss": 0.2949, + "num_input_tokens_seen": 8134536, + "step": 10665 + }, + { + "epoch": 22.18295218295218, + "grad_norm": 0.0009745433926582336, + "learning_rate": 0.250349161008196, + "loss": 0.2546, + "num_input_tokens_seen": 8138344, + "step": 10670 + }, + { + "epoch": 22.193347193347194, + "grad_norm": 0.00014149340859148651, + "learning_rate": 0.2503053712057409, + "loss": 0.2646, + "num_input_tokens_seen": 8142152, + "step": 10675 + }, + { + "epoch": 22.203742203742205, + "grad_norm": 0.00045154045801609755, + "learning_rate": 0.25026156593493715, + "loss": 0.2546, + "num_input_tokens_seen": 8145960, + "step": 10680 + }, + { + "epoch": 22.214137214137214, + "grad_norm": 9.074409899767488e-05, + "learning_rate": 0.2502177452025399, + "loss": 0.2938, + "num_input_tokens_seen": 8149768, + "step": 10685 + }, + { + "epoch": 22.224532224532226, + "grad_norm": 0.00021942477906122804, + "learning_rate": 0.25017390901530695, + "loss": 0.2605, + "num_input_tokens_seen": 8153544, + "step": 10690 + }, + { + "epoch": 22.234927234927234, + "grad_norm": 0.00031422381289303303, + "learning_rate": 0.2501300573799984, + "loss": 0.2528, + "num_input_tokens_seen": 8157288, + "step": 10695 + }, + { + "epoch": 22.245322245322246, + "grad_norm": 0.0004385676875244826, + "learning_rate": 0.2500861903033766, + "loss": 0.2441, + "num_input_tokens_seen": 8160936, + "step": 10700 + }, + { + "epoch": 22.255717255717254, + "grad_norm": 0.0007493385346606374, + "learning_rate": 0.25004230779220654, + "loss": 0.2647, + "num_input_tokens_seen": 8164776, + "step": 10705 + }, + { + "epoch": 22.266112266112266, + "grad_norm": 0.0003376492240931839, + "learning_rate": 0.24999840985325542, + "loss": 0.2726, + "num_input_tokens_seen": 8168584, + "step": 10710 + }, + { + "epoch": 22.276507276507278, + "grad_norm": 0.0004350741219241172, + "learning_rate": 0.24995449649329285, + "loss": 0.2726, + "num_input_tokens_seen": 8172360, + "step": 10715 + }, + { + "epoch": 22.286902286902286, + "grad_norm": 0.000273155135801062, + "learning_rate": 0.2499105677190908, + "loss": 0.2665, + "num_input_tokens_seen": 8176264, + "step": 10720 + }, + { + "epoch": 22.2972972972973, + "grad_norm": 0.0009184036171063781, + "learning_rate": 0.24986662353742364, + "loss": 0.2738, + "num_input_tokens_seen": 8179944, + "step": 10725 + }, + { + "epoch": 22.307692307692307, + "grad_norm": 0.000325720408000052, + "learning_rate": 0.24982266395506814, + "loss": 0.2811, + "num_input_tokens_seen": 8183752, + "step": 10730 + }, + { + "epoch": 22.31808731808732, + "grad_norm": 0.00035290457890369, + "learning_rate": 0.2497786889788034, + "loss": 0.2788, + "num_input_tokens_seen": 8187720, + "step": 10735 + }, + { + "epoch": 22.328482328482327, + "grad_norm": 0.00017647678032517433, + "learning_rate": 0.24973469861541095, + "loss": 0.28, + "num_input_tokens_seen": 8191624, + "step": 10740 + }, + { + "epoch": 22.33887733887734, + "grad_norm": 0.0009805572917684913, + "learning_rate": 0.24969069287167456, + "loss": 0.2778, + "num_input_tokens_seen": 8195368, + "step": 10745 + }, + { + "epoch": 22.34927234927235, + "grad_norm": 0.0004555070190690458, + "learning_rate": 0.2496466717543806, + "loss": 0.2788, + "num_input_tokens_seen": 8199240, + "step": 10750 + }, + { + "epoch": 22.35966735966736, + "grad_norm": 0.00021970391389913857, + "learning_rate": 0.24960263527031762, + "loss": 0.2623, + "num_input_tokens_seen": 8203048, + "step": 10755 + }, + { + "epoch": 22.37006237006237, + "grad_norm": 0.0008720891200937331, + "learning_rate": 0.24955858342627657, + "loss": 0.2662, + "num_input_tokens_seen": 8206760, + "step": 10760 + }, + { + "epoch": 22.38045738045738, + "grad_norm": 0.00029956205980852246, + "learning_rate": 0.24951451622905083, + "loss": 0.2703, + "num_input_tokens_seen": 8210632, + "step": 10765 + }, + { + "epoch": 22.39085239085239, + "grad_norm": 0.00047952140448614955, + "learning_rate": 0.24947043368543612, + "loss": 0.2759, + "num_input_tokens_seen": 8214472, + "step": 10770 + }, + { + "epoch": 22.401247401247403, + "grad_norm": 0.0010888060787692666, + "learning_rate": 0.2494263358022305, + "loss": 0.2806, + "num_input_tokens_seen": 8218440, + "step": 10775 + }, + { + "epoch": 22.41164241164241, + "grad_norm": 0.00041226629400625825, + "learning_rate": 0.24938222258623444, + "loss": 0.3006, + "num_input_tokens_seen": 8222312, + "step": 10780 + }, + { + "epoch": 22.422037422037423, + "grad_norm": 0.0010342691093683243, + "learning_rate": 0.24933809404425075, + "loss": 0.2554, + "num_input_tokens_seen": 8226152, + "step": 10785 + }, + { + "epoch": 22.43243243243243, + "grad_norm": 0.0007386417128145695, + "learning_rate": 0.24929395018308453, + "loss": 0.2692, + "num_input_tokens_seen": 8230024, + "step": 10790 + }, + { + "epoch": 22.442827442827443, + "grad_norm": 0.00018341877148486674, + "learning_rate": 0.24924979100954348, + "loss": 0.292, + "num_input_tokens_seen": 8233992, + "step": 10795 + }, + { + "epoch": 22.453222453222452, + "grad_norm": 0.0004787735524587333, + "learning_rate": 0.24920561653043735, + "loss": 0.2724, + "num_input_tokens_seen": 8237768, + "step": 10800 + }, + { + "epoch": 22.453222453222452, + "eval_loss": 0.2622421383857727, + "eval_runtime": 13.4181, + "eval_samples_per_second": 63.794, + "eval_steps_per_second": 15.949, + "num_input_tokens_seen": 8237768, + "step": 10800 + }, + { + "epoch": 22.463617463617464, + "grad_norm": 0.0007335016271099448, + "learning_rate": 0.24916142675257846, + "loss": 0.2642, + "num_input_tokens_seen": 8241608, + "step": 10805 + }, + { + "epoch": 22.474012474012476, + "grad_norm": 0.0007495825411751866, + "learning_rate": 0.24911722168278144, + "loss": 0.2505, + "num_input_tokens_seen": 8245512, + "step": 10810 + }, + { + "epoch": 22.484407484407484, + "grad_norm": 0.0004769162624143064, + "learning_rate": 0.24907300132786328, + "loss": 0.2689, + "num_input_tokens_seen": 8249480, + "step": 10815 + }, + { + "epoch": 22.494802494802496, + "grad_norm": 0.0006044330657459795, + "learning_rate": 0.24902876569464322, + "loss": 0.2618, + "num_input_tokens_seen": 8253352, + "step": 10820 + }, + { + "epoch": 22.505197505197504, + "grad_norm": 0.00044620202970691025, + "learning_rate": 0.24898451478994305, + "loss": 0.2788, + "num_input_tokens_seen": 8257096, + "step": 10825 + }, + { + "epoch": 22.515592515592516, + "grad_norm": 0.0002337045007152483, + "learning_rate": 0.2489402486205868, + "loss": 0.2742, + "num_input_tokens_seen": 8261000, + "step": 10830 + }, + { + "epoch": 22.525987525987524, + "grad_norm": 0.00014708917296957225, + "learning_rate": 0.24889596719340085, + "loss": 0.2469, + "num_input_tokens_seen": 8264744, + "step": 10835 + }, + { + "epoch": 22.536382536382536, + "grad_norm": 0.0004472598375286907, + "learning_rate": 0.24885167051521392, + "loss": 0.2702, + "num_input_tokens_seen": 8268648, + "step": 10840 + }, + { + "epoch": 22.546777546777548, + "grad_norm": 0.0010834215208888054, + "learning_rate": 0.24880735859285716, + "loss": 0.2689, + "num_input_tokens_seen": 8272264, + "step": 10845 + }, + { + "epoch": 22.557172557172557, + "grad_norm": 0.0001862489734776318, + "learning_rate": 0.24876303143316406, + "loss": 0.2574, + "num_input_tokens_seen": 8276104, + "step": 10850 + }, + { + "epoch": 22.56756756756757, + "grad_norm": 0.0005557636031880975, + "learning_rate": 0.24871868904297031, + "loss": 0.2988, + "num_input_tokens_seen": 8279912, + "step": 10855 + }, + { + "epoch": 22.577962577962577, + "grad_norm": 0.00017783230578061193, + "learning_rate": 0.24867433142911416, + "loss": 0.2831, + "num_input_tokens_seen": 8283784, + "step": 10860 + }, + { + "epoch": 22.58835758835759, + "grad_norm": 0.0013421468902379274, + "learning_rate": 0.24862995859843612, + "loss": 0.2909, + "num_input_tokens_seen": 8287496, + "step": 10865 + }, + { + "epoch": 22.598752598752597, + "grad_norm": 0.00035398625186644495, + "learning_rate": 0.24858557055777897, + "loss": 0.2671, + "num_input_tokens_seen": 8291272, + "step": 10870 + }, + { + "epoch": 22.60914760914761, + "grad_norm": 0.0004928258713334799, + "learning_rate": 0.24854116731398793, + "loss": 0.2533, + "num_input_tokens_seen": 8294984, + "step": 10875 + }, + { + "epoch": 22.61954261954262, + "grad_norm": 8.907296432880685e-05, + "learning_rate": 0.24849674887391052, + "loss": 0.2273, + "num_input_tokens_seen": 8298760, + "step": 10880 + }, + { + "epoch": 22.62993762993763, + "grad_norm": 0.00015555019490420818, + "learning_rate": 0.2484523152443967, + "loss": 0.2996, + "num_input_tokens_seen": 8302600, + "step": 10885 + }, + { + "epoch": 22.64033264033264, + "grad_norm": 0.0002797841443680227, + "learning_rate": 0.24840786643229862, + "loss": 0.2746, + "num_input_tokens_seen": 8306248, + "step": 10890 + }, + { + "epoch": 22.65072765072765, + "grad_norm": 5.0095448386855423e-05, + "learning_rate": 0.2483634024444709, + "loss": 0.2634, + "num_input_tokens_seen": 8310088, + "step": 10895 + }, + { + "epoch": 22.66112266112266, + "grad_norm": 0.0008946954039856791, + "learning_rate": 0.24831892328777033, + "loss": 0.2885, + "num_input_tokens_seen": 8314056, + "step": 10900 + }, + { + "epoch": 22.671517671517673, + "grad_norm": 0.0001729989453451708, + "learning_rate": 0.2482744289690563, + "loss": 0.2724, + "num_input_tokens_seen": 8317832, + "step": 10905 + }, + { + "epoch": 22.68191268191268, + "grad_norm": 0.0004232304636389017, + "learning_rate": 0.2482299194951903, + "loss": 0.2625, + "num_input_tokens_seen": 8321576, + "step": 10910 + }, + { + "epoch": 22.692307692307693, + "grad_norm": 0.00023504953423980623, + "learning_rate": 0.2481853948730363, + "loss": 0.281, + "num_input_tokens_seen": 8325448, + "step": 10915 + }, + { + "epoch": 22.7027027027027, + "grad_norm": 0.0006898775463923812, + "learning_rate": 0.24814085510946052, + "loss": 0.2712, + "num_input_tokens_seen": 8329192, + "step": 10920 + }, + { + "epoch": 22.713097713097714, + "grad_norm": 0.0007428527460433543, + "learning_rate": 0.24809630021133158, + "loss": 0.2682, + "num_input_tokens_seen": 8333128, + "step": 10925 + }, + { + "epoch": 22.723492723492722, + "grad_norm": 0.0003109718963969499, + "learning_rate": 0.24805173018552037, + "loss": 0.2618, + "num_input_tokens_seen": 8336968, + "step": 10930 + }, + { + "epoch": 22.733887733887734, + "grad_norm": 0.0007977247005328536, + "learning_rate": 0.2480071450389002, + "loss": 0.2667, + "num_input_tokens_seen": 8340712, + "step": 10935 + }, + { + "epoch": 22.744282744282746, + "grad_norm": 0.00029371693381108344, + "learning_rate": 0.24796254477834662, + "loss": 0.273, + "num_input_tokens_seen": 8344424, + "step": 10940 + }, + { + "epoch": 22.754677754677754, + "grad_norm": 0.0003132218844257295, + "learning_rate": 0.24791792941073754, + "loss": 0.2712, + "num_input_tokens_seen": 8348360, + "step": 10945 + }, + { + "epoch": 22.765072765072766, + "grad_norm": 0.0008319218759424984, + "learning_rate": 0.2478732989429533, + "loss": 0.2616, + "num_input_tokens_seen": 8352232, + "step": 10950 + }, + { + "epoch": 22.775467775467774, + "grad_norm": 0.0007291169022209942, + "learning_rate": 0.24782865338187632, + "loss": 0.2632, + "num_input_tokens_seen": 8356136, + "step": 10955 + }, + { + "epoch": 22.785862785862786, + "grad_norm": 0.0003768869210034609, + "learning_rate": 0.2477839927343916, + "loss": 0.3096, + "num_input_tokens_seen": 8359912, + "step": 10960 + }, + { + "epoch": 22.796257796257795, + "grad_norm": 0.0004930169088765979, + "learning_rate": 0.2477393170073864, + "loss": 0.2733, + "num_input_tokens_seen": 8363656, + "step": 10965 + }, + { + "epoch": 22.806652806652806, + "grad_norm": 0.0010685250163078308, + "learning_rate": 0.2476946262077503, + "loss": 0.272, + "num_input_tokens_seen": 8367592, + "step": 10970 + }, + { + "epoch": 22.81704781704782, + "grad_norm": 0.0009760446846485138, + "learning_rate": 0.24764992034237507, + "loss": 0.2965, + "num_input_tokens_seen": 8371464, + "step": 10975 + }, + { + "epoch": 22.827442827442827, + "grad_norm": 0.00044547097058966756, + "learning_rate": 0.24760519941815498, + "loss": 0.2619, + "num_input_tokens_seen": 8375208, + "step": 10980 + }, + { + "epoch": 22.83783783783784, + "grad_norm": 0.0008346649119630456, + "learning_rate": 0.2475604634419866, + "loss": 0.2865, + "num_input_tokens_seen": 8379080, + "step": 10985 + }, + { + "epoch": 22.848232848232847, + "grad_norm": 0.0001985155831789598, + "learning_rate": 0.24751571242076872, + "loss": 0.2763, + "num_input_tokens_seen": 8382856, + "step": 10990 + }, + { + "epoch": 22.85862785862786, + "grad_norm": 0.00023423205129802227, + "learning_rate": 0.2474709463614025, + "loss": 0.2828, + "num_input_tokens_seen": 8386920, + "step": 10995 + }, + { + "epoch": 22.86902286902287, + "grad_norm": 0.00041810987750068307, + "learning_rate": 0.24742616527079145, + "loss": 0.2279, + "num_input_tokens_seen": 8390664, + "step": 11000 + }, + { + "epoch": 22.86902286902287, + "eval_loss": 0.2504565417766571, + "eval_runtime": 13.4057, + "eval_samples_per_second": 63.854, + "eval_steps_per_second": 15.963, + "num_input_tokens_seen": 8390664, + "step": 11000 + }, + { + "epoch": 22.87941787941788, + "grad_norm": 0.0005937424139119685, + "learning_rate": 0.24738136915584139, + "loss": 0.3025, + "num_input_tokens_seen": 8394504, + "step": 11005 + }, + { + "epoch": 22.88981288981289, + "grad_norm": 0.000998524483293295, + "learning_rate": 0.24733655802346047, + "loss": 0.2674, + "num_input_tokens_seen": 8398312, + "step": 11010 + }, + { + "epoch": 22.9002079002079, + "grad_norm": 0.00020585885795298964, + "learning_rate": 0.24729173188055906, + "loss": 0.2725, + "num_input_tokens_seen": 8402088, + "step": 11015 + }, + { + "epoch": 22.91060291060291, + "grad_norm": 0.0009814698714762926, + "learning_rate": 0.24724689073404996, + "loss": 0.2724, + "num_input_tokens_seen": 8405928, + "step": 11020 + }, + { + "epoch": 22.92099792099792, + "grad_norm": 0.0003191218711435795, + "learning_rate": 0.24720203459084822, + "loss": 0.2721, + "num_input_tokens_seen": 8409576, + "step": 11025 + }, + { + "epoch": 22.93139293139293, + "grad_norm": 0.00011392906890250742, + "learning_rate": 0.24715716345787123, + "loss": 0.2535, + "num_input_tokens_seen": 8413384, + "step": 11030 + }, + { + "epoch": 22.941787941787943, + "grad_norm": 0.00046227345592342317, + "learning_rate": 0.2471122773420387, + "loss": 0.2346, + "num_input_tokens_seen": 8417256, + "step": 11035 + }, + { + "epoch": 22.95218295218295, + "grad_norm": 0.00015766163414809853, + "learning_rate": 0.24706737625027259, + "loss": 0.2415, + "num_input_tokens_seen": 8421096, + "step": 11040 + }, + { + "epoch": 22.962577962577964, + "grad_norm": 0.00012834918743465096, + "learning_rate": 0.24702246018949725, + "loss": 0.2935, + "num_input_tokens_seen": 8424872, + "step": 11045 + }, + { + "epoch": 22.972972972972972, + "grad_norm": 0.00014894672494847327, + "learning_rate": 0.2469775291666393, + "loss": 0.2685, + "num_input_tokens_seen": 8428648, + "step": 11050 + }, + { + "epoch": 22.983367983367984, + "grad_norm": 0.0003273191978223622, + "learning_rate": 0.24693258318862765, + "loss": 0.2626, + "num_input_tokens_seen": 8432616, + "step": 11055 + }, + { + "epoch": 22.993762993762992, + "grad_norm": 0.0008077337988652289, + "learning_rate": 0.2468876222623935, + "loss": 0.265, + "num_input_tokens_seen": 8436616, + "step": 11060 + }, + { + "epoch": 23.004158004158004, + "grad_norm": 0.0002598306746222079, + "learning_rate": 0.2468426463948705, + "loss": 0.3022, + "num_input_tokens_seen": 8440208, + "step": 11065 + }, + { + "epoch": 23.014553014553016, + "grad_norm": 0.000348987290635705, + "learning_rate": 0.24679765559299438, + "loss": 0.2744, + "num_input_tokens_seen": 8444112, + "step": 11070 + }, + { + "epoch": 23.024948024948024, + "grad_norm": 0.000650043657515198, + "learning_rate": 0.24675264986370332, + "loss": 0.2776, + "num_input_tokens_seen": 8447920, + "step": 11075 + }, + { + "epoch": 23.035343035343036, + "grad_norm": 0.00030308536952361465, + "learning_rate": 0.2467076292139378, + "loss": 0.2516, + "num_input_tokens_seen": 8451792, + "step": 11080 + }, + { + "epoch": 23.045738045738045, + "grad_norm": 0.00021021509019192308, + "learning_rate": 0.24666259365064055, + "loss": 0.2305, + "num_input_tokens_seen": 8455504, + "step": 11085 + }, + { + "epoch": 23.056133056133056, + "grad_norm": 0.00029759813332930207, + "learning_rate": 0.24661754318075663, + "loss": 0.2575, + "num_input_tokens_seen": 8459408, + "step": 11090 + }, + { + "epoch": 23.066528066528065, + "grad_norm": 0.000461197312688455, + "learning_rate": 0.2465724778112334, + "loss": 0.2767, + "num_input_tokens_seen": 8463280, + "step": 11095 + }, + { + "epoch": 23.076923076923077, + "grad_norm": 0.0005435289931483567, + "learning_rate": 0.24652739754902042, + "loss": 0.2807, + "num_input_tokens_seen": 8467120, + "step": 11100 + }, + { + "epoch": 23.08731808731809, + "grad_norm": 0.0008591612568125129, + "learning_rate": 0.24648230240106975, + "loss": 0.2692, + "num_input_tokens_seen": 8470960, + "step": 11105 + }, + { + "epoch": 23.097713097713097, + "grad_norm": 0.0005267381202429533, + "learning_rate": 0.2464371923743356, + "loss": 0.2306, + "num_input_tokens_seen": 8474736, + "step": 11110 + }, + { + "epoch": 23.10810810810811, + "grad_norm": 0.00031147280242294073, + "learning_rate": 0.24639206747577444, + "loss": 0.262, + "num_input_tokens_seen": 8478672, + "step": 11115 + }, + { + "epoch": 23.118503118503117, + "grad_norm": 0.0003172263677697629, + "learning_rate": 0.24634692771234515, + "loss": 0.2708, + "num_input_tokens_seen": 8482608, + "step": 11120 + }, + { + "epoch": 23.12889812889813, + "grad_norm": 0.0001414748840034008, + "learning_rate": 0.2463017730910088, + "loss": 0.2639, + "num_input_tokens_seen": 8486480, + "step": 11125 + }, + { + "epoch": 23.13929313929314, + "grad_norm": 0.000844207766931504, + "learning_rate": 0.2462566036187289, + "loss": 0.2635, + "num_input_tokens_seen": 8490224, + "step": 11130 + }, + { + "epoch": 23.14968814968815, + "grad_norm": 0.0010493883164599538, + "learning_rate": 0.24621141930247106, + "loss": 0.2633, + "num_input_tokens_seen": 8493968, + "step": 11135 + }, + { + "epoch": 23.16008316008316, + "grad_norm": 0.00019993902242276818, + "learning_rate": 0.2461662201492033, + "loss": 0.2601, + "num_input_tokens_seen": 8497712, + "step": 11140 + }, + { + "epoch": 23.17047817047817, + "grad_norm": 0.00027402755222283304, + "learning_rate": 0.24612100616589586, + "loss": 0.2663, + "num_input_tokens_seen": 8501328, + "step": 11145 + }, + { + "epoch": 23.18087318087318, + "grad_norm": 0.0005629007937386632, + "learning_rate": 0.24607577735952135, + "loss": 0.2785, + "num_input_tokens_seen": 8505136, + "step": 11150 + }, + { + "epoch": 23.19126819126819, + "grad_norm": 0.0006390362395904958, + "learning_rate": 0.24603053373705464, + "loss": 0.2748, + "num_input_tokens_seen": 8508912, + "step": 11155 + }, + { + "epoch": 23.2016632016632, + "grad_norm": 0.0005824014660902321, + "learning_rate": 0.2459852753054728, + "loss": 0.2833, + "num_input_tokens_seen": 8512784, + "step": 11160 + }, + { + "epoch": 23.212058212058214, + "grad_norm": 0.0002615506818983704, + "learning_rate": 0.24594000207175526, + "loss": 0.2548, + "num_input_tokens_seen": 8516560, + "step": 11165 + }, + { + "epoch": 23.222453222453222, + "grad_norm": 0.00040190783329308033, + "learning_rate": 0.2458947140428838, + "loss": 0.2687, + "num_input_tokens_seen": 8520400, + "step": 11170 + }, + { + "epoch": 23.232848232848234, + "grad_norm": 0.0004516924964264035, + "learning_rate": 0.24584941122584233, + "loss": 0.2866, + "num_input_tokens_seen": 8524208, + "step": 11175 + }, + { + "epoch": 23.243243243243242, + "grad_norm": 0.00045322030200622976, + "learning_rate": 0.24580409362761713, + "loss": 0.2819, + "num_input_tokens_seen": 8528048, + "step": 11180 + }, + { + "epoch": 23.253638253638254, + "grad_norm": 9.581913036527112e-05, + "learning_rate": 0.2457587612551967, + "loss": 0.2712, + "num_input_tokens_seen": 8531920, + "step": 11185 + }, + { + "epoch": 23.264033264033262, + "grad_norm": 0.0005042356206104159, + "learning_rate": 0.24571341411557193, + "loss": 0.2512, + "num_input_tokens_seen": 8535696, + "step": 11190 + }, + { + "epoch": 23.274428274428274, + "grad_norm": 0.00011410010483814403, + "learning_rate": 0.2456680522157359, + "loss": 0.2696, + "num_input_tokens_seen": 8539568, + "step": 11195 + }, + { + "epoch": 23.284823284823286, + "grad_norm": 0.0006625992245972157, + "learning_rate": 0.245622675562684, + "loss": 0.266, + "num_input_tokens_seen": 8543280, + "step": 11200 + }, + { + "epoch": 23.284823284823286, + "eval_loss": 0.2535899877548218, + "eval_runtime": 13.4784, + "eval_samples_per_second": 63.509, + "eval_steps_per_second": 15.877, + "num_input_tokens_seen": 8543280, + "step": 11200 + }, + { + "epoch": 23.295218295218294, + "grad_norm": 0.0005809700815007091, + "learning_rate": 0.24557728416341384, + "loss": 0.2772, + "num_input_tokens_seen": 8547184, + "step": 11205 + }, + { + "epoch": 23.305613305613306, + "grad_norm": 0.0003124606446363032, + "learning_rate": 0.24553187802492538, + "loss": 0.2534, + "num_input_tokens_seen": 8551024, + "step": 11210 + }, + { + "epoch": 23.316008316008315, + "grad_norm": 0.0001347609650110826, + "learning_rate": 0.24548645715422074, + "loss": 0.2506, + "num_input_tokens_seen": 8554864, + "step": 11215 + }, + { + "epoch": 23.326403326403327, + "grad_norm": 0.00015566272486466914, + "learning_rate": 0.2454410215583045, + "loss": 0.2262, + "num_input_tokens_seen": 8558704, + "step": 11220 + }, + { + "epoch": 23.33679833679834, + "grad_norm": 0.000481221271911636, + "learning_rate": 0.24539557124418332, + "loss": 0.2455, + "num_input_tokens_seen": 8562448, + "step": 11225 + }, + { + "epoch": 23.347193347193347, + "grad_norm": 0.0002525653981138021, + "learning_rate": 0.24535010621886624, + "loss": 0.2774, + "num_input_tokens_seen": 8566256, + "step": 11230 + }, + { + "epoch": 23.35758835758836, + "grad_norm": 0.0013254560763016343, + "learning_rate": 0.2453046264893646, + "loss": 0.2895, + "num_input_tokens_seen": 8569968, + "step": 11235 + }, + { + "epoch": 23.367983367983367, + "grad_norm": 0.0009029508219100535, + "learning_rate": 0.24525913206269184, + "loss": 0.2852, + "num_input_tokens_seen": 8574000, + "step": 11240 + }, + { + "epoch": 23.37837837837838, + "grad_norm": 0.0008508173050358891, + "learning_rate": 0.2452136229458638, + "loss": 0.2937, + "num_input_tokens_seen": 8577840, + "step": 11245 + }, + { + "epoch": 23.388773388773387, + "grad_norm": 0.0013674009824171662, + "learning_rate": 0.24516809914589857, + "loss": 0.2486, + "num_input_tokens_seen": 8581648, + "step": 11250 + }, + { + "epoch": 23.3991683991684, + "grad_norm": 0.0006002847221679986, + "learning_rate": 0.2451225606698165, + "loss": 0.2697, + "num_input_tokens_seen": 8585680, + "step": 11255 + }, + { + "epoch": 23.40956340956341, + "grad_norm": 0.000300979329040274, + "learning_rate": 0.2450770075246402, + "loss": 0.2789, + "num_input_tokens_seen": 8589392, + "step": 11260 + }, + { + "epoch": 23.41995841995842, + "grad_norm": 0.0005880179814994335, + "learning_rate": 0.24503143971739455, + "loss": 0.2747, + "num_input_tokens_seen": 8593328, + "step": 11265 + }, + { + "epoch": 23.43035343035343, + "grad_norm": 0.0005758588085882366, + "learning_rate": 0.24498585725510663, + "loss": 0.2753, + "num_input_tokens_seen": 8597296, + "step": 11270 + }, + { + "epoch": 23.44074844074844, + "grad_norm": 0.00017160146671812981, + "learning_rate": 0.24494026014480583, + "loss": 0.2649, + "num_input_tokens_seen": 8601264, + "step": 11275 + }, + { + "epoch": 23.45114345114345, + "grad_norm": 0.0007787612266838551, + "learning_rate": 0.24489464839352387, + "loss": 0.2883, + "num_input_tokens_seen": 8605040, + "step": 11280 + }, + { + "epoch": 23.46153846153846, + "grad_norm": 0.0004692948132287711, + "learning_rate": 0.2448490220082946, + "loss": 0.2557, + "num_input_tokens_seen": 8608752, + "step": 11285 + }, + { + "epoch": 23.471933471933472, + "grad_norm": 0.00012573765707202256, + "learning_rate": 0.24480338099615415, + "loss": 0.2581, + "num_input_tokens_seen": 8612464, + "step": 11290 + }, + { + "epoch": 23.482328482328484, + "grad_norm": 0.000173399384948425, + "learning_rate": 0.244757725364141, + "loss": 0.2545, + "num_input_tokens_seen": 8616240, + "step": 11295 + }, + { + "epoch": 23.492723492723492, + "grad_norm": 0.0003453368553891778, + "learning_rate": 0.24471205511929583, + "loss": 0.2901, + "num_input_tokens_seen": 8620048, + "step": 11300 + }, + { + "epoch": 23.503118503118504, + "grad_norm": 0.00042090981150977314, + "learning_rate": 0.24466637026866145, + "loss": 0.2741, + "num_input_tokens_seen": 8623824, + "step": 11305 + }, + { + "epoch": 23.513513513513512, + "grad_norm": 0.00013157064677216113, + "learning_rate": 0.2446206708192832, + "loss": 0.2674, + "num_input_tokens_seen": 8627664, + "step": 11310 + }, + { + "epoch": 23.523908523908524, + "grad_norm": 0.00015192497812677175, + "learning_rate": 0.2445749567782084, + "loss": 0.2688, + "num_input_tokens_seen": 8631504, + "step": 11315 + }, + { + "epoch": 23.534303534303533, + "grad_norm": 0.0004945768741890788, + "learning_rate": 0.2445292281524868, + "loss": 0.2697, + "num_input_tokens_seen": 8635472, + "step": 11320 + }, + { + "epoch": 23.544698544698544, + "grad_norm": 0.00030011392664164305, + "learning_rate": 0.24448348494917022, + "loss": 0.2665, + "num_input_tokens_seen": 8639376, + "step": 11325 + }, + { + "epoch": 23.555093555093556, + "grad_norm": 0.00026054843328893185, + "learning_rate": 0.24443772717531295, + "loss": 0.2785, + "num_input_tokens_seen": 8643120, + "step": 11330 + }, + { + "epoch": 23.565488565488565, + "grad_norm": 0.00013831326214130968, + "learning_rate": 0.24439195483797138, + "loss": 0.2847, + "num_input_tokens_seen": 8646896, + "step": 11335 + }, + { + "epoch": 23.575883575883577, + "grad_norm": 0.0006735522183589637, + "learning_rate": 0.24434616794420416, + "loss": 0.2752, + "num_input_tokens_seen": 8650768, + "step": 11340 + }, + { + "epoch": 23.586278586278585, + "grad_norm": 0.0002660851750988513, + "learning_rate": 0.24430036650107223, + "loss": 0.2653, + "num_input_tokens_seen": 8654576, + "step": 11345 + }, + { + "epoch": 23.596673596673597, + "grad_norm": 0.0005186255439184606, + "learning_rate": 0.2442545505156387, + "loss": 0.2764, + "num_input_tokens_seen": 8658448, + "step": 11350 + }, + { + "epoch": 23.60706860706861, + "grad_norm": 0.0005912862252444029, + "learning_rate": 0.24420871999496904, + "loss": 0.2757, + "num_input_tokens_seen": 8662352, + "step": 11355 + }, + { + "epoch": 23.617463617463617, + "grad_norm": 0.0006086938083171844, + "learning_rate": 0.24416287494613084, + "loss": 0.2787, + "num_input_tokens_seen": 8666192, + "step": 11360 + }, + { + "epoch": 23.62785862785863, + "grad_norm": 0.00010927175753749907, + "learning_rate": 0.24411701537619399, + "loss": 0.2588, + "num_input_tokens_seen": 8670032, + "step": 11365 + }, + { + "epoch": 23.638253638253637, + "grad_norm": 0.0005654322449117899, + "learning_rate": 0.24407114129223062, + "loss": 0.2489, + "num_input_tokens_seen": 8673840, + "step": 11370 + }, + { + "epoch": 23.64864864864865, + "grad_norm": 7.375364657491446e-05, + "learning_rate": 0.2440252527013151, + "loss": 0.2614, + "num_input_tokens_seen": 8677456, + "step": 11375 + }, + { + "epoch": 23.659043659043657, + "grad_norm": 3.828847184195183e-05, + "learning_rate": 0.24397934961052403, + "loss": 0.2649, + "num_input_tokens_seen": 8681136, + "step": 11380 + }, + { + "epoch": 23.66943866943867, + "grad_norm": 0.0005979792331345379, + "learning_rate": 0.24393343202693618, + "loss": 0.235, + "num_input_tokens_seen": 8684912, + "step": 11385 + }, + { + "epoch": 23.67983367983368, + "grad_norm": 0.00038252881495282054, + "learning_rate": 0.2438874999576327, + "loss": 0.2851, + "num_input_tokens_seen": 8688816, + "step": 11390 + }, + { + "epoch": 23.69022869022869, + "grad_norm": 8.783923840383068e-05, + "learning_rate": 0.24384155340969688, + "loss": 0.2697, + "num_input_tokens_seen": 8692720, + "step": 11395 + }, + { + "epoch": 23.7006237006237, + "grad_norm": 0.0006030662334524095, + "learning_rate": 0.24379559239021423, + "loss": 0.2712, + "num_input_tokens_seen": 8696432, + "step": 11400 + }, + { + "epoch": 23.7006237006237, + "eval_loss": 0.25823086500167847, + "eval_runtime": 13.4282, + "eval_samples_per_second": 63.747, + "eval_steps_per_second": 15.937, + "num_input_tokens_seen": 8696432, + "step": 11400 + }, + { + "epoch": 23.71101871101871, + "grad_norm": 0.00014482211554422975, + "learning_rate": 0.2437496169062725, + "loss": 0.2818, + "num_input_tokens_seen": 8700272, + "step": 11405 + }, + { + "epoch": 23.72141372141372, + "grad_norm": 0.0003956988512072712, + "learning_rate": 0.24370362696496176, + "loss": 0.2845, + "num_input_tokens_seen": 8704112, + "step": 11410 + }, + { + "epoch": 23.731808731808734, + "grad_norm": 0.00016388812218792737, + "learning_rate": 0.24365762257337417, + "loss": 0.2573, + "num_input_tokens_seen": 8707952, + "step": 11415 + }, + { + "epoch": 23.742203742203742, + "grad_norm": 0.001138166873715818, + "learning_rate": 0.2436116037386042, + "loss": 0.3075, + "num_input_tokens_seen": 8711696, + "step": 11420 + }, + { + "epoch": 23.752598752598754, + "grad_norm": 0.0001774019910953939, + "learning_rate": 0.24356557046774852, + "loss": 0.2921, + "num_input_tokens_seen": 8715728, + "step": 11425 + }, + { + "epoch": 23.762993762993762, + "grad_norm": 0.0004788069927599281, + "learning_rate": 0.24351952276790606, + "loss": 0.295, + "num_input_tokens_seen": 8719536, + "step": 11430 + }, + { + "epoch": 23.773388773388774, + "grad_norm": 0.0005475771613419056, + "learning_rate": 0.24347346064617797, + "loss": 0.275, + "num_input_tokens_seen": 8723344, + "step": 11435 + }, + { + "epoch": 23.783783783783782, + "grad_norm": 0.00013213448983151466, + "learning_rate": 0.24342738410966758, + "loss": 0.2439, + "num_input_tokens_seen": 8727184, + "step": 11440 + }, + { + "epoch": 23.794178794178794, + "grad_norm": 0.00032229532371275127, + "learning_rate": 0.24338129316548046, + "loss": 0.251, + "num_input_tokens_seen": 8730992, + "step": 11445 + }, + { + "epoch": 23.804573804573806, + "grad_norm": 0.0009105876088142395, + "learning_rate": 0.24333518782072444, + "loss": 0.2757, + "num_input_tokens_seen": 8734704, + "step": 11450 + }, + { + "epoch": 23.814968814968815, + "grad_norm": 0.0008148581255227327, + "learning_rate": 0.24328906808250952, + "loss": 0.2849, + "num_input_tokens_seen": 8738512, + "step": 11455 + }, + { + "epoch": 23.825363825363826, + "grad_norm": 0.0008986066095530987, + "learning_rate": 0.243242933957948, + "loss": 0.2796, + "num_input_tokens_seen": 8742256, + "step": 11460 + }, + { + "epoch": 23.835758835758835, + "grad_norm": 0.0007245682063512504, + "learning_rate": 0.24319678545415427, + "loss": 0.297, + "num_input_tokens_seen": 8746064, + "step": 11465 + }, + { + "epoch": 23.846153846153847, + "grad_norm": 0.0003781277628149837, + "learning_rate": 0.24315062257824507, + "loss": 0.275, + "num_input_tokens_seen": 8749840, + "step": 11470 + }, + { + "epoch": 23.856548856548855, + "grad_norm": 8.64934190758504e-05, + "learning_rate": 0.24310444533733921, + "loss": 0.2807, + "num_input_tokens_seen": 8753616, + "step": 11475 + }, + { + "epoch": 23.866943866943867, + "grad_norm": 0.00019423867342993617, + "learning_rate": 0.2430582537385579, + "loss": 0.2953, + "num_input_tokens_seen": 8757264, + "step": 11480 + }, + { + "epoch": 23.87733887733888, + "grad_norm": 0.0004998295335099101, + "learning_rate": 0.2430120477890244, + "loss": 0.2753, + "num_input_tokens_seen": 8761104, + "step": 11485 + }, + { + "epoch": 23.887733887733887, + "grad_norm": 0.00035150619805790484, + "learning_rate": 0.24296582749586426, + "loss": 0.2569, + "num_input_tokens_seen": 8764848, + "step": 11490 + }, + { + "epoch": 23.8981288981289, + "grad_norm": 0.00010583133553154767, + "learning_rate": 0.24291959286620526, + "loss": 0.2734, + "num_input_tokens_seen": 8768624, + "step": 11495 + }, + { + "epoch": 23.908523908523907, + "grad_norm": 0.00037805380998179317, + "learning_rate": 0.24287334390717738, + "loss": 0.241, + "num_input_tokens_seen": 8772464, + "step": 11500 + }, + { + "epoch": 23.91891891891892, + "grad_norm": 0.00040706092840991914, + "learning_rate": 0.24282708062591268, + "loss": 0.2647, + "num_input_tokens_seen": 8776368, + "step": 11505 + }, + { + "epoch": 23.929313929313928, + "grad_norm": 8.585912291891873e-05, + "learning_rate": 0.24278080302954563, + "loss": 0.2689, + "num_input_tokens_seen": 8780208, + "step": 11510 + }, + { + "epoch": 23.93970893970894, + "grad_norm": 0.0005100580747239292, + "learning_rate": 0.24273451112521283, + "loss": 0.2674, + "num_input_tokens_seen": 8783984, + "step": 11515 + }, + { + "epoch": 23.95010395010395, + "grad_norm": 0.00020650606893468648, + "learning_rate": 0.242688204920053, + "loss": 0.2573, + "num_input_tokens_seen": 8787856, + "step": 11520 + }, + { + "epoch": 23.96049896049896, + "grad_norm": 0.0003401414433028549, + "learning_rate": 0.24264188442120715, + "loss": 0.2629, + "num_input_tokens_seen": 8791600, + "step": 11525 + }, + { + "epoch": 23.97089397089397, + "grad_norm": 0.0001000309202936478, + "learning_rate": 0.24259554963581853, + "loss": 0.2295, + "num_input_tokens_seen": 8795312, + "step": 11530 + }, + { + "epoch": 23.98128898128898, + "grad_norm": 0.000363660859875381, + "learning_rate": 0.24254920057103257, + "loss": 0.273, + "num_input_tokens_seen": 8799184, + "step": 11535 + }, + { + "epoch": 23.991683991683992, + "grad_norm": 0.00042738919728435576, + "learning_rate": 0.24250283723399685, + "loss": 0.2189, + "num_input_tokens_seen": 8802960, + "step": 11540 + }, + { + "epoch": 24.002079002079004, + "grad_norm": 0.0002777199260890484, + "learning_rate": 0.24245645963186108, + "loss": 0.2775, + "num_input_tokens_seen": 8806816, + "step": 11545 + }, + { + "epoch": 24.012474012474012, + "grad_norm": 0.0008512705680914223, + "learning_rate": 0.2424100677717774, + "loss": 0.3022, + "num_input_tokens_seen": 8810720, + "step": 11550 + }, + { + "epoch": 24.022869022869024, + "grad_norm": 0.00034911895636469126, + "learning_rate": 0.24236366166090004, + "loss": 0.282, + "num_input_tokens_seen": 8814816, + "step": 11555 + }, + { + "epoch": 24.033264033264032, + "grad_norm": 0.0013787307543680072, + "learning_rate": 0.24231724130638527, + "loss": 0.2805, + "num_input_tokens_seen": 8818688, + "step": 11560 + }, + { + "epoch": 24.043659043659044, + "grad_norm": 0.0007107564015313983, + "learning_rate": 0.2422708067153917, + "loss": 0.2755, + "num_input_tokens_seen": 8822528, + "step": 11565 + }, + { + "epoch": 24.054054054054053, + "grad_norm": 0.0005409236764535308, + "learning_rate": 0.24222435789508026, + "loss": 0.2691, + "num_input_tokens_seen": 8826528, + "step": 11570 + }, + { + "epoch": 24.064449064449065, + "grad_norm": 0.0004387928347568959, + "learning_rate": 0.24217789485261387, + "loss": 0.2698, + "num_input_tokens_seen": 8830368, + "step": 11575 + }, + { + "epoch": 24.074844074844076, + "grad_norm": 0.0003292463661637157, + "learning_rate": 0.2421314175951577, + "loss": 0.2717, + "num_input_tokens_seen": 8834176, + "step": 11580 + }, + { + "epoch": 24.085239085239085, + "grad_norm": 0.00030610713292844594, + "learning_rate": 0.2420849261298791, + "loss": 0.2943, + "num_input_tokens_seen": 8838144, + "step": 11585 + }, + { + "epoch": 24.095634095634097, + "grad_norm": 0.0003340231196489185, + "learning_rate": 0.24203842046394775, + "loss": 0.2839, + "num_input_tokens_seen": 8841952, + "step": 11590 + }, + { + "epoch": 24.106029106029105, + "grad_norm": 0.0007471473072655499, + "learning_rate": 0.24199190060453535, + "loss": 0.2716, + "num_input_tokens_seen": 8845632, + "step": 11595 + }, + { + "epoch": 24.116424116424117, + "grad_norm": 7.737662235740572e-05, + "learning_rate": 0.2419453665588158, + "loss": 0.2664, + "num_input_tokens_seen": 8849408, + "step": 11600 + }, + { + "epoch": 24.116424116424117, + "eval_loss": 0.24997226893901825, + "eval_runtime": 13.4528, + "eval_samples_per_second": 63.63, + "eval_steps_per_second": 15.907, + "num_input_tokens_seen": 8849408, + "step": 11600 + }, + { + "epoch": 24.126819126819125, + "grad_norm": 0.00013961437798570842, + "learning_rate": 0.24189881833396523, + "loss": 0.259, + "num_input_tokens_seen": 8853184, + "step": 11605 + }, + { + "epoch": 24.137214137214137, + "grad_norm": 0.00033145881025120616, + "learning_rate": 0.24185225593716203, + "loss": 0.2518, + "num_input_tokens_seen": 8856896, + "step": 11610 + }, + { + "epoch": 24.14760914760915, + "grad_norm": 0.00014620805450249463, + "learning_rate": 0.2418056793755867, + "loss": 0.2595, + "num_input_tokens_seen": 8860672, + "step": 11615 + }, + { + "epoch": 24.158004158004157, + "grad_norm": 0.0008421497186645865, + "learning_rate": 0.24175908865642187, + "loss": 0.2813, + "num_input_tokens_seen": 8864544, + "step": 11620 + }, + { + "epoch": 24.16839916839917, + "grad_norm": 0.0003480222949292511, + "learning_rate": 0.24171248378685248, + "loss": 0.2805, + "num_input_tokens_seen": 8868352, + "step": 11625 + }, + { + "epoch": 24.178794178794178, + "grad_norm": 0.0004227267927490175, + "learning_rate": 0.24166586477406554, + "loss": 0.2793, + "num_input_tokens_seen": 8872128, + "step": 11630 + }, + { + "epoch": 24.18918918918919, + "grad_norm": 5.2247785788495094e-05, + "learning_rate": 0.24161923162525034, + "loss": 0.2808, + "num_input_tokens_seen": 8875840, + "step": 11635 + }, + { + "epoch": 24.1995841995842, + "grad_norm": 0.000608329544775188, + "learning_rate": 0.2415725843475982, + "loss": 0.2802, + "num_input_tokens_seen": 8879648, + "step": 11640 + }, + { + "epoch": 24.20997920997921, + "grad_norm": 3.049782208108809e-05, + "learning_rate": 0.24152592294830286, + "loss": 0.2641, + "num_input_tokens_seen": 8883648, + "step": 11645 + }, + { + "epoch": 24.22037422037422, + "grad_norm": 0.0006523304618895054, + "learning_rate": 0.24147924743455995, + "loss": 0.2821, + "num_input_tokens_seen": 8887552, + "step": 11650 + }, + { + "epoch": 24.23076923076923, + "grad_norm": 8.20455388748087e-05, + "learning_rate": 0.24143255781356754, + "loss": 0.2531, + "num_input_tokens_seen": 8891424, + "step": 11655 + }, + { + "epoch": 24.241164241164242, + "grad_norm": 0.0001101844827644527, + "learning_rate": 0.24138585409252566, + "loss": 0.2807, + "num_input_tokens_seen": 8895264, + "step": 11660 + }, + { + "epoch": 24.25155925155925, + "grad_norm": 0.0004913162556476891, + "learning_rate": 0.24133913627863662, + "loss": 0.2643, + "num_input_tokens_seen": 8898944, + "step": 11665 + }, + { + "epoch": 24.261954261954262, + "grad_norm": 0.0002816529886331409, + "learning_rate": 0.241292404379105, + "loss": 0.2619, + "num_input_tokens_seen": 8902848, + "step": 11670 + }, + { + "epoch": 24.272349272349274, + "grad_norm": 0.00026957373484037817, + "learning_rate": 0.24124565840113735, + "loss": 0.2499, + "num_input_tokens_seen": 8906624, + "step": 11675 + }, + { + "epoch": 24.282744282744282, + "grad_norm": 0.00021040673891548067, + "learning_rate": 0.2411988983519425, + "loss": 0.2637, + "num_input_tokens_seen": 8910432, + "step": 11680 + }, + { + "epoch": 24.293139293139294, + "grad_norm": 0.00018961050955113024, + "learning_rate": 0.24115212423873145, + "loss": 0.2671, + "num_input_tokens_seen": 8914208, + "step": 11685 + }, + { + "epoch": 24.303534303534303, + "grad_norm": 0.00011383459786884487, + "learning_rate": 0.24110533606871737, + "loss": 0.2785, + "num_input_tokens_seen": 8918112, + "step": 11690 + }, + { + "epoch": 24.313929313929314, + "grad_norm": 0.00026835172320716083, + "learning_rate": 0.24105853384911552, + "loss": 0.2727, + "num_input_tokens_seen": 8921920, + "step": 11695 + }, + { + "epoch": 24.324324324324323, + "grad_norm": 0.00013970643340144306, + "learning_rate": 0.24101171758714346, + "loss": 0.2615, + "num_input_tokens_seen": 8925632, + "step": 11700 + }, + { + "epoch": 24.334719334719335, + "grad_norm": 0.0005839199875481427, + "learning_rate": 0.24096488729002086, + "loss": 0.2357, + "num_input_tokens_seen": 8929344, + "step": 11705 + }, + { + "epoch": 24.345114345114347, + "grad_norm": 0.0003611715801525861, + "learning_rate": 0.24091804296496946, + "loss": 0.2977, + "num_input_tokens_seen": 8933056, + "step": 11710 + }, + { + "epoch": 24.355509355509355, + "grad_norm": 0.0005658221780322492, + "learning_rate": 0.2408711846192133, + "loss": 0.2586, + "num_input_tokens_seen": 8936768, + "step": 11715 + }, + { + "epoch": 24.365904365904367, + "grad_norm": 0.0004351457755547017, + "learning_rate": 0.24082431225997855, + "loss": 0.2817, + "num_input_tokens_seen": 8940672, + "step": 11720 + }, + { + "epoch": 24.376299376299375, + "grad_norm": 0.0003449256473686546, + "learning_rate": 0.24077742589449344, + "loss": 0.262, + "num_input_tokens_seen": 8944512, + "step": 11725 + }, + { + "epoch": 24.386694386694387, + "grad_norm": 0.00021875255333725363, + "learning_rate": 0.24073052552998844, + "loss": 0.2464, + "num_input_tokens_seen": 8948256, + "step": 11730 + }, + { + "epoch": 24.397089397089395, + "grad_norm": 0.00021527346689254045, + "learning_rate": 0.2406836111736963, + "loss": 0.2698, + "num_input_tokens_seen": 8952032, + "step": 11735 + }, + { + "epoch": 24.407484407484407, + "grad_norm": 0.00019175326451659203, + "learning_rate": 0.2406366828328517, + "loss": 0.2813, + "num_input_tokens_seen": 8955872, + "step": 11740 + }, + { + "epoch": 24.41787941787942, + "grad_norm": 0.00028428129735402763, + "learning_rate": 0.2405897405146915, + "loss": 0.2885, + "num_input_tokens_seen": 8959680, + "step": 11745 + }, + { + "epoch": 24.428274428274428, + "grad_norm": 0.0006020207656547427, + "learning_rate": 0.240542784226455, + "loss": 0.2801, + "num_input_tokens_seen": 8963584, + "step": 11750 + }, + { + "epoch": 24.43866943866944, + "grad_norm": 0.0001576445938553661, + "learning_rate": 0.24049581397538328, + "loss": 0.2067, + "num_input_tokens_seen": 8967232, + "step": 11755 + }, + { + "epoch": 24.449064449064448, + "grad_norm": 0.0008181019220501184, + "learning_rate": 0.24044882976871984, + "loss": 0.2633, + "num_input_tokens_seen": 8971008, + "step": 11760 + }, + { + "epoch": 24.45945945945946, + "grad_norm": 0.00036928593181073666, + "learning_rate": 0.2404018316137102, + "loss": 0.2393, + "num_input_tokens_seen": 8974944, + "step": 11765 + }, + { + "epoch": 24.46985446985447, + "grad_norm": 0.0008843201212584972, + "learning_rate": 0.24035481951760204, + "loss": 0.2743, + "num_input_tokens_seen": 8978688, + "step": 11770 + }, + { + "epoch": 24.48024948024948, + "grad_norm": 0.000123951758723706, + "learning_rate": 0.2403077934876452, + "loss": 0.2674, + "num_input_tokens_seen": 8982432, + "step": 11775 + }, + { + "epoch": 24.490644490644492, + "grad_norm": 0.0010666167363524437, + "learning_rate": 0.2402607535310918, + "loss": 0.2826, + "num_input_tokens_seen": 8986272, + "step": 11780 + }, + { + "epoch": 24.5010395010395, + "grad_norm": 9.37069344217889e-05, + "learning_rate": 0.2402136996551959, + "loss": 0.2666, + "num_input_tokens_seen": 8990016, + "step": 11785 + }, + { + "epoch": 24.511434511434512, + "grad_norm": 0.00044376743608154356, + "learning_rate": 0.24016663186721376, + "loss": 0.3267, + "num_input_tokens_seen": 8993728, + "step": 11790 + }, + { + "epoch": 24.52182952182952, + "grad_norm": 0.0006809383048675954, + "learning_rate": 0.24011955017440395, + "loss": 0.2668, + "num_input_tokens_seen": 8997568, + "step": 11795 + }, + { + "epoch": 24.532224532224532, + "grad_norm": 0.0003921407333109528, + "learning_rate": 0.24007245458402696, + "loss": 0.2668, + "num_input_tokens_seen": 9001408, + "step": 11800 + }, + { + "epoch": 24.532224532224532, + "eval_loss": 0.2539379298686981, + "eval_runtime": 13.4097, + "eval_samples_per_second": 63.834, + "eval_steps_per_second": 15.959, + "num_input_tokens_seen": 9001408, + "step": 11800 + }, + { + "epoch": 24.542619542619544, + "grad_norm": 0.00016116558981593698, + "learning_rate": 0.2400253451033456, + "loss": 0.2579, + "num_input_tokens_seen": 9005152, + "step": 11805 + }, + { + "epoch": 24.553014553014552, + "grad_norm": 0.00011709408863680437, + "learning_rate": 0.23997822173962463, + "loss": 0.2544, + "num_input_tokens_seen": 9009056, + "step": 11810 + }, + { + "epoch": 24.563409563409564, + "grad_norm": 0.0003134087019134313, + "learning_rate": 0.23993108450013118, + "loss": 0.2778, + "num_input_tokens_seen": 9012832, + "step": 11815 + }, + { + "epoch": 24.573804573804573, + "grad_norm": 9.469052019994706e-05, + "learning_rate": 0.2398839333921343, + "loss": 0.2647, + "num_input_tokens_seen": 9016576, + "step": 11820 + }, + { + "epoch": 24.584199584199585, + "grad_norm": 0.00024380724062211812, + "learning_rate": 0.23983676842290536, + "loss": 0.2607, + "num_input_tokens_seen": 9020224, + "step": 11825 + }, + { + "epoch": 24.594594594594593, + "grad_norm": 0.0002000228123506531, + "learning_rate": 0.2397895895997178, + "loss": 0.2735, + "num_input_tokens_seen": 9024000, + "step": 11830 + }, + { + "epoch": 24.604989604989605, + "grad_norm": 0.0010783580364659429, + "learning_rate": 0.23974239692984714, + "loss": 0.278, + "num_input_tokens_seen": 9027776, + "step": 11835 + }, + { + "epoch": 24.615384615384617, + "grad_norm": 0.0009282429818995297, + "learning_rate": 0.2396951904205711, + "loss": 0.2704, + "num_input_tokens_seen": 9031520, + "step": 11840 + }, + { + "epoch": 24.625779625779625, + "grad_norm": 0.0003448710194788873, + "learning_rate": 0.23964797007916952, + "loss": 0.2776, + "num_input_tokens_seen": 9035360, + "step": 11845 + }, + { + "epoch": 24.636174636174637, + "grad_norm": 0.0006336988299153745, + "learning_rate": 0.23960073591292436, + "loss": 0.2617, + "num_input_tokens_seen": 9039104, + "step": 11850 + }, + { + "epoch": 24.646569646569645, + "grad_norm": 0.0003483004111330956, + "learning_rate": 0.2395534879291197, + "loss": 0.2606, + "num_input_tokens_seen": 9043008, + "step": 11855 + }, + { + "epoch": 24.656964656964657, + "grad_norm": 0.00030497240368276834, + "learning_rate": 0.23950622613504186, + "loss": 0.245, + "num_input_tokens_seen": 9046912, + "step": 11860 + }, + { + "epoch": 24.66735966735967, + "grad_norm": 0.00017458910588175058, + "learning_rate": 0.2394589505379791, + "loss": 0.2615, + "num_input_tokens_seen": 9050656, + "step": 11865 + }, + { + "epoch": 24.677754677754677, + "grad_norm": 0.0002103852020809427, + "learning_rate": 0.23941166114522197, + "loss": 0.2662, + "num_input_tokens_seen": 9054400, + "step": 11870 + }, + { + "epoch": 24.68814968814969, + "grad_norm": 0.00023987595341168344, + "learning_rate": 0.23936435796406308, + "loss": 0.2771, + "num_input_tokens_seen": 9058208, + "step": 11875 + }, + { + "epoch": 24.698544698544698, + "grad_norm": 0.0006227315752767026, + "learning_rate": 0.23931704100179715, + "loss": 0.267, + "num_input_tokens_seen": 9061952, + "step": 11880 + }, + { + "epoch": 24.70893970893971, + "grad_norm": 0.0004599630774464458, + "learning_rate": 0.2392697102657211, + "loss": 0.2836, + "num_input_tokens_seen": 9065664, + "step": 11885 + }, + { + "epoch": 24.719334719334718, + "grad_norm": 0.0006134499562904239, + "learning_rate": 0.23922236576313388, + "loss": 0.2832, + "num_input_tokens_seen": 9069344, + "step": 11890 + }, + { + "epoch": 24.72972972972973, + "grad_norm": 0.0002611627278383821, + "learning_rate": 0.2391750075013366, + "loss": 0.2769, + "num_input_tokens_seen": 9073184, + "step": 11895 + }, + { + "epoch": 24.74012474012474, + "grad_norm": 0.0002813110768329352, + "learning_rate": 0.2391276354876326, + "loss": 0.2613, + "num_input_tokens_seen": 9077088, + "step": 11900 + }, + { + "epoch": 24.75051975051975, + "grad_norm": 0.0008468155283480883, + "learning_rate": 0.23908024972932707, + "loss": 0.2888, + "num_input_tokens_seen": 9081024, + "step": 11905 + }, + { + "epoch": 24.760914760914762, + "grad_norm": 0.00011971390631515533, + "learning_rate": 0.2390328502337276, + "loss": 0.2704, + "num_input_tokens_seen": 9084896, + "step": 11910 + }, + { + "epoch": 24.77130977130977, + "grad_norm": 8.70052317623049e-05, + "learning_rate": 0.23898543700814376, + "loss": 0.2683, + "num_input_tokens_seen": 9088704, + "step": 11915 + }, + { + "epoch": 24.781704781704782, + "grad_norm": 0.00014342175563797355, + "learning_rate": 0.2389380100598873, + "loss": 0.2706, + "num_input_tokens_seen": 9092416, + "step": 11920 + }, + { + "epoch": 24.79209979209979, + "grad_norm": 0.00019683194113895297, + "learning_rate": 0.23889056939627207, + "loss": 0.2795, + "num_input_tokens_seen": 9096128, + "step": 11925 + }, + { + "epoch": 24.802494802494802, + "grad_norm": 0.0006490256637334824, + "learning_rate": 0.23884311502461386, + "loss": 0.2779, + "num_input_tokens_seen": 9099840, + "step": 11930 + }, + { + "epoch": 24.812889812889814, + "grad_norm": 0.00020424398826435208, + "learning_rate": 0.23879564695223088, + "loss": 0.2545, + "num_input_tokens_seen": 9103712, + "step": 11935 + }, + { + "epoch": 24.823284823284823, + "grad_norm": 0.00016514574235770851, + "learning_rate": 0.23874816518644332, + "loss": 0.2767, + "num_input_tokens_seen": 9107520, + "step": 11940 + }, + { + "epoch": 24.833679833679835, + "grad_norm": 0.00014729250688105822, + "learning_rate": 0.23870066973457335, + "loss": 0.2736, + "num_input_tokens_seen": 9111392, + "step": 11945 + }, + { + "epoch": 24.844074844074843, + "grad_norm": 7.94215957284905e-05, + "learning_rate": 0.23865316060394545, + "loss": 0.2413, + "num_input_tokens_seen": 9115168, + "step": 11950 + }, + { + "epoch": 24.854469854469855, + "grad_norm": 0.001129868789575994, + "learning_rate": 0.2386056378018861, + "loss": 0.2707, + "num_input_tokens_seen": 9118848, + "step": 11955 + }, + { + "epoch": 24.864864864864863, + "grad_norm": 0.000570828327909112, + "learning_rate": 0.2385581013357239, + "loss": 0.1991, + "num_input_tokens_seen": 9122880, + "step": 11960 + }, + { + "epoch": 24.875259875259875, + "grad_norm": 0.0002486240118741989, + "learning_rate": 0.23851055121278958, + "loss": 0.2431, + "num_input_tokens_seen": 9126688, + "step": 11965 + }, + { + "epoch": 24.885654885654887, + "grad_norm": 0.0006722926045767963, + "learning_rate": 0.23846298744041594, + "loss": 0.2749, + "num_input_tokens_seen": 9130464, + "step": 11970 + }, + { + "epoch": 24.896049896049895, + "grad_norm": 0.0006204298115335405, + "learning_rate": 0.23841541002593802, + "loss": 0.225, + "num_input_tokens_seen": 9134080, + "step": 11975 + }, + { + "epoch": 24.906444906444907, + "grad_norm": 0.00014286961231846362, + "learning_rate": 0.23836781897669276, + "loss": 0.2916, + "num_input_tokens_seen": 9137824, + "step": 11980 + }, + { + "epoch": 24.916839916839916, + "grad_norm": 0.00010744509927462786, + "learning_rate": 0.23832021430001926, + "loss": 0.2515, + "num_input_tokens_seen": 9141760, + "step": 11985 + }, + { + "epoch": 24.927234927234927, + "grad_norm": 0.00016330509970430285, + "learning_rate": 0.2382725960032588, + "loss": 0.2861, + "num_input_tokens_seen": 9145696, + "step": 11990 + }, + { + "epoch": 24.93762993762994, + "grad_norm": 0.000831234036013484, + "learning_rate": 0.23822496409375482, + "loss": 0.2813, + "num_input_tokens_seen": 9149760, + "step": 11995 + }, + { + "epoch": 24.948024948024948, + "grad_norm": 0.00016153437900356948, + "learning_rate": 0.2381773185788526, + "loss": 0.2822, + "num_input_tokens_seen": 9153696, + "step": 12000 + }, + { + "epoch": 24.948024948024948, + "eval_loss": 0.2769145369529724, + "eval_runtime": 13.4503, + "eval_samples_per_second": 63.642, + "eval_steps_per_second": 15.91, + "num_input_tokens_seen": 9153696, + "step": 12000 + }, + { + "epoch": 24.95841995841996, + "grad_norm": 0.0004421588673721999, + "learning_rate": 0.2381296594658998, + "loss": 0.2743, + "num_input_tokens_seen": 9157440, + "step": 12005 + }, + { + "epoch": 24.968814968814968, + "grad_norm": 0.00010778033902170137, + "learning_rate": 0.238081986762246, + "loss": 0.26, + "num_input_tokens_seen": 9161248, + "step": 12010 + }, + { + "epoch": 24.97920997920998, + "grad_norm": 0.0003534658462740481, + "learning_rate": 0.23803430047524293, + "loss": 0.2621, + "num_input_tokens_seen": 9164960, + "step": 12015 + }, + { + "epoch": 24.989604989604988, + "grad_norm": 0.0002858824154827744, + "learning_rate": 0.23798660061224441, + "loss": 0.2564, + "num_input_tokens_seen": 9168704, + "step": 12020 + }, + { + "epoch": 25.0, + "grad_norm": 0.000365712505299598, + "learning_rate": 0.23793888718060632, + "loss": 0.2494, + "num_input_tokens_seen": 9172624, + "step": 12025 + }, + { + "epoch": 25.010395010395012, + "grad_norm": 0.00014775973977521062, + "learning_rate": 0.23789116018768675, + "loss": 0.2672, + "num_input_tokens_seen": 9176400, + "step": 12030 + }, + { + "epoch": 25.02079002079002, + "grad_norm": 0.0006968433735892177, + "learning_rate": 0.2378434196408458, + "loss": 0.2701, + "num_input_tokens_seen": 9180272, + "step": 12035 + }, + { + "epoch": 25.031185031185032, + "grad_norm": 7.712693331995979e-05, + "learning_rate": 0.23779566554744563, + "loss": 0.2756, + "num_input_tokens_seen": 9184080, + "step": 12040 + }, + { + "epoch": 25.04158004158004, + "grad_norm": 0.0003730101161636412, + "learning_rate": 0.23774789791485051, + "loss": 0.2769, + "num_input_tokens_seen": 9187824, + "step": 12045 + }, + { + "epoch": 25.051975051975052, + "grad_norm": 0.0001803341438062489, + "learning_rate": 0.2377001167504268, + "loss": 0.2576, + "num_input_tokens_seen": 9191536, + "step": 12050 + }, + { + "epoch": 25.06237006237006, + "grad_norm": 0.0001336833811365068, + "learning_rate": 0.23765232206154302, + "loss": 0.2448, + "num_input_tokens_seen": 9195408, + "step": 12055 + }, + { + "epoch": 25.072765072765073, + "grad_norm": 0.00039558636490255594, + "learning_rate": 0.23760451385556966, + "loss": 0.2943, + "num_input_tokens_seen": 9199152, + "step": 12060 + }, + { + "epoch": 25.083160083160084, + "grad_norm": 0.00046535010915249586, + "learning_rate": 0.23755669213987932, + "loss": 0.2656, + "num_input_tokens_seen": 9202960, + "step": 12065 + }, + { + "epoch": 25.093555093555093, + "grad_norm": 0.0004747312341351062, + "learning_rate": 0.23750885692184676, + "loss": 0.2673, + "num_input_tokens_seen": 9206832, + "step": 12070 + }, + { + "epoch": 25.103950103950105, + "grad_norm": 0.0007404095958918333, + "learning_rate": 0.23746100820884875, + "loss": 0.2673, + "num_input_tokens_seen": 9210736, + "step": 12075 + }, + { + "epoch": 25.114345114345113, + "grad_norm": 7.103076495695859e-05, + "learning_rate": 0.23741314600826421, + "loss": 0.2747, + "num_input_tokens_seen": 9214576, + "step": 12080 + }, + { + "epoch": 25.124740124740125, + "grad_norm": 0.0006793164066039026, + "learning_rate": 0.23736527032747406, + "loss": 0.2906, + "num_input_tokens_seen": 9218256, + "step": 12085 + }, + { + "epoch": 25.135135135135137, + "grad_norm": 0.0004902869113720953, + "learning_rate": 0.23731738117386128, + "loss": 0.2769, + "num_input_tokens_seen": 9221968, + "step": 12090 + }, + { + "epoch": 25.145530145530145, + "grad_norm": 0.0008474405622109771, + "learning_rate": 0.237269478554811, + "loss": 0.2653, + "num_input_tokens_seen": 9225872, + "step": 12095 + }, + { + "epoch": 25.155925155925157, + "grad_norm": 0.00038160482654348016, + "learning_rate": 0.23722156247771053, + "loss": 0.2949, + "num_input_tokens_seen": 9229808, + "step": 12100 + }, + { + "epoch": 25.166320166320165, + "grad_norm": 4.230980630381964e-05, + "learning_rate": 0.23717363294994895, + "loss": 0.2766, + "num_input_tokens_seen": 9233456, + "step": 12105 + }, + { + "epoch": 25.176715176715177, + "grad_norm": 0.0004975068150088191, + "learning_rate": 0.2371256899789177, + "loss": 0.2682, + "num_input_tokens_seen": 9237360, + "step": 12110 + }, + { + "epoch": 25.187110187110186, + "grad_norm": 0.00021768540318589658, + "learning_rate": 0.23707773357201017, + "loss": 0.2703, + "num_input_tokens_seen": 9241360, + "step": 12115 + }, + { + "epoch": 25.197505197505198, + "grad_norm": 0.0001609144965186715, + "learning_rate": 0.2370297637366218, + "loss": 0.2596, + "num_input_tokens_seen": 9245264, + "step": 12120 + }, + { + "epoch": 25.20790020790021, + "grad_norm": 0.0005136209656484425, + "learning_rate": 0.23698178048015026, + "loss": 0.2835, + "num_input_tokens_seen": 9249072, + "step": 12125 + }, + { + "epoch": 25.218295218295218, + "grad_norm": 0.0002003306581173092, + "learning_rate": 0.236933783809995, + "loss": 0.2702, + "num_input_tokens_seen": 9252880, + "step": 12130 + }, + { + "epoch": 25.22869022869023, + "grad_norm": 3.456450212979689e-05, + "learning_rate": 0.23688577373355785, + "loss": 0.2677, + "num_input_tokens_seen": 9256656, + "step": 12135 + }, + { + "epoch": 25.239085239085238, + "grad_norm": 0.0002603501779958606, + "learning_rate": 0.23683775025824247, + "loss": 0.275, + "num_input_tokens_seen": 9260400, + "step": 12140 + }, + { + "epoch": 25.24948024948025, + "grad_norm": 0.0003529172099661082, + "learning_rate": 0.2367897133914548, + "loss": 0.258, + "num_input_tokens_seen": 9264336, + "step": 12145 + }, + { + "epoch": 25.25987525987526, + "grad_norm": 0.000611013441812247, + "learning_rate": 0.2367416631406026, + "loss": 0.2657, + "num_input_tokens_seen": 9268400, + "step": 12150 + }, + { + "epoch": 25.27027027027027, + "grad_norm": 0.0006457085255533457, + "learning_rate": 0.23669359951309588, + "loss": 0.2756, + "num_input_tokens_seen": 9272208, + "step": 12155 + }, + { + "epoch": 25.280665280665282, + "grad_norm": 0.0004691470239777118, + "learning_rate": 0.23664552251634666, + "loss": 0.2663, + "num_input_tokens_seen": 9276176, + "step": 12160 + }, + { + "epoch": 25.29106029106029, + "grad_norm": 0.00025904655922204256, + "learning_rate": 0.23659743215776907, + "loss": 0.2796, + "num_input_tokens_seen": 9280048, + "step": 12165 + }, + { + "epoch": 25.301455301455302, + "grad_norm": 7.312667730730027e-05, + "learning_rate": 0.23654932844477908, + "loss": 0.264, + "num_input_tokens_seen": 9284016, + "step": 12170 + }, + { + "epoch": 25.31185031185031, + "grad_norm": 0.0007526806439273059, + "learning_rate": 0.23650121138479507, + "loss": 0.2645, + "num_input_tokens_seen": 9287920, + "step": 12175 + }, + { + "epoch": 25.322245322245323, + "grad_norm": 0.0001987564901355654, + "learning_rate": 0.23645308098523724, + "loss": 0.2382, + "num_input_tokens_seen": 9291824, + "step": 12180 + }, + { + "epoch": 25.33264033264033, + "grad_norm": 0.00020318591850809753, + "learning_rate": 0.23640493725352785, + "loss": 0.2647, + "num_input_tokens_seen": 9295632, + "step": 12185 + }, + { + "epoch": 25.343035343035343, + "grad_norm": 0.00010177848162129521, + "learning_rate": 0.2363567801970913, + "loss": 0.2636, + "num_input_tokens_seen": 9299472, + "step": 12190 + }, + { + "epoch": 25.353430353430355, + "grad_norm": 0.00040496455039829016, + "learning_rate": 0.236308609823354, + "loss": 0.269, + "num_input_tokens_seen": 9303344, + "step": 12195 + }, + { + "epoch": 25.363825363825363, + "grad_norm": 0.0003969160025008023, + "learning_rate": 0.23626042613974452, + "loss": 0.2745, + "num_input_tokens_seen": 9307088, + "step": 12200 + }, + { + "epoch": 25.363825363825363, + "eval_loss": 0.2504133880138397, + "eval_runtime": 13.4638, + "eval_samples_per_second": 63.578, + "eval_steps_per_second": 15.894, + "num_input_tokens_seen": 9307088, + "step": 12200 + }, + { + "epoch": 25.374220374220375, + "grad_norm": 0.00026065215934067965, + "learning_rate": 0.23621222915369325, + "loss": 0.2615, + "num_input_tokens_seen": 9310928, + "step": 12205 + }, + { + "epoch": 25.384615384615383, + "grad_norm": 0.0005557139520533383, + "learning_rate": 0.23616401887263283, + "loss": 0.2371, + "num_input_tokens_seen": 9314672, + "step": 12210 + }, + { + "epoch": 25.395010395010395, + "grad_norm": 5.077619061921723e-05, + "learning_rate": 0.23611579530399793, + "loss": 0.2606, + "num_input_tokens_seen": 9318416, + "step": 12215 + }, + { + "epoch": 25.405405405405407, + "grad_norm": 0.00044916279148310423, + "learning_rate": 0.23606755845522517, + "loss": 0.2587, + "num_input_tokens_seen": 9322128, + "step": 12220 + }, + { + "epoch": 25.415800415800415, + "grad_norm": 0.00030319116194732487, + "learning_rate": 0.23601930833375329, + "loss": 0.2487, + "num_input_tokens_seen": 9326000, + "step": 12225 + }, + { + "epoch": 25.426195426195427, + "grad_norm": 0.0004046520043630153, + "learning_rate": 0.23597104494702312, + "loss": 0.2696, + "num_input_tokens_seen": 9329904, + "step": 12230 + }, + { + "epoch": 25.436590436590436, + "grad_norm": 0.000509279256220907, + "learning_rate": 0.23592276830247744, + "loss": 0.2629, + "num_input_tokens_seen": 9333648, + "step": 12235 + }, + { + "epoch": 25.446985446985448, + "grad_norm": 0.0005246573709882796, + "learning_rate": 0.2358744784075611, + "loss": 0.2783, + "num_input_tokens_seen": 9337360, + "step": 12240 + }, + { + "epoch": 25.457380457380456, + "grad_norm": 0.000660305842757225, + "learning_rate": 0.235826175269721, + "loss": 0.273, + "num_input_tokens_seen": 9341232, + "step": 12245 + }, + { + "epoch": 25.467775467775468, + "grad_norm": 0.00025600846856832504, + "learning_rate": 0.23577785889640612, + "loss": 0.2568, + "num_input_tokens_seen": 9345104, + "step": 12250 + }, + { + "epoch": 25.47817047817048, + "grad_norm": 0.00022463449568022043, + "learning_rate": 0.23572952929506744, + "loss": 0.2548, + "num_input_tokens_seen": 9348912, + "step": 12255 + }, + { + "epoch": 25.488565488565488, + "grad_norm": 0.0007313215173780918, + "learning_rate": 0.23568118647315803, + "loss": 0.2761, + "num_input_tokens_seen": 9352784, + "step": 12260 + }, + { + "epoch": 25.4989604989605, + "grad_norm": 5.804266038467176e-05, + "learning_rate": 0.23563283043813296, + "loss": 0.2666, + "num_input_tokens_seen": 9356528, + "step": 12265 + }, + { + "epoch": 25.509355509355508, + "grad_norm": 0.00037733608041889966, + "learning_rate": 0.23558446119744922, + "loss": 0.2586, + "num_input_tokens_seen": 9360432, + "step": 12270 + }, + { + "epoch": 25.51975051975052, + "grad_norm": 0.00020216168195474893, + "learning_rate": 0.23553607875856608, + "loss": 0.2632, + "num_input_tokens_seen": 9364144, + "step": 12275 + }, + { + "epoch": 25.53014553014553, + "grad_norm": 0.0003941018076147884, + "learning_rate": 0.2354876831289447, + "loss": 0.2653, + "num_input_tokens_seen": 9367888, + "step": 12280 + }, + { + "epoch": 25.54054054054054, + "grad_norm": 0.0004069653805345297, + "learning_rate": 0.23543927431604827, + "loss": 0.2553, + "num_input_tokens_seen": 9371792, + "step": 12285 + }, + { + "epoch": 25.550935550935552, + "grad_norm": 0.00014301339979283512, + "learning_rate": 0.23539085232734203, + "loss": 0.2451, + "num_input_tokens_seen": 9375504, + "step": 12290 + }, + { + "epoch": 25.56133056133056, + "grad_norm": 0.00013156987552065402, + "learning_rate": 0.2353424171702933, + "loss": 0.2976, + "num_input_tokens_seen": 9379152, + "step": 12295 + }, + { + "epoch": 25.571725571725572, + "grad_norm": 0.0006048521609045565, + "learning_rate": 0.23529396885237133, + "loss": 0.2722, + "num_input_tokens_seen": 9382960, + "step": 12300 + }, + { + "epoch": 25.58212058212058, + "grad_norm": 9.623035293770954e-05, + "learning_rate": 0.2352455073810475, + "loss": 0.2744, + "num_input_tokens_seen": 9386864, + "step": 12305 + }, + { + "epoch": 25.592515592515593, + "grad_norm": 0.00025088590336963534, + "learning_rate": 0.23519703276379517, + "loss": 0.2298, + "num_input_tokens_seen": 9390736, + "step": 12310 + }, + { + "epoch": 25.602910602910605, + "grad_norm": 0.0004715774266514927, + "learning_rate": 0.2351485450080897, + "loss": 0.2988, + "num_input_tokens_seen": 9394576, + "step": 12315 + }, + { + "epoch": 25.613305613305613, + "grad_norm": 0.00013741556904278696, + "learning_rate": 0.2351000441214086, + "loss": 0.2709, + "num_input_tokens_seen": 9398384, + "step": 12320 + }, + { + "epoch": 25.623700623700625, + "grad_norm": 0.000391049514291808, + "learning_rate": 0.23505153011123125, + "loss": 0.2734, + "num_input_tokens_seen": 9402256, + "step": 12325 + }, + { + "epoch": 25.634095634095633, + "grad_norm": 0.0010021297493949533, + "learning_rate": 0.23500300298503912, + "loss": 0.2923, + "num_input_tokens_seen": 9406224, + "step": 12330 + }, + { + "epoch": 25.644490644490645, + "grad_norm": 0.00020896481873933226, + "learning_rate": 0.23495446275031576, + "loss": 0.2692, + "num_input_tokens_seen": 9410160, + "step": 12335 + }, + { + "epoch": 25.654885654885653, + "grad_norm": 0.0005280310288071632, + "learning_rate": 0.2349059094145466, + "loss": 0.2783, + "num_input_tokens_seen": 9413936, + "step": 12340 + }, + { + "epoch": 25.665280665280665, + "grad_norm": 0.00010060082422569394, + "learning_rate": 0.2348573429852192, + "loss": 0.2325, + "num_input_tokens_seen": 9417744, + "step": 12345 + }, + { + "epoch": 25.675675675675677, + "grad_norm": 0.00010296085383743048, + "learning_rate": 0.23480876346982313, + "loss": 0.2518, + "num_input_tokens_seen": 9421520, + "step": 12350 + }, + { + "epoch": 25.686070686070686, + "grad_norm": 0.00018870858184527606, + "learning_rate": 0.23476017087585, + "loss": 0.2844, + "num_input_tokens_seen": 9425456, + "step": 12355 + }, + { + "epoch": 25.696465696465697, + "grad_norm": 0.00025114015443250537, + "learning_rate": 0.23471156521079334, + "loss": 0.2705, + "num_input_tokens_seen": 9429200, + "step": 12360 + }, + { + "epoch": 25.706860706860706, + "grad_norm": 0.000310354633256793, + "learning_rate": 0.23466294648214875, + "loss": 0.2762, + "num_input_tokens_seen": 9433040, + "step": 12365 + }, + { + "epoch": 25.717255717255718, + "grad_norm": 0.00012162828352302313, + "learning_rate": 0.2346143146974139, + "loss": 0.2281, + "num_input_tokens_seen": 9436848, + "step": 12370 + }, + { + "epoch": 25.727650727650726, + "grad_norm": 0.00043991600978188217, + "learning_rate": 0.23456566986408836, + "loss": 0.3021, + "num_input_tokens_seen": 9440752, + "step": 12375 + }, + { + "epoch": 25.738045738045738, + "grad_norm": 0.0002981819270644337, + "learning_rate": 0.23451701198967384, + "loss": 0.2587, + "num_input_tokens_seen": 9444592, + "step": 12380 + }, + { + "epoch": 25.74844074844075, + "grad_norm": 0.00028357660630717874, + "learning_rate": 0.23446834108167397, + "loss": 0.2792, + "num_input_tokens_seen": 9448400, + "step": 12385 + }, + { + "epoch": 25.758835758835758, + "grad_norm": 0.00033950238139368594, + "learning_rate": 0.23441965714759438, + "loss": 0.2724, + "num_input_tokens_seen": 9452304, + "step": 12390 + }, + { + "epoch": 25.76923076923077, + "grad_norm": 0.0009178064065054059, + "learning_rate": 0.23437096019494277, + "loss": 0.2677, + "num_input_tokens_seen": 9455984, + "step": 12395 + }, + { + "epoch": 25.77962577962578, + "grad_norm": 0.0008330817800015211, + "learning_rate": 0.23432225023122885, + "loss": 0.2536, + "num_input_tokens_seen": 9459824, + "step": 12400 + }, + { + "epoch": 25.77962577962578, + "eval_loss": 0.24817043542861938, + "eval_runtime": 13.4113, + "eval_samples_per_second": 63.827, + "eval_steps_per_second": 15.957, + "num_input_tokens_seen": 9459824, + "step": 12400 + }, + { + "epoch": 25.79002079002079, + "grad_norm": 0.0006262730457819998, + "learning_rate": 0.23427352726396428, + "loss": 0.274, + "num_input_tokens_seen": 9463664, + "step": 12405 + }, + { + "epoch": 25.8004158004158, + "grad_norm": 0.00023454749316442758, + "learning_rate": 0.2342247913006628, + "loss": 0.272, + "num_input_tokens_seen": 9467472, + "step": 12410 + }, + { + "epoch": 25.81081081081081, + "grad_norm": 0.00041848912951536477, + "learning_rate": 0.23417604234883999, + "loss": 0.2767, + "num_input_tokens_seen": 9471088, + "step": 12415 + }, + { + "epoch": 25.821205821205822, + "grad_norm": 0.00023458755458705127, + "learning_rate": 0.23412728041601363, + "loss": 0.2786, + "num_input_tokens_seen": 9474864, + "step": 12420 + }, + { + "epoch": 25.83160083160083, + "grad_norm": 0.00040733005153015256, + "learning_rate": 0.23407850550970347, + "loss": 0.277, + "num_input_tokens_seen": 9478576, + "step": 12425 + }, + { + "epoch": 25.841995841995843, + "grad_norm": 0.00040534278377890587, + "learning_rate": 0.23402971763743116, + "loss": 0.2601, + "num_input_tokens_seen": 9482384, + "step": 12430 + }, + { + "epoch": 25.85239085239085, + "grad_norm": 0.0006408787448890507, + "learning_rate": 0.23398091680672037, + "loss": 0.2786, + "num_input_tokens_seen": 9486288, + "step": 12435 + }, + { + "epoch": 25.862785862785863, + "grad_norm": 0.00020789298287127167, + "learning_rate": 0.23393210302509687, + "loss": 0.2902, + "num_input_tokens_seen": 9490032, + "step": 12440 + }, + { + "epoch": 25.873180873180875, + "grad_norm": 0.00016575797053519636, + "learning_rate": 0.23388327630008832, + "loss": 0.2932, + "num_input_tokens_seen": 9493808, + "step": 12445 + }, + { + "epoch": 25.883575883575883, + "grad_norm": 5.5754659115336835e-05, + "learning_rate": 0.23383443663922443, + "loss": 0.2488, + "num_input_tokens_seen": 9497616, + "step": 12450 + }, + { + "epoch": 25.893970893970895, + "grad_norm": 0.0003425349132157862, + "learning_rate": 0.23378558405003685, + "loss": 0.3026, + "num_input_tokens_seen": 9501520, + "step": 12455 + }, + { + "epoch": 25.904365904365903, + "grad_norm": 0.0005627562059089541, + "learning_rate": 0.2337367185400593, + "loss": 0.2786, + "num_input_tokens_seen": 9505520, + "step": 12460 + }, + { + "epoch": 25.914760914760915, + "grad_norm": 0.0004024420923087746, + "learning_rate": 0.23368784011682747, + "loss": 0.2746, + "num_input_tokens_seen": 9509328, + "step": 12465 + }, + { + "epoch": 25.925155925155924, + "grad_norm": 9.865907486528158e-05, + "learning_rate": 0.23363894878787902, + "loss": 0.2612, + "num_input_tokens_seen": 9513168, + "step": 12470 + }, + { + "epoch": 25.935550935550935, + "grad_norm": 0.0007322691380977631, + "learning_rate": 0.23359004456075352, + "loss": 0.2718, + "num_input_tokens_seen": 9516976, + "step": 12475 + }, + { + "epoch": 25.945945945945947, + "grad_norm": 0.00012380574480630457, + "learning_rate": 0.23354112744299277, + "loss": 0.2899, + "num_input_tokens_seen": 9520816, + "step": 12480 + }, + { + "epoch": 25.956340956340956, + "grad_norm": 0.0006794001092202961, + "learning_rate": 0.2334921974421403, + "loss": 0.2575, + "num_input_tokens_seen": 9524592, + "step": 12485 + }, + { + "epoch": 25.966735966735968, + "grad_norm": 0.0005183862522244453, + "learning_rate": 0.23344325456574178, + "loss": 0.2586, + "num_input_tokens_seen": 9528336, + "step": 12490 + }, + { + "epoch": 25.977130977130976, + "grad_norm": 0.0009788044262677431, + "learning_rate": 0.23339429882134477, + "loss": 0.2903, + "num_input_tokens_seen": 9532080, + "step": 12495 + }, + { + "epoch": 25.987525987525988, + "grad_norm": 0.00023358248290605843, + "learning_rate": 0.23334533021649884, + "loss": 0.2761, + "num_input_tokens_seen": 9535760, + "step": 12500 + }, + { + "epoch": 25.997920997921, + "grad_norm": 0.00012860505376011133, + "learning_rate": 0.23329634875875566, + "loss": 0.2841, + "num_input_tokens_seen": 9539440, + "step": 12505 + }, + { + "epoch": 26.008316008316008, + "grad_norm": 0.00021904618188273162, + "learning_rate": 0.23324735445566874, + "loss": 0.2708, + "num_input_tokens_seen": 9543000, + "step": 12510 + }, + { + "epoch": 26.01871101871102, + "grad_norm": 0.00015079065633472055, + "learning_rate": 0.2331983473147936, + "loss": 0.2574, + "num_input_tokens_seen": 9546840, + "step": 12515 + }, + { + "epoch": 26.02910602910603, + "grad_norm": 0.00047588636516593397, + "learning_rate": 0.23314932734368776, + "loss": 0.3206, + "num_input_tokens_seen": 9550808, + "step": 12520 + }, + { + "epoch": 26.03950103950104, + "grad_norm": 0.0009797702077776194, + "learning_rate": 0.2331002945499107, + "loss": 0.2406, + "num_input_tokens_seen": 9554520, + "step": 12525 + }, + { + "epoch": 26.04989604989605, + "grad_norm": 0.0001158341474365443, + "learning_rate": 0.23305124894102397, + "loss": 0.2722, + "num_input_tokens_seen": 9558328, + "step": 12530 + }, + { + "epoch": 26.06029106029106, + "grad_norm": 0.0003536955046001822, + "learning_rate": 0.23300219052459092, + "loss": 0.2813, + "num_input_tokens_seen": 9562104, + "step": 12535 + }, + { + "epoch": 26.070686070686072, + "grad_norm": 0.0009927182691171765, + "learning_rate": 0.23295311930817708, + "loss": 0.2656, + "num_input_tokens_seen": 9565976, + "step": 12540 + }, + { + "epoch": 26.08108108108108, + "grad_norm": 0.00011536591046024114, + "learning_rate": 0.23290403529934972, + "loss": 0.264, + "num_input_tokens_seen": 9569912, + "step": 12545 + }, + { + "epoch": 26.091476091476093, + "grad_norm": 0.0006257393979467452, + "learning_rate": 0.23285493850567832, + "loss": 0.2761, + "num_input_tokens_seen": 9573816, + "step": 12550 + }, + { + "epoch": 26.1018711018711, + "grad_norm": 0.0006008932250551879, + "learning_rate": 0.23280582893473414, + "loss": 0.2807, + "num_input_tokens_seen": 9577656, + "step": 12555 + }, + { + "epoch": 26.112266112266113, + "grad_norm": 0.0003209675778634846, + "learning_rate": 0.2327567065940906, + "loss": 0.2795, + "num_input_tokens_seen": 9581368, + "step": 12560 + }, + { + "epoch": 26.12266112266112, + "grad_norm": 0.00016421207692474127, + "learning_rate": 0.23270757149132285, + "loss": 0.2745, + "num_input_tokens_seen": 9585176, + "step": 12565 + }, + { + "epoch": 26.133056133056133, + "grad_norm": 8.873855404090136e-05, + "learning_rate": 0.23265842363400827, + "loss": 0.2714, + "num_input_tokens_seen": 9588952, + "step": 12570 + }, + { + "epoch": 26.143451143451145, + "grad_norm": 0.0004526890115812421, + "learning_rate": 0.23260926302972595, + "loss": 0.25, + "num_input_tokens_seen": 9592568, + "step": 12575 + }, + { + "epoch": 26.153846153846153, + "grad_norm": 0.000568743736948818, + "learning_rate": 0.2325600896860572, + "loss": 0.2757, + "num_input_tokens_seen": 9596408, + "step": 12580 + }, + { + "epoch": 26.164241164241165, + "grad_norm": 7.831402035662904e-05, + "learning_rate": 0.23251090361058505, + "loss": 0.277, + "num_input_tokens_seen": 9600152, + "step": 12585 + }, + { + "epoch": 26.174636174636174, + "grad_norm": 0.0004367585643194616, + "learning_rate": 0.23246170481089476, + "loss": 0.2763, + "num_input_tokens_seen": 9603992, + "step": 12590 + }, + { + "epoch": 26.185031185031185, + "grad_norm": 0.00038059434155002236, + "learning_rate": 0.23241249329457317, + "loss": 0.2798, + "num_input_tokens_seen": 9607800, + "step": 12595 + }, + { + "epoch": 26.195426195426194, + "grad_norm": 9.448687342228368e-05, + "learning_rate": 0.23236326906920957, + "loss": 0.2723, + "num_input_tokens_seen": 9611704, + "step": 12600 + }, + { + "epoch": 26.195426195426194, + "eval_loss": 0.24951080977916718, + "eval_runtime": 13.4167, + "eval_samples_per_second": 63.801, + "eval_steps_per_second": 15.95, + "num_input_tokens_seen": 9611704, + "step": 12600 + }, + { + "epoch": 26.205821205821206, + "grad_norm": 0.0008909075404517353, + "learning_rate": 0.2323140321423948, + "loss": 0.2801, + "num_input_tokens_seen": 9615352, + "step": 12605 + }, + { + "epoch": 26.216216216216218, + "grad_norm": 0.0003795934026129544, + "learning_rate": 0.23226478252172184, + "loss": 0.2354, + "num_input_tokens_seen": 9619064, + "step": 12610 + }, + { + "epoch": 26.226611226611226, + "grad_norm": 0.0008331681019626558, + "learning_rate": 0.23221552021478561, + "loss": 0.2773, + "num_input_tokens_seen": 9622904, + "step": 12615 + }, + { + "epoch": 26.237006237006238, + "grad_norm": 7.846447988413274e-05, + "learning_rate": 0.232166245229183, + "loss": 0.2573, + "num_input_tokens_seen": 9626776, + "step": 12620 + }, + { + "epoch": 26.247401247401246, + "grad_norm": 0.00021828798344358802, + "learning_rate": 0.2321169575725128, + "loss": 0.271, + "num_input_tokens_seen": 9630712, + "step": 12625 + }, + { + "epoch": 26.257796257796258, + "grad_norm": 0.0002588498464319855, + "learning_rate": 0.23206765725237577, + "loss": 0.2802, + "num_input_tokens_seen": 9634520, + "step": 12630 + }, + { + "epoch": 26.26819126819127, + "grad_norm": 6.29895948804915e-05, + "learning_rate": 0.2320183442763747, + "loss": 0.2517, + "num_input_tokens_seen": 9638296, + "step": 12635 + }, + { + "epoch": 26.27858627858628, + "grad_norm": 0.00019017315935343504, + "learning_rate": 0.23196901865211422, + "loss": 0.277, + "num_input_tokens_seen": 9642328, + "step": 12640 + }, + { + "epoch": 26.28898128898129, + "grad_norm": 0.0003713190380949527, + "learning_rate": 0.231919680387201, + "loss": 0.2782, + "num_input_tokens_seen": 9646008, + "step": 12645 + }, + { + "epoch": 26.2993762993763, + "grad_norm": 0.0006231119041331112, + "learning_rate": 0.23187032948924358, + "loss": 0.2599, + "num_input_tokens_seen": 9649848, + "step": 12650 + }, + { + "epoch": 26.30977130977131, + "grad_norm": 0.00022300405544228852, + "learning_rate": 0.23182096596585247, + "loss": 0.2363, + "num_input_tokens_seen": 9653592, + "step": 12655 + }, + { + "epoch": 26.32016632016632, + "grad_norm": 0.001334618660621345, + "learning_rate": 0.23177158982464025, + "loss": 0.3357, + "num_input_tokens_seen": 9657464, + "step": 12660 + }, + { + "epoch": 26.33056133056133, + "grad_norm": 0.0001812281843740493, + "learning_rate": 0.23172220107322122, + "loss": 0.262, + "num_input_tokens_seen": 9661144, + "step": 12665 + }, + { + "epoch": 26.340956340956343, + "grad_norm": 0.0002520125708542764, + "learning_rate": 0.23167279971921184, + "loss": 0.283, + "num_input_tokens_seen": 9665048, + "step": 12670 + }, + { + "epoch": 26.35135135135135, + "grad_norm": 0.00042743273661471903, + "learning_rate": 0.23162338577023034, + "loss": 0.2877, + "num_input_tokens_seen": 9668728, + "step": 12675 + }, + { + "epoch": 26.361746361746363, + "grad_norm": 0.0008316849125549197, + "learning_rate": 0.23157395923389704, + "loss": 0.2605, + "num_input_tokens_seen": 9672664, + "step": 12680 + }, + { + "epoch": 26.37214137214137, + "grad_norm": 0.0007197829545475543, + "learning_rate": 0.2315245201178341, + "loss": 0.3093, + "num_input_tokens_seen": 9676760, + "step": 12685 + }, + { + "epoch": 26.382536382536383, + "grad_norm": 0.0007787124486640096, + "learning_rate": 0.23147506842966564, + "loss": 0.2814, + "num_input_tokens_seen": 9680728, + "step": 12690 + }, + { + "epoch": 26.39293139293139, + "grad_norm": 9.129822865361348e-05, + "learning_rate": 0.23142560417701774, + "loss": 0.2679, + "num_input_tokens_seen": 9684440, + "step": 12695 + }, + { + "epoch": 26.403326403326403, + "grad_norm": 0.0009400932467542589, + "learning_rate": 0.23137612736751845, + "loss": 0.2754, + "num_input_tokens_seen": 9688248, + "step": 12700 + }, + { + "epoch": 26.413721413721415, + "grad_norm": 0.0003207334375474602, + "learning_rate": 0.23132663800879766, + "loss": 0.234, + "num_input_tokens_seen": 9692152, + "step": 12705 + }, + { + "epoch": 26.424116424116423, + "grad_norm": 8.331677236128598e-05, + "learning_rate": 0.2312771361084873, + "loss": 0.2579, + "num_input_tokens_seen": 9695800, + "step": 12710 + }, + { + "epoch": 26.434511434511435, + "grad_norm": 0.0001733617827994749, + "learning_rate": 0.23122762167422112, + "loss": 0.2591, + "num_input_tokens_seen": 9699512, + "step": 12715 + }, + { + "epoch": 26.444906444906444, + "grad_norm": 0.0002785191172733903, + "learning_rate": 0.23117809471363493, + "loss": 0.2672, + "num_input_tokens_seen": 9703384, + "step": 12720 + }, + { + "epoch": 26.455301455301456, + "grad_norm": 0.0005105927702970803, + "learning_rate": 0.23112855523436637, + "loss": 0.2628, + "num_input_tokens_seen": 9707032, + "step": 12725 + }, + { + "epoch": 26.465696465696467, + "grad_norm": 0.00028112190193496644, + "learning_rate": 0.23107900324405511, + "loss": 0.2648, + "num_input_tokens_seen": 9711000, + "step": 12730 + }, + { + "epoch": 26.476091476091476, + "grad_norm": 0.0001392694830428809, + "learning_rate": 0.2310294387503426, + "loss": 0.2557, + "num_input_tokens_seen": 9714744, + "step": 12735 + }, + { + "epoch": 26.486486486486488, + "grad_norm": 0.00015754684864077717, + "learning_rate": 0.23097986176087237, + "loss": 0.2304, + "num_input_tokens_seen": 9718712, + "step": 12740 + }, + { + "epoch": 26.496881496881496, + "grad_norm": 0.0005974313244223595, + "learning_rate": 0.23093027228328986, + "loss": 0.3189, + "num_input_tokens_seen": 9722456, + "step": 12745 + }, + { + "epoch": 26.507276507276508, + "grad_norm": 0.0004982036189176142, + "learning_rate": 0.23088067032524226, + "loss": 0.2692, + "num_input_tokens_seen": 9726296, + "step": 12750 + }, + { + "epoch": 26.517671517671516, + "grad_norm": 0.00012270480510778725, + "learning_rate": 0.23083105589437888, + "loss": 0.2662, + "num_input_tokens_seen": 9729976, + "step": 12755 + }, + { + "epoch": 26.528066528066528, + "grad_norm": 0.00018452580843586475, + "learning_rate": 0.23078142899835094, + "loss": 0.2667, + "num_input_tokens_seen": 9733784, + "step": 12760 + }, + { + "epoch": 26.53846153846154, + "grad_norm": 0.00037734053330495954, + "learning_rate": 0.23073178964481147, + "loss": 0.2489, + "num_input_tokens_seen": 9737528, + "step": 12765 + }, + { + "epoch": 26.54885654885655, + "grad_norm": 0.0004967832937836647, + "learning_rate": 0.2306821378414155, + "loss": 0.2487, + "num_input_tokens_seen": 9741368, + "step": 12770 + }, + { + "epoch": 26.55925155925156, + "grad_norm": 0.0004886838141828775, + "learning_rate": 0.2306324735958199, + "loss": 0.2509, + "num_input_tokens_seen": 9745144, + "step": 12775 + }, + { + "epoch": 26.56964656964657, + "grad_norm": 0.0005577904521487653, + "learning_rate": 0.23058279691568362, + "loss": 0.2528, + "num_input_tokens_seen": 9748984, + "step": 12780 + }, + { + "epoch": 26.58004158004158, + "grad_norm": 0.00034506028168834746, + "learning_rate": 0.23053310780866745, + "loss": 0.2582, + "num_input_tokens_seen": 9752728, + "step": 12785 + }, + { + "epoch": 26.59043659043659, + "grad_norm": 0.00026751452242024243, + "learning_rate": 0.23048340628243397, + "loss": 0.2529, + "num_input_tokens_seen": 9756568, + "step": 12790 + }, + { + "epoch": 26.6008316008316, + "grad_norm": 0.00030433418578468263, + "learning_rate": 0.23043369234464783, + "loss": 0.2541, + "num_input_tokens_seen": 9760440, + "step": 12795 + }, + { + "epoch": 26.611226611226613, + "grad_norm": 0.00019172928296029568, + "learning_rate": 0.2303839660029755, + "loss": 0.2621, + "num_input_tokens_seen": 9764344, + "step": 12800 + }, + { + "epoch": 26.611226611226613, + "eval_loss": 0.24915602803230286, + "eval_runtime": 13.5127, + "eval_samples_per_second": 63.348, + "eval_steps_per_second": 15.837, + "num_input_tokens_seen": 9764344, + "step": 12800 + }, + { + "epoch": 26.62162162162162, + "grad_norm": 0.00030297087505459785, + "learning_rate": 0.23033422726508548, + "loss": 0.2427, + "num_input_tokens_seen": 9768280, + "step": 12805 + }, + { + "epoch": 26.632016632016633, + "grad_norm": 0.0012150837574154139, + "learning_rate": 0.23028447613864808, + "loss": 0.26, + "num_input_tokens_seen": 9772024, + "step": 12810 + }, + { + "epoch": 26.64241164241164, + "grad_norm": 0.0001291695807594806, + "learning_rate": 0.2302347126313355, + "loss": 0.276, + "num_input_tokens_seen": 9775960, + "step": 12815 + }, + { + "epoch": 26.652806652806653, + "grad_norm": 0.0006311475881375372, + "learning_rate": 0.23018493675082197, + "loss": 0.2812, + "num_input_tokens_seen": 9779672, + "step": 12820 + }, + { + "epoch": 26.66320166320166, + "grad_norm": 0.0001517148339189589, + "learning_rate": 0.2301351485047835, + "loss": 0.2699, + "num_input_tokens_seen": 9783512, + "step": 12825 + }, + { + "epoch": 26.673596673596673, + "grad_norm": 0.00045714061707258224, + "learning_rate": 0.23008534790089813, + "loss": 0.2958, + "num_input_tokens_seen": 9787256, + "step": 12830 + }, + { + "epoch": 26.683991683991685, + "grad_norm": 0.0007368313963524997, + "learning_rate": 0.2300355349468457, + "loss": 0.2443, + "num_input_tokens_seen": 9791032, + "step": 12835 + }, + { + "epoch": 26.694386694386694, + "grad_norm": 0.0004565578419715166, + "learning_rate": 0.22998570965030793, + "loss": 0.2571, + "num_input_tokens_seen": 9794808, + "step": 12840 + }, + { + "epoch": 26.704781704781706, + "grad_norm": 9.455796680413187e-05, + "learning_rate": 0.22993587201896862, + "loss": 0.2644, + "num_input_tokens_seen": 9798616, + "step": 12845 + }, + { + "epoch": 26.715176715176714, + "grad_norm": 0.0002607906353659928, + "learning_rate": 0.2298860220605133, + "loss": 0.2785, + "num_input_tokens_seen": 9802584, + "step": 12850 + }, + { + "epoch": 26.725571725571726, + "grad_norm": 0.0001764680491760373, + "learning_rate": 0.22983615978262942, + "loss": 0.2805, + "num_input_tokens_seen": 9806488, + "step": 12855 + }, + { + "epoch": 26.735966735966738, + "grad_norm": 0.0003495899436529726, + "learning_rate": 0.22978628519300648, + "loss": 0.2658, + "num_input_tokens_seen": 9810520, + "step": 12860 + }, + { + "epoch": 26.746361746361746, + "grad_norm": 0.0006536575965583324, + "learning_rate": 0.22973639829933568, + "loss": 0.2675, + "num_input_tokens_seen": 9814328, + "step": 12865 + }, + { + "epoch": 26.756756756756758, + "grad_norm": 0.00024923565797507763, + "learning_rate": 0.22968649910931027, + "loss": 0.2925, + "num_input_tokens_seen": 9818232, + "step": 12870 + }, + { + "epoch": 26.767151767151766, + "grad_norm": 0.0004650003102142364, + "learning_rate": 0.22963658763062528, + "loss": 0.2707, + "num_input_tokens_seen": 9822072, + "step": 12875 + }, + { + "epoch": 26.777546777546778, + "grad_norm": 0.0009208290721289814, + "learning_rate": 0.22958666387097765, + "loss": 0.2629, + "num_input_tokens_seen": 9825816, + "step": 12880 + }, + { + "epoch": 26.787941787941786, + "grad_norm": 9.822161518968642e-05, + "learning_rate": 0.22953672783806633, + "loss": 0.2524, + "num_input_tokens_seen": 9829528, + "step": 12885 + }, + { + "epoch": 26.7983367983368, + "grad_norm": 0.0002308362745679915, + "learning_rate": 0.22948677953959207, + "loss": 0.297, + "num_input_tokens_seen": 9833400, + "step": 12890 + }, + { + "epoch": 26.80873180873181, + "grad_norm": 0.0004099719226360321, + "learning_rate": 0.2294368189832575, + "loss": 0.2599, + "num_input_tokens_seen": 9837176, + "step": 12895 + }, + { + "epoch": 26.81912681912682, + "grad_norm": 0.00016004088683985174, + "learning_rate": 0.2293868461767672, + "loss": 0.2691, + "num_input_tokens_seen": 9840856, + "step": 12900 + }, + { + "epoch": 26.82952182952183, + "grad_norm": 0.00046312747872434556, + "learning_rate": 0.22933686112782758, + "loss": 0.2744, + "num_input_tokens_seen": 9844792, + "step": 12905 + }, + { + "epoch": 26.83991683991684, + "grad_norm": 0.0005662898765876889, + "learning_rate": 0.22928686384414698, + "loss": 0.2773, + "num_input_tokens_seen": 9848600, + "step": 12910 + }, + { + "epoch": 26.85031185031185, + "grad_norm": 0.00025487018865533173, + "learning_rate": 0.22923685433343552, + "loss": 0.2741, + "num_input_tokens_seen": 9852248, + "step": 12915 + }, + { + "epoch": 26.86070686070686, + "grad_norm": 0.00038627488538622856, + "learning_rate": 0.22918683260340542, + "loss": 0.279, + "num_input_tokens_seen": 9856216, + "step": 12920 + }, + { + "epoch": 26.87110187110187, + "grad_norm": 0.00028804840985685587, + "learning_rate": 0.2291367986617706, + "loss": 0.2785, + "num_input_tokens_seen": 9859992, + "step": 12925 + }, + { + "epoch": 26.881496881496883, + "grad_norm": 0.0003801829880103469, + "learning_rate": 0.22908675251624697, + "loss": 0.2818, + "num_input_tokens_seen": 9863992, + "step": 12930 + }, + { + "epoch": 26.89189189189189, + "grad_norm": 8.225213241530582e-05, + "learning_rate": 0.22903669417455216, + "loss": 0.2544, + "num_input_tokens_seen": 9867928, + "step": 12935 + }, + { + "epoch": 26.902286902286903, + "grad_norm": 0.00010724605090217665, + "learning_rate": 0.22898662364440592, + "loss": 0.2732, + "num_input_tokens_seen": 9871672, + "step": 12940 + }, + { + "epoch": 26.91268191268191, + "grad_norm": 0.00022391023230738938, + "learning_rate": 0.2289365409335297, + "loss": 0.2939, + "num_input_tokens_seen": 9875448, + "step": 12945 + }, + { + "epoch": 26.923076923076923, + "grad_norm": 0.0007258779369294643, + "learning_rate": 0.2288864460496469, + "loss": 0.2779, + "num_input_tokens_seen": 9879256, + "step": 12950 + }, + { + "epoch": 26.933471933471935, + "grad_norm": 0.0004381911421660334, + "learning_rate": 0.22883633900048272, + "loss": 0.2625, + "num_input_tokens_seen": 9883000, + "step": 12955 + }, + { + "epoch": 26.943866943866944, + "grad_norm": 0.0005304153892211616, + "learning_rate": 0.2287862197937644, + "loss": 0.2793, + "num_input_tokens_seen": 9886680, + "step": 12960 + }, + { + "epoch": 26.954261954261955, + "grad_norm": 0.00025625419220887125, + "learning_rate": 0.2287360884372209, + "loss": 0.2455, + "num_input_tokens_seen": 9890584, + "step": 12965 + }, + { + "epoch": 26.964656964656964, + "grad_norm": 0.00018429012561682612, + "learning_rate": 0.22868594493858307, + "loss": 0.2798, + "num_input_tokens_seen": 9894360, + "step": 12970 + }, + { + "epoch": 26.975051975051976, + "grad_norm": 0.000454979483038187, + "learning_rate": 0.2286357893055837, + "loss": 0.2829, + "num_input_tokens_seen": 9898200, + "step": 12975 + }, + { + "epoch": 26.985446985446984, + "grad_norm": 0.0005798538913950324, + "learning_rate": 0.22858562154595746, + "loss": 0.2734, + "num_input_tokens_seen": 9902008, + "step": 12980 + }, + { + "epoch": 26.995841995841996, + "grad_norm": 0.0006047403439879417, + "learning_rate": 0.22853544166744078, + "loss": 0.2668, + "num_input_tokens_seen": 9905688, + "step": 12985 + }, + { + "epoch": 27.006237006237008, + "grad_norm": 0.0003344781289342791, + "learning_rate": 0.22848524967777206, + "loss": 0.26, + "num_input_tokens_seen": 9909512, + "step": 12990 + }, + { + "epoch": 27.016632016632016, + "grad_norm": 0.00031634216429665685, + "learning_rate": 0.22843504558469152, + "loss": 0.2729, + "num_input_tokens_seen": 9913160, + "step": 12995 + }, + { + "epoch": 27.027027027027028, + "grad_norm": 0.0003347745514474809, + "learning_rate": 0.2283848293959413, + "loss": 0.2696, + "num_input_tokens_seen": 9917064, + "step": 13000 + }, + { + "epoch": 27.027027027027028, + "eval_loss": 0.25639572739601135, + "eval_runtime": 13.4265, + "eval_samples_per_second": 63.755, + "eval_steps_per_second": 15.939, + "num_input_tokens_seen": 9917064, + "step": 13000 + }, + { + "epoch": 27.037422037422036, + "grad_norm": 0.0002382049133302644, + "learning_rate": 0.22833460111926532, + "loss": 0.2637, + "num_input_tokens_seen": 9920744, + "step": 13005 + }, + { + "epoch": 27.04781704781705, + "grad_norm": 0.0004500375653151423, + "learning_rate": 0.22828436076240946, + "loss": 0.2537, + "num_input_tokens_seen": 9924552, + "step": 13010 + }, + { + "epoch": 27.058212058212057, + "grad_norm": 0.0002097799879265949, + "learning_rate": 0.22823410833312135, + "loss": 0.2492, + "num_input_tokens_seen": 9928296, + "step": 13015 + }, + { + "epoch": 27.06860706860707, + "grad_norm": 0.00050159334205091, + "learning_rate": 0.2281838438391506, + "loss": 0.2639, + "num_input_tokens_seen": 9932136, + "step": 13020 + }, + { + "epoch": 27.07900207900208, + "grad_norm": 0.0001650533522479236, + "learning_rate": 0.22813356728824863, + "loss": 0.2499, + "num_input_tokens_seen": 9935784, + "step": 13025 + }, + { + "epoch": 27.08939708939709, + "grad_norm": 0.00020827454864047468, + "learning_rate": 0.2280832786881687, + "loss": 0.2349, + "num_input_tokens_seen": 9939464, + "step": 13030 + }, + { + "epoch": 27.0997920997921, + "grad_norm": 0.0002079051046166569, + "learning_rate": 0.22803297804666592, + "loss": 0.2995, + "num_input_tokens_seen": 9943336, + "step": 13035 + }, + { + "epoch": 27.11018711018711, + "grad_norm": 0.0003163835790473968, + "learning_rate": 0.22798266537149728, + "loss": 0.2871, + "num_input_tokens_seen": 9947112, + "step": 13040 + }, + { + "epoch": 27.12058212058212, + "grad_norm": 0.0006993371061980724, + "learning_rate": 0.22793234067042167, + "loss": 0.2769, + "num_input_tokens_seen": 9950792, + "step": 13045 + }, + { + "epoch": 27.13097713097713, + "grad_norm": 0.000580761581659317, + "learning_rate": 0.22788200395119979, + "loss": 0.2916, + "num_input_tokens_seen": 9954824, + "step": 13050 + }, + { + "epoch": 27.14137214137214, + "grad_norm": 0.0004978504148311913, + "learning_rate": 0.2278316552215942, + "loss": 0.2711, + "num_input_tokens_seen": 9958600, + "step": 13055 + }, + { + "epoch": 27.151767151767153, + "grad_norm": 0.0003250467998441309, + "learning_rate": 0.22778129448936918, + "loss": 0.2676, + "num_input_tokens_seen": 9962312, + "step": 13060 + }, + { + "epoch": 27.16216216216216, + "grad_norm": 0.00018614131840877235, + "learning_rate": 0.22773092176229118, + "loss": 0.2689, + "num_input_tokens_seen": 9966056, + "step": 13065 + }, + { + "epoch": 27.172557172557173, + "grad_norm": 7.626097794855013e-05, + "learning_rate": 0.22768053704812816, + "loss": 0.2613, + "num_input_tokens_seen": 9969928, + "step": 13070 + }, + { + "epoch": 27.18295218295218, + "grad_norm": 0.0005300771445035934, + "learning_rate": 0.22763014035465018, + "loss": 0.2463, + "num_input_tokens_seen": 9973736, + "step": 13075 + }, + { + "epoch": 27.193347193347194, + "grad_norm": 0.00015333088231272995, + "learning_rate": 0.22757973168962892, + "loss": 0.2422, + "num_input_tokens_seen": 9977448, + "step": 13080 + }, + { + "epoch": 27.203742203742205, + "grad_norm": 0.00033755411277525127, + "learning_rate": 0.22752931106083818, + "loss": 0.2873, + "num_input_tokens_seen": 9981512, + "step": 13085 + }, + { + "epoch": 27.214137214137214, + "grad_norm": 0.00017177635163534433, + "learning_rate": 0.22747887847605341, + "loss": 0.262, + "num_input_tokens_seen": 9985384, + "step": 13090 + }, + { + "epoch": 27.224532224532226, + "grad_norm": 0.0002843157562892884, + "learning_rate": 0.22742843394305184, + "loss": 0.2592, + "num_input_tokens_seen": 9989256, + "step": 13095 + }, + { + "epoch": 27.234927234927234, + "grad_norm": 0.0003438879793975502, + "learning_rate": 0.22737797746961272, + "loss": 0.2713, + "num_input_tokens_seen": 9993000, + "step": 13100 + }, + { + "epoch": 27.245322245322246, + "grad_norm": 0.0005375894252210855, + "learning_rate": 0.22732750906351712, + "loss": 0.2554, + "num_input_tokens_seen": 9996872, + "step": 13105 + }, + { + "epoch": 27.255717255717254, + "grad_norm": 0.00018800931866280735, + "learning_rate": 0.22727702873254785, + "loss": 0.2754, + "num_input_tokens_seen": 10000648, + "step": 13110 + }, + { + "epoch": 27.266112266112266, + "grad_norm": 0.0009014391689561307, + "learning_rate": 0.22722653648448968, + "loss": 0.2667, + "num_input_tokens_seen": 10004264, + "step": 13115 + }, + { + "epoch": 27.276507276507278, + "grad_norm": 0.0004581070097628981, + "learning_rate": 0.22717603232712902, + "loss": 0.2675, + "num_input_tokens_seen": 10008136, + "step": 13120 + }, + { + "epoch": 27.286902286902286, + "grad_norm": 0.00024369248421862721, + "learning_rate": 0.22712551626825436, + "loss": 0.2313, + "num_input_tokens_seen": 10012072, + "step": 13125 + }, + { + "epoch": 27.2972972972973, + "grad_norm": 0.00024073383247014135, + "learning_rate": 0.2270749883156559, + "loss": 0.3025, + "num_input_tokens_seen": 10015848, + "step": 13130 + }, + { + "epoch": 27.307692307692307, + "grad_norm": 0.0002042127016466111, + "learning_rate": 0.22702444847712563, + "loss": 0.2685, + "num_input_tokens_seen": 10019560, + "step": 13135 + }, + { + "epoch": 27.31808731808732, + "grad_norm": 0.0004922857624478638, + "learning_rate": 0.22697389676045743, + "loss": 0.2796, + "num_input_tokens_seen": 10023496, + "step": 13140 + }, + { + "epoch": 27.328482328482327, + "grad_norm": 0.0010805350029841065, + "learning_rate": 0.22692333317344704, + "loss": 0.2942, + "num_input_tokens_seen": 10027304, + "step": 13145 + }, + { + "epoch": 27.33887733887734, + "grad_norm": 0.0006932936958037317, + "learning_rate": 0.22687275772389198, + "loss": 0.2934, + "num_input_tokens_seen": 10031048, + "step": 13150 + }, + { + "epoch": 27.34927234927235, + "grad_norm": 8.872978651197627e-05, + "learning_rate": 0.22682217041959168, + "loss": 0.2689, + "num_input_tokens_seen": 10034792, + "step": 13155 + }, + { + "epoch": 27.35966735966736, + "grad_norm": 0.0003817516262643039, + "learning_rate": 0.2267715712683473, + "loss": 0.2648, + "num_input_tokens_seen": 10038568, + "step": 13160 + }, + { + "epoch": 27.37006237006237, + "grad_norm": 0.0007749811629764736, + "learning_rate": 0.22672096027796182, + "loss": 0.2695, + "num_input_tokens_seen": 10042280, + "step": 13165 + }, + { + "epoch": 27.38045738045738, + "grad_norm": 0.00026218913262709975, + "learning_rate": 0.22667033745624016, + "loss": 0.2647, + "num_input_tokens_seen": 10046088, + "step": 13170 + }, + { + "epoch": 27.39085239085239, + "grad_norm": 0.0003157384635414928, + "learning_rate": 0.22661970281098895, + "loss": 0.2825, + "num_input_tokens_seen": 10049864, + "step": 13175 + }, + { + "epoch": 27.401247401247403, + "grad_norm": 0.00024935463443398476, + "learning_rate": 0.22656905635001667, + "loss": 0.2646, + "num_input_tokens_seen": 10053576, + "step": 13180 + }, + { + "epoch": 27.41164241164241, + "grad_norm": 0.000323864835081622, + "learning_rate": 0.2265183980811337, + "loss": 0.2819, + "num_input_tokens_seen": 10057320, + "step": 13185 + }, + { + "epoch": 27.422037422037423, + "grad_norm": 0.0004008161195088178, + "learning_rate": 0.22646772801215218, + "loss": 0.2521, + "num_input_tokens_seen": 10061000, + "step": 13190 + }, + { + "epoch": 27.43243243243243, + "grad_norm": 0.0002013692574109882, + "learning_rate": 0.22641704615088598, + "loss": 0.237, + "num_input_tokens_seen": 10064712, + "step": 13195 + }, + { + "epoch": 27.442827442827443, + "grad_norm": 0.00044108665315434337, + "learning_rate": 0.22636635250515103, + "loss": 0.2466, + "num_input_tokens_seen": 10068520, + "step": 13200 + }, + { + "epoch": 27.442827442827443, + "eval_loss": 0.25269845128059387, + "eval_runtime": 13.4039, + "eval_samples_per_second": 63.862, + "eval_steps_per_second": 15.966, + "num_input_tokens_seen": 10068520, + "step": 13200 + }, + { + "epoch": 27.453222453222452, + "grad_norm": 0.00015968404477462173, + "learning_rate": 0.2263156470827648, + "loss": 0.267, + "num_input_tokens_seen": 10072328, + "step": 13205 + }, + { + "epoch": 27.463617463617464, + "grad_norm": 0.0003435423714108765, + "learning_rate": 0.22626492989154678, + "loss": 0.2779, + "num_input_tokens_seen": 10076136, + "step": 13210 + }, + { + "epoch": 27.474012474012476, + "grad_norm": 0.0005049800965934992, + "learning_rate": 0.22621420093931813, + "loss": 0.2752, + "num_input_tokens_seen": 10079912, + "step": 13215 + }, + { + "epoch": 27.484407484407484, + "grad_norm": 0.0006863938760943711, + "learning_rate": 0.22616346023390194, + "loss": 0.2753, + "num_input_tokens_seen": 10083816, + "step": 13220 + }, + { + "epoch": 27.494802494802496, + "grad_norm": 0.00016316442633979023, + "learning_rate": 0.22611270778312306, + "loss": 0.2788, + "num_input_tokens_seen": 10087560, + "step": 13225 + }, + { + "epoch": 27.505197505197504, + "grad_norm": 5.5427837651222944e-05, + "learning_rate": 0.2260619435948081, + "loss": 0.2638, + "num_input_tokens_seen": 10091432, + "step": 13230 + }, + { + "epoch": 27.515592515592516, + "grad_norm": 8.78640275914222e-05, + "learning_rate": 0.22601116767678567, + "loss": 0.2536, + "num_input_tokens_seen": 10095368, + "step": 13235 + }, + { + "epoch": 27.525987525987524, + "grad_norm": 0.00015075386909302324, + "learning_rate": 0.2259603800368859, + "loss": 0.2693, + "num_input_tokens_seen": 10099176, + "step": 13240 + }, + { + "epoch": 27.536382536382536, + "grad_norm": 0.00029906249255873263, + "learning_rate": 0.22590958068294098, + "loss": 0.2655, + "num_input_tokens_seen": 10102824, + "step": 13245 + }, + { + "epoch": 27.546777546777548, + "grad_norm": 0.00022729103511665016, + "learning_rate": 0.22585876962278478, + "loss": 0.2421, + "num_input_tokens_seen": 10106696, + "step": 13250 + }, + { + "epoch": 27.557172557172557, + "grad_norm": 0.0001592343469383195, + "learning_rate": 0.22580794686425298, + "loss": 0.2426, + "num_input_tokens_seen": 10110568, + "step": 13255 + }, + { + "epoch": 27.56756756756757, + "grad_norm": 0.0001673367660259828, + "learning_rate": 0.22575711241518312, + "loss": 0.2883, + "num_input_tokens_seen": 10114376, + "step": 13260 + }, + { + "epoch": 27.577962577962577, + "grad_norm": 0.0004015630984213203, + "learning_rate": 0.22570626628341453, + "loss": 0.2657, + "num_input_tokens_seen": 10118152, + "step": 13265 + }, + { + "epoch": 27.58835758835759, + "grad_norm": 9.43072373047471e-05, + "learning_rate": 0.22565540847678828, + "loss": 0.2819, + "num_input_tokens_seen": 10121896, + "step": 13270 + }, + { + "epoch": 27.598752598752597, + "grad_norm": 0.00030626184889115393, + "learning_rate": 0.2256045390031473, + "loss": 0.2721, + "num_input_tokens_seen": 10125608, + "step": 13275 + }, + { + "epoch": 27.60914760914761, + "grad_norm": 0.00025307227042503655, + "learning_rate": 0.22555365787033627, + "loss": 0.2644, + "num_input_tokens_seen": 10129512, + "step": 13280 + }, + { + "epoch": 27.61954261954262, + "grad_norm": 0.00017517949163448066, + "learning_rate": 0.22550276508620173, + "loss": 0.2628, + "num_input_tokens_seen": 10133352, + "step": 13285 + }, + { + "epoch": 27.62993762993763, + "grad_norm": 6.868541822768748e-05, + "learning_rate": 0.22545186065859202, + "loss": 0.2735, + "num_input_tokens_seen": 10137064, + "step": 13290 + }, + { + "epoch": 27.64033264033264, + "grad_norm": 0.0003172597207594663, + "learning_rate": 0.2254009445953572, + "loss": 0.2764, + "num_input_tokens_seen": 10140968, + "step": 13295 + }, + { + "epoch": 27.65072765072765, + "grad_norm": 0.00017866762937046587, + "learning_rate": 0.22535001690434917, + "loss": 0.2689, + "num_input_tokens_seen": 10144872, + "step": 13300 + }, + { + "epoch": 27.66112266112266, + "grad_norm": 0.0004021819040644914, + "learning_rate": 0.22529907759342163, + "loss": 0.2643, + "num_input_tokens_seen": 10148680, + "step": 13305 + }, + { + "epoch": 27.671517671517673, + "grad_norm": 0.00021828350145369768, + "learning_rate": 0.22524812667043007, + "loss": 0.2824, + "num_input_tokens_seen": 10152712, + "step": 13310 + }, + { + "epoch": 27.68191268191268, + "grad_norm": 0.0008664872730150819, + "learning_rate": 0.22519716414323177, + "loss": 0.2692, + "num_input_tokens_seen": 10156456, + "step": 13315 + }, + { + "epoch": 27.692307692307693, + "grad_norm": 9.309778397437185e-05, + "learning_rate": 0.22514619001968567, + "loss": 0.2727, + "num_input_tokens_seen": 10160168, + "step": 13320 + }, + { + "epoch": 27.7027027027027, + "grad_norm": 0.0008709369576536119, + "learning_rate": 0.2250952043076528, + "loss": 0.2531, + "num_input_tokens_seen": 10164104, + "step": 13325 + }, + { + "epoch": 27.713097713097714, + "grad_norm": 0.0003518385929055512, + "learning_rate": 0.2250442070149957, + "loss": 0.2565, + "num_input_tokens_seen": 10167720, + "step": 13330 + }, + { + "epoch": 27.723492723492722, + "grad_norm": 0.0005521838320419192, + "learning_rate": 0.22499319814957885, + "loss": 0.2522, + "num_input_tokens_seen": 10171400, + "step": 13335 + }, + { + "epoch": 27.733887733887734, + "grad_norm": 6.336794467642903e-05, + "learning_rate": 0.2249421777192684, + "loss": 0.3081, + "num_input_tokens_seen": 10175208, + "step": 13340 + }, + { + "epoch": 27.744282744282746, + "grad_norm": 0.0005448810989037156, + "learning_rate": 0.22489114573193236, + "loss": 0.2875, + "num_input_tokens_seen": 10179112, + "step": 13345 + }, + { + "epoch": 27.754677754677754, + "grad_norm": 6.213696178747341e-05, + "learning_rate": 0.2248401021954405, + "loss": 0.2788, + "num_input_tokens_seen": 10182984, + "step": 13350 + }, + { + "epoch": 27.765072765072766, + "grad_norm": 0.00028042099438607693, + "learning_rate": 0.22478904711766443, + "loss": 0.2714, + "num_input_tokens_seen": 10186920, + "step": 13355 + }, + { + "epoch": 27.775467775467774, + "grad_norm": 0.00021156042930670083, + "learning_rate": 0.22473798050647734, + "loss": 0.2738, + "num_input_tokens_seen": 10190824, + "step": 13360 + }, + { + "epoch": 27.785862785862786, + "grad_norm": 9.774912905413657e-05, + "learning_rate": 0.22468690236975453, + "loss": 0.2632, + "num_input_tokens_seen": 10194664, + "step": 13365 + }, + { + "epoch": 27.796257796257795, + "grad_norm": 0.0005764098023064435, + "learning_rate": 0.22463581271537272, + "loss": 0.2857, + "num_input_tokens_seen": 10198440, + "step": 13370 + }, + { + "epoch": 27.806652806652806, + "grad_norm": 0.0001233753573615104, + "learning_rate": 0.22458471155121076, + "loss": 0.2803, + "num_input_tokens_seen": 10202280, + "step": 13375 + }, + { + "epoch": 27.81704781704782, + "grad_norm": 0.00018434102821629494, + "learning_rate": 0.2245335988851489, + "loss": 0.2679, + "num_input_tokens_seen": 10206056, + "step": 13380 + }, + { + "epoch": 27.827442827442827, + "grad_norm": 0.0005305905360728502, + "learning_rate": 0.2244824747250695, + "loss": 0.2676, + "num_input_tokens_seen": 10209800, + "step": 13385 + }, + { + "epoch": 27.83783783783784, + "grad_norm": 0.00022734608501195908, + "learning_rate": 0.22443133907885646, + "loss": 0.2796, + "num_input_tokens_seen": 10213576, + "step": 13390 + }, + { + "epoch": 27.848232848232847, + "grad_norm": 0.0002604887413326651, + "learning_rate": 0.22438019195439557, + "loss": 0.2626, + "num_input_tokens_seen": 10217352, + "step": 13395 + }, + { + "epoch": 27.85862785862786, + "grad_norm": 0.00034578944905661047, + "learning_rate": 0.22432903335957435, + "loss": 0.283, + "num_input_tokens_seen": 10221224, + "step": 13400 + }, + { + "epoch": 27.85862785862786, + "eval_loss": 0.2568046748638153, + "eval_runtime": 13.4539, + "eval_samples_per_second": 63.625, + "eval_steps_per_second": 15.906, + "num_input_tokens_seen": 10221224, + "step": 13400 + }, + { + "epoch": 27.86902286902287, + "grad_norm": 0.0003184888046234846, + "learning_rate": 0.22427786330228214, + "loss": 0.2681, + "num_input_tokens_seen": 10225096, + "step": 13405 + }, + { + "epoch": 27.87941787941788, + "grad_norm": 0.00016574117762502283, + "learning_rate": 0.22422668179040997, + "loss": 0.2694, + "num_input_tokens_seen": 10228840, + "step": 13410 + }, + { + "epoch": 27.88981288981289, + "grad_norm": 0.0006649466813541949, + "learning_rate": 0.2241754888318507, + "loss": 0.2347, + "num_input_tokens_seen": 10232648, + "step": 13415 + }, + { + "epoch": 27.9002079002079, + "grad_norm": 0.0010515139438211918, + "learning_rate": 0.22412428443449886, + "loss": 0.2817, + "num_input_tokens_seen": 10236680, + "step": 13420 + }, + { + "epoch": 27.91060291060291, + "grad_norm": 0.00011317091411910951, + "learning_rate": 0.22407306860625087, + "loss": 0.2779, + "num_input_tokens_seen": 10240392, + "step": 13425 + }, + { + "epoch": 27.92099792099792, + "grad_norm": 0.00038695023977197707, + "learning_rate": 0.22402184135500483, + "loss": 0.2738, + "num_input_tokens_seen": 10244232, + "step": 13430 + }, + { + "epoch": 27.93139293139293, + "grad_norm": 0.0007046873797662556, + "learning_rate": 0.22397060268866067, + "loss": 0.2757, + "num_input_tokens_seen": 10247976, + "step": 13435 + }, + { + "epoch": 27.941787941787943, + "grad_norm": 0.000474357046186924, + "learning_rate": 0.22391935261511994, + "loss": 0.2508, + "num_input_tokens_seen": 10251784, + "step": 13440 + }, + { + "epoch": 27.95218295218295, + "grad_norm": 0.00031123822554945946, + "learning_rate": 0.22386809114228615, + "loss": 0.2569, + "num_input_tokens_seen": 10255656, + "step": 13445 + }, + { + "epoch": 27.962577962577964, + "grad_norm": 0.00018421078857500106, + "learning_rate": 0.22381681827806446, + "loss": 0.2427, + "num_input_tokens_seen": 10259400, + "step": 13450 + }, + { + "epoch": 27.972972972972972, + "grad_norm": 4.176851871307008e-05, + "learning_rate": 0.22376553403036173, + "loss": 0.2832, + "num_input_tokens_seen": 10263272, + "step": 13455 + }, + { + "epoch": 27.983367983367984, + "grad_norm": 0.0004083602107129991, + "learning_rate": 0.22371423840708662, + "loss": 0.255, + "num_input_tokens_seen": 10267144, + "step": 13460 + }, + { + "epoch": 27.993762993762992, + "grad_norm": 0.00045082971337251365, + "learning_rate": 0.22366293141614962, + "loss": 0.277, + "num_input_tokens_seen": 10271048, + "step": 13465 + }, + { + "epoch": 28.004158004158004, + "grad_norm": 0.0003402531147003174, + "learning_rate": 0.22361161306546287, + "loss": 0.2765, + "num_input_tokens_seen": 10274840, + "step": 13470 + }, + { + "epoch": 28.014553014553016, + "grad_norm": 0.0006915408303029835, + "learning_rate": 0.22356028336294037, + "loss": 0.2627, + "num_input_tokens_seen": 10278488, + "step": 13475 + }, + { + "epoch": 28.024948024948024, + "grad_norm": 0.00031890012905932963, + "learning_rate": 0.2235089423164977, + "loss": 0.2674, + "num_input_tokens_seen": 10282328, + "step": 13480 + }, + { + "epoch": 28.035343035343036, + "grad_norm": 0.00023372222494799644, + "learning_rate": 0.22345758993405243, + "loss": 0.2429, + "num_input_tokens_seen": 10286008, + "step": 13485 + }, + { + "epoch": 28.045738045738045, + "grad_norm": 0.00025254394859075546, + "learning_rate": 0.2234062262235236, + "loss": 0.2541, + "num_input_tokens_seen": 10289720, + "step": 13490 + }, + { + "epoch": 28.056133056133056, + "grad_norm": 6.45907421130687e-05, + "learning_rate": 0.22335485119283222, + "loss": 0.2674, + "num_input_tokens_seen": 10293400, + "step": 13495 + }, + { + "epoch": 28.066528066528065, + "grad_norm": 0.0006500897579826415, + "learning_rate": 0.22330346484990093, + "loss": 0.2743, + "num_input_tokens_seen": 10297144, + "step": 13500 + }, + { + "epoch": 28.076923076923077, + "grad_norm": 0.00016336972475983202, + "learning_rate": 0.22325206720265425, + "loss": 0.2645, + "num_input_tokens_seen": 10301208, + "step": 13505 + }, + { + "epoch": 28.08731808731809, + "grad_norm": 0.00020031094027217478, + "learning_rate": 0.2232006582590182, + "loss": 0.2799, + "num_input_tokens_seen": 10305016, + "step": 13510 + }, + { + "epoch": 28.097713097713097, + "grad_norm": 0.00023100509133655578, + "learning_rate": 0.22314923802692077, + "loss": 0.2546, + "num_input_tokens_seen": 10308792, + "step": 13515 + }, + { + "epoch": 28.10810810810811, + "grad_norm": 0.000627058616373688, + "learning_rate": 0.22309780651429156, + "loss": 0.2682, + "num_input_tokens_seen": 10312600, + "step": 13520 + }, + { + "epoch": 28.118503118503117, + "grad_norm": 0.00032175221713259816, + "learning_rate": 0.22304636372906203, + "loss": 0.2452, + "num_input_tokens_seen": 10316408, + "step": 13525 + }, + { + "epoch": 28.12889812889813, + "grad_norm": 0.0003232719318475574, + "learning_rate": 0.22299490967916522, + "loss": 0.2879, + "num_input_tokens_seen": 10320408, + "step": 13530 + }, + { + "epoch": 28.13929313929314, + "grad_norm": 0.00021448101324494928, + "learning_rate": 0.22294344437253602, + "loss": 0.2762, + "num_input_tokens_seen": 10324280, + "step": 13535 + }, + { + "epoch": 28.14968814968815, + "grad_norm": 0.0006172165740281343, + "learning_rate": 0.22289196781711101, + "loss": 0.2744, + "num_input_tokens_seen": 10328248, + "step": 13540 + }, + { + "epoch": 28.16008316008316, + "grad_norm": 0.00017873563047032803, + "learning_rate": 0.2228404800208286, + "loss": 0.2747, + "num_input_tokens_seen": 10331896, + "step": 13545 + }, + { + "epoch": 28.17047817047817, + "grad_norm": 0.00047356702270917594, + "learning_rate": 0.22278898099162875, + "loss": 0.2649, + "num_input_tokens_seen": 10335640, + "step": 13550 + }, + { + "epoch": 28.18087318087318, + "grad_norm": 0.0002550750505179167, + "learning_rate": 0.22273747073745337, + "loss": 0.2821, + "num_input_tokens_seen": 10339608, + "step": 13555 + }, + { + "epoch": 28.19126819126819, + "grad_norm": 0.0002169163926737383, + "learning_rate": 0.22268594926624588, + "loss": 0.261, + "num_input_tokens_seen": 10343384, + "step": 13560 + }, + { + "epoch": 28.2016632016632, + "grad_norm": 0.00020018444047309458, + "learning_rate": 0.22263441658595162, + "loss": 0.2755, + "num_input_tokens_seen": 10347288, + "step": 13565 + }, + { + "epoch": 28.212058212058214, + "grad_norm": 0.00021457232651300728, + "learning_rate": 0.2225828727045175, + "loss": 0.257, + "num_input_tokens_seen": 10351064, + "step": 13570 + }, + { + "epoch": 28.222453222453222, + "grad_norm": 0.000828791584353894, + "learning_rate": 0.22253131762989228, + "loss": 0.2812, + "num_input_tokens_seen": 10354904, + "step": 13575 + }, + { + "epoch": 28.232848232848234, + "grad_norm": 0.00044811420957557857, + "learning_rate": 0.2224797513700264, + "loss": 0.2427, + "num_input_tokens_seen": 10358680, + "step": 13580 + }, + { + "epoch": 28.243243243243242, + "grad_norm": 8.97665013326332e-05, + "learning_rate": 0.22242817393287204, + "loss": 0.2806, + "num_input_tokens_seen": 10362520, + "step": 13585 + }, + { + "epoch": 28.253638253638254, + "grad_norm": 0.0005119292181916535, + "learning_rate": 0.22237658532638305, + "loss": 0.2634, + "num_input_tokens_seen": 10366392, + "step": 13590 + }, + { + "epoch": 28.264033264033262, + "grad_norm": 0.0004910401185043156, + "learning_rate": 0.22232498555851513, + "loss": 0.2645, + "num_input_tokens_seen": 10370136, + "step": 13595 + }, + { + "epoch": 28.274428274428274, + "grad_norm": 9.295009658671916e-05, + "learning_rate": 0.22227337463722546, + "loss": 0.2742, + "num_input_tokens_seen": 10373912, + "step": 13600 + }, + { + "epoch": 28.274428274428274, + "eval_loss": 0.24840591847896576, + "eval_runtime": 13.4017, + "eval_samples_per_second": 63.873, + "eval_steps_per_second": 15.968, + "num_input_tokens_seen": 10373912, + "step": 13600 + }, + { + "epoch": 28.284823284823286, + "grad_norm": 0.0004236153035890311, + "learning_rate": 0.2222217525704732, + "loss": 0.268, + "num_input_tokens_seen": 10377688, + "step": 13605 + }, + { + "epoch": 28.295218295218294, + "grad_norm": 0.00020817089534830302, + "learning_rate": 0.22217011936621908, + "loss": 0.278, + "num_input_tokens_seen": 10381464, + "step": 13610 + }, + { + "epoch": 28.305613305613306, + "grad_norm": 0.0009431368671357632, + "learning_rate": 0.22211847503242566, + "loss": 0.2661, + "num_input_tokens_seen": 10385336, + "step": 13615 + }, + { + "epoch": 28.316008316008315, + "grad_norm": 0.00024732365272939205, + "learning_rate": 0.22206681957705704, + "loss": 0.2672, + "num_input_tokens_seen": 10389208, + "step": 13620 + }, + { + "epoch": 28.326403326403327, + "grad_norm": 0.00034605455584824085, + "learning_rate": 0.2220151530080792, + "loss": 0.2895, + "num_input_tokens_seen": 10393048, + "step": 13625 + }, + { + "epoch": 28.33679833679834, + "grad_norm": 0.00024305492115672678, + "learning_rate": 0.2219634753334598, + "loss": 0.2513, + "num_input_tokens_seen": 10396888, + "step": 13630 + }, + { + "epoch": 28.347193347193347, + "grad_norm": 0.00016207690350711346, + "learning_rate": 0.22191178656116817, + "loss": 0.2446, + "num_input_tokens_seen": 10400696, + "step": 13635 + }, + { + "epoch": 28.35758835758836, + "grad_norm": 0.0005197438294999301, + "learning_rate": 0.2218600866991753, + "loss": 0.2822, + "num_input_tokens_seen": 10404504, + "step": 13640 + }, + { + "epoch": 28.367983367983367, + "grad_norm": 0.0002536851679906249, + "learning_rate": 0.221808375755454, + "loss": 0.2358, + "num_input_tokens_seen": 10408472, + "step": 13645 + }, + { + "epoch": 28.37837837837838, + "grad_norm": 0.00018766659195534885, + "learning_rate": 0.22175665373797881, + "loss": 0.3053, + "num_input_tokens_seen": 10412344, + "step": 13650 + }, + { + "epoch": 28.388773388773387, + "grad_norm": 0.0006162977661006153, + "learning_rate": 0.22170492065472583, + "loss": 0.2757, + "num_input_tokens_seen": 10416248, + "step": 13655 + }, + { + "epoch": 28.3991683991684, + "grad_norm": 0.0006687522400170565, + "learning_rate": 0.221653176513673, + "loss": 0.2617, + "num_input_tokens_seen": 10420056, + "step": 13660 + }, + { + "epoch": 28.40956340956341, + "grad_norm": 0.0001514917385065928, + "learning_rate": 0.2216014213227999, + "loss": 0.2663, + "num_input_tokens_seen": 10423800, + "step": 13665 + }, + { + "epoch": 28.41995841995842, + "grad_norm": 3.96703981095925e-05, + "learning_rate": 0.22154965509008784, + "loss": 0.25, + "num_input_tokens_seen": 10427544, + "step": 13670 + }, + { + "epoch": 28.43035343035343, + "grad_norm": 0.00032525864662602544, + "learning_rate": 0.2214978778235198, + "loss": 0.2847, + "num_input_tokens_seen": 10431320, + "step": 13675 + }, + { + "epoch": 28.44074844074844, + "grad_norm": 0.0009113075793720782, + "learning_rate": 0.2214460895310805, + "loss": 0.2556, + "num_input_tokens_seen": 10435192, + "step": 13680 + }, + { + "epoch": 28.45114345114345, + "grad_norm": 0.0002059585676761344, + "learning_rate": 0.22139429022075635, + "loss": 0.2411, + "num_input_tokens_seen": 10438872, + "step": 13685 + }, + { + "epoch": 28.46153846153846, + "grad_norm": 0.0006226692348718643, + "learning_rate": 0.22134247990053546, + "loss": 0.2778, + "num_input_tokens_seen": 10442744, + "step": 13690 + }, + { + "epoch": 28.471933471933472, + "grad_norm": 0.0007250407943502069, + "learning_rate": 0.2212906585784076, + "loss": 0.2591, + "num_input_tokens_seen": 10446552, + "step": 13695 + }, + { + "epoch": 28.482328482328484, + "grad_norm": 0.00031673398916609585, + "learning_rate": 0.22123882626236432, + "loss": 0.2693, + "num_input_tokens_seen": 10450392, + "step": 13700 + }, + { + "epoch": 28.492723492723492, + "grad_norm": 0.0001294183894060552, + "learning_rate": 0.2211869829603988, + "loss": 0.2741, + "num_input_tokens_seen": 10454200, + "step": 13705 + }, + { + "epoch": 28.503118503118504, + "grad_norm": 0.0002462098782416433, + "learning_rate": 0.22113512868050592, + "loss": 0.2515, + "num_input_tokens_seen": 10457944, + "step": 13710 + }, + { + "epoch": 28.513513513513512, + "grad_norm": 0.00043582136277109385, + "learning_rate": 0.2210832634306822, + "loss": 0.2284, + "num_input_tokens_seen": 10461816, + "step": 13715 + }, + { + "epoch": 28.523908523908524, + "grad_norm": 0.00022136872576083988, + "learning_rate": 0.22103138721892598, + "loss": 0.3063, + "num_input_tokens_seen": 10465816, + "step": 13720 + }, + { + "epoch": 28.534303534303533, + "grad_norm": 3.260079029132612e-05, + "learning_rate": 0.22097950005323724, + "loss": 0.2756, + "num_input_tokens_seen": 10469496, + "step": 13725 + }, + { + "epoch": 28.544698544698544, + "grad_norm": 6.137879972811788e-05, + "learning_rate": 0.22092760194161762, + "loss": 0.2795, + "num_input_tokens_seen": 10473368, + "step": 13730 + }, + { + "epoch": 28.555093555093556, + "grad_norm": 0.0002104137820424512, + "learning_rate": 0.2208756928920704, + "loss": 0.2812, + "num_input_tokens_seen": 10477144, + "step": 13735 + }, + { + "epoch": 28.565488565488565, + "grad_norm": 3.5347711673239246e-05, + "learning_rate": 0.22082377291260072, + "loss": 0.2711, + "num_input_tokens_seen": 10480856, + "step": 13740 + }, + { + "epoch": 28.575883575883577, + "grad_norm": 0.00022669402824249119, + "learning_rate": 0.2207718420112152, + "loss": 0.2506, + "num_input_tokens_seen": 10484696, + "step": 13745 + }, + { + "epoch": 28.586278586278585, + "grad_norm": 0.0003218489873688668, + "learning_rate": 0.22071990019592228, + "loss": 0.2412, + "num_input_tokens_seen": 10488472, + "step": 13750 + }, + { + "epoch": 28.596673596673597, + "grad_norm": 0.00044010457349941134, + "learning_rate": 0.22066794747473198, + "loss": 0.2697, + "num_input_tokens_seen": 10492312, + "step": 13755 + }, + { + "epoch": 28.60706860706861, + "grad_norm": 0.00035100351669825613, + "learning_rate": 0.2206159838556562, + "loss": 0.2526, + "num_input_tokens_seen": 10496216, + "step": 13760 + }, + { + "epoch": 28.617463617463617, + "grad_norm": 0.00087849295232445, + "learning_rate": 0.2205640093467082, + "loss": 0.283, + "num_input_tokens_seen": 10500024, + "step": 13765 + }, + { + "epoch": 28.62785862785863, + "grad_norm": 0.00043509266106411815, + "learning_rate": 0.22051202395590322, + "loss": 0.2847, + "num_input_tokens_seen": 10503800, + "step": 13770 + }, + { + "epoch": 28.638253638253637, + "grad_norm": 0.00039357872446998954, + "learning_rate": 0.22046002769125808, + "loss": 0.2867, + "num_input_tokens_seen": 10507608, + "step": 13775 + }, + { + "epoch": 28.64864864864865, + "grad_norm": 0.00011844849359476939, + "learning_rate": 0.2204080205607912, + "loss": 0.2858, + "num_input_tokens_seen": 10511608, + "step": 13780 + }, + { + "epoch": 28.659043659043657, + "grad_norm": 6.902393943164498e-05, + "learning_rate": 0.22035600257252272, + "loss": 0.2789, + "num_input_tokens_seen": 10515320, + "step": 13785 + }, + { + "epoch": 28.66943866943867, + "grad_norm": 0.0005170084768906236, + "learning_rate": 0.2203039737344745, + "loss": 0.2836, + "num_input_tokens_seen": 10519160, + "step": 13790 + }, + { + "epoch": 28.67983367983368, + "grad_norm": 4.670147973229177e-05, + "learning_rate": 0.22025193405467003, + "loss": 0.2647, + "num_input_tokens_seen": 10523000, + "step": 13795 + }, + { + "epoch": 28.69022869022869, + "grad_norm": 0.0003400059649720788, + "learning_rate": 0.2201998835411345, + "loss": 0.2789, + "num_input_tokens_seen": 10526808, + "step": 13800 + }, + { + "epoch": 28.69022869022869, + "eval_loss": 0.2496807724237442, + "eval_runtime": 13.4138, + "eval_samples_per_second": 63.815, + "eval_steps_per_second": 15.954, + "num_input_tokens_seen": 10526808, + "step": 13800 + }, + { + "epoch": 28.7006237006237, + "grad_norm": 8.334021549671888e-05, + "learning_rate": 0.22014782220189474, + "loss": 0.2616, + "num_input_tokens_seen": 10530584, + "step": 13805 + }, + { + "epoch": 28.71101871101871, + "grad_norm": 0.00017785215459298342, + "learning_rate": 0.2200957500449793, + "loss": 0.259, + "num_input_tokens_seen": 10534488, + "step": 13810 + }, + { + "epoch": 28.72141372141372, + "grad_norm": 0.0004352965042926371, + "learning_rate": 0.22004366707841827, + "loss": 0.2642, + "num_input_tokens_seen": 10538264, + "step": 13815 + }, + { + "epoch": 28.731808731808734, + "grad_norm": 0.0003644624666776508, + "learning_rate": 0.21999157331024358, + "loss": 0.2577, + "num_input_tokens_seen": 10541912, + "step": 13820 + }, + { + "epoch": 28.742203742203742, + "grad_norm": 0.0003521043690852821, + "learning_rate": 0.21993946874848871, + "loss": 0.2608, + "num_input_tokens_seen": 10545720, + "step": 13825 + }, + { + "epoch": 28.752598752598754, + "grad_norm": 0.0003447293129283935, + "learning_rate": 0.2198873534011888, + "loss": 0.2725, + "num_input_tokens_seen": 10549624, + "step": 13830 + }, + { + "epoch": 28.762993762993762, + "grad_norm": 0.00019379006698727608, + "learning_rate": 0.2198352272763808, + "loss": 0.2613, + "num_input_tokens_seen": 10553368, + "step": 13835 + }, + { + "epoch": 28.773388773388774, + "grad_norm": 0.00016407031216658652, + "learning_rate": 0.2197830903821031, + "loss": 0.2623, + "num_input_tokens_seen": 10557144, + "step": 13840 + }, + { + "epoch": 28.783783783783782, + "grad_norm": 0.0004928882699459791, + "learning_rate": 0.21973094272639598, + "loss": 0.2589, + "num_input_tokens_seen": 10560984, + "step": 13845 + }, + { + "epoch": 28.794178794178794, + "grad_norm": 0.00047189852921292186, + "learning_rate": 0.21967878431730117, + "loss": 0.2779, + "num_input_tokens_seen": 10564824, + "step": 13850 + }, + { + "epoch": 28.804573804573806, + "grad_norm": 0.00022804511536378413, + "learning_rate": 0.21962661516286217, + "loss": 0.2681, + "num_input_tokens_seen": 10568664, + "step": 13855 + }, + { + "epoch": 28.814968814968815, + "grad_norm": 5.3561681852443144e-05, + "learning_rate": 0.21957443527112414, + "loss": 0.2734, + "num_input_tokens_seen": 10572376, + "step": 13860 + }, + { + "epoch": 28.825363825363826, + "grad_norm": 0.00013434651191346347, + "learning_rate": 0.21952224465013384, + "loss": 0.2639, + "num_input_tokens_seen": 10576088, + "step": 13865 + }, + { + "epoch": 28.835758835758835, + "grad_norm": 0.0008785526151768863, + "learning_rate": 0.21947004330793976, + "loss": 0.2856, + "num_input_tokens_seen": 10579960, + "step": 13870 + }, + { + "epoch": 28.846153846153847, + "grad_norm": 0.0004586655995808542, + "learning_rate": 0.21941783125259198, + "loss": 0.2255, + "num_input_tokens_seen": 10583544, + "step": 13875 + }, + { + "epoch": 28.856548856548855, + "grad_norm": 0.0003975137078668922, + "learning_rate": 0.21936560849214226, + "loss": 0.2606, + "num_input_tokens_seen": 10587288, + "step": 13880 + }, + { + "epoch": 28.866943866943867, + "grad_norm": 0.00015181452909018844, + "learning_rate": 0.21931337503464404, + "loss": 0.2874, + "num_input_tokens_seen": 10591160, + "step": 13885 + }, + { + "epoch": 28.87733887733888, + "grad_norm": 0.00035508748260326684, + "learning_rate": 0.21926113088815233, + "loss": 0.2657, + "num_input_tokens_seen": 10595032, + "step": 13890 + }, + { + "epoch": 28.887733887733887, + "grad_norm": 0.0006310552125796676, + "learning_rate": 0.2192088760607238, + "loss": 0.2897, + "num_input_tokens_seen": 10598968, + "step": 13895 + }, + { + "epoch": 28.8981288981289, + "grad_norm": 0.0010681103449314833, + "learning_rate": 0.2191566105604169, + "loss": 0.2591, + "num_input_tokens_seen": 10602680, + "step": 13900 + }, + { + "epoch": 28.908523908523907, + "grad_norm": 0.0005205380730330944, + "learning_rate": 0.21910433439529153, + "loss": 0.2653, + "num_input_tokens_seen": 10606488, + "step": 13905 + }, + { + "epoch": 28.91891891891892, + "grad_norm": 0.00010886816744459793, + "learning_rate": 0.2190520475734094, + "loss": 0.2227, + "num_input_tokens_seen": 10610328, + "step": 13910 + }, + { + "epoch": 28.929313929313928, + "grad_norm": 0.00032817869214341044, + "learning_rate": 0.2189997501028338, + "loss": 0.2586, + "num_input_tokens_seen": 10614104, + "step": 13915 + }, + { + "epoch": 28.93970893970894, + "grad_norm": 0.00036502041621133685, + "learning_rate": 0.2189474419916296, + "loss": 0.2733, + "num_input_tokens_seen": 10617976, + "step": 13920 + }, + { + "epoch": 28.95010395010395, + "grad_norm": 0.0002529247140046209, + "learning_rate": 0.21889512324786342, + "loss": 0.2694, + "num_input_tokens_seen": 10621688, + "step": 13925 + }, + { + "epoch": 28.96049896049896, + "grad_norm": 0.00014868068683426827, + "learning_rate": 0.21884279387960345, + "loss": 0.2727, + "num_input_tokens_seen": 10625368, + "step": 13930 + }, + { + "epoch": 28.97089397089397, + "grad_norm": 0.000550108146853745, + "learning_rate": 0.2187904538949195, + "loss": 0.2802, + "num_input_tokens_seen": 10629112, + "step": 13935 + }, + { + "epoch": 28.98128898128898, + "grad_norm": 0.00043433523387648165, + "learning_rate": 0.2187381033018831, + "loss": 0.2756, + "num_input_tokens_seen": 10633048, + "step": 13940 + }, + { + "epoch": 28.991683991683992, + "grad_norm": 0.0002551070647314191, + "learning_rate": 0.2186857421085673, + "loss": 0.2707, + "num_input_tokens_seen": 10636824, + "step": 13945 + }, + { + "epoch": 29.002079002079004, + "grad_norm": 0.0006948256050236523, + "learning_rate": 0.21863337032304697, + "loss": 0.2716, + "num_input_tokens_seen": 10640672, + "step": 13950 + }, + { + "epoch": 29.012474012474012, + "grad_norm": 0.00031978514743968844, + "learning_rate": 0.21858098795339845, + "loss": 0.2705, + "num_input_tokens_seen": 10644512, + "step": 13955 + }, + { + "epoch": 29.022869022869024, + "grad_norm": 0.00045618126750923693, + "learning_rate": 0.21852859500769975, + "loss": 0.2534, + "num_input_tokens_seen": 10648384, + "step": 13960 + }, + { + "epoch": 29.033264033264032, + "grad_norm": 0.00010564591502770782, + "learning_rate": 0.21847619149403044, + "loss": 0.2418, + "num_input_tokens_seen": 10652288, + "step": 13965 + }, + { + "epoch": 29.043659043659044, + "grad_norm": 0.00027221511118113995, + "learning_rate": 0.21842377742047195, + "loss": 0.2495, + "num_input_tokens_seen": 10656000, + "step": 13970 + }, + { + "epoch": 29.054054054054053, + "grad_norm": 0.00017686275532469153, + "learning_rate": 0.21837135279510705, + "loss": 0.2643, + "num_input_tokens_seen": 10659872, + "step": 13975 + }, + { + "epoch": 29.064449064449065, + "grad_norm": 0.00022571589215658605, + "learning_rate": 0.21831891762602038, + "loss": 0.2604, + "num_input_tokens_seen": 10663648, + "step": 13980 + }, + { + "epoch": 29.074844074844076, + "grad_norm": 0.00042226180084981024, + "learning_rate": 0.21826647192129806, + "loss": 0.2637, + "num_input_tokens_seen": 10667456, + "step": 13985 + }, + { + "epoch": 29.085239085239085, + "grad_norm": 5.466501534101553e-05, + "learning_rate": 0.21821401568902787, + "loss": 0.2717, + "num_input_tokens_seen": 10671264, + "step": 13990 + }, + { + "epoch": 29.095634095634097, + "grad_norm": 3.547640881151892e-05, + "learning_rate": 0.21816154893729925, + "loss": 0.288, + "num_input_tokens_seen": 10675264, + "step": 13995 + }, + { + "epoch": 29.106029106029105, + "grad_norm": 8.872819307725877e-05, + "learning_rate": 0.2181090716742032, + "loss": 0.273, + "num_input_tokens_seen": 10678976, + "step": 14000 + }, + { + "epoch": 29.106029106029105, + "eval_loss": 0.25908181071281433, + "eval_runtime": 13.4319, + "eval_samples_per_second": 63.729, + "eval_steps_per_second": 15.932, + "num_input_tokens_seen": 10678976, + "step": 14000 + }, + { + "epoch": 29.116424116424117, + "grad_norm": 0.0005775134195573628, + "learning_rate": 0.21805658390783236, + "loss": 0.2599, + "num_input_tokens_seen": 10682720, + "step": 14005 + }, + { + "epoch": 29.126819126819125, + "grad_norm": 0.0008774096495471895, + "learning_rate": 0.21800408564628107, + "loss": 0.26, + "num_input_tokens_seen": 10686592, + "step": 14010 + }, + { + "epoch": 29.137214137214137, + "grad_norm": 0.00013676991511601955, + "learning_rate": 0.21795157689764516, + "loss": 0.2927, + "num_input_tokens_seen": 10690464, + "step": 14015 + }, + { + "epoch": 29.14760914760915, + "grad_norm": 0.0003483971522655338, + "learning_rate": 0.21789905767002216, + "loss": 0.2726, + "num_input_tokens_seen": 10694304, + "step": 14020 + }, + { + "epoch": 29.158004158004157, + "grad_norm": 8.366950351046398e-05, + "learning_rate": 0.2178465279715112, + "loss": 0.2579, + "num_input_tokens_seen": 10698208, + "step": 14025 + }, + { + "epoch": 29.16839916839917, + "grad_norm": 0.00020679635053966194, + "learning_rate": 0.21779398781021303, + "loss": 0.2474, + "num_input_tokens_seen": 10701888, + "step": 14030 + }, + { + "epoch": 29.178794178794178, + "grad_norm": 0.0008092079660855234, + "learning_rate": 0.21774143719422998, + "loss": 0.3004, + "num_input_tokens_seen": 10705792, + "step": 14035 + }, + { + "epoch": 29.18918918918919, + "grad_norm": 0.0009311113390140235, + "learning_rate": 0.21768887613166601, + "loss": 0.2824, + "num_input_tokens_seen": 10709632, + "step": 14040 + }, + { + "epoch": 29.1995841995842, + "grad_norm": 0.0006800752598792315, + "learning_rate": 0.2176363046306267, + "loss": 0.2897, + "num_input_tokens_seen": 10713536, + "step": 14045 + }, + { + "epoch": 29.20997920997921, + "grad_norm": 0.00022434488346334547, + "learning_rate": 0.21758372269921925, + "loss": 0.2876, + "num_input_tokens_seen": 10717376, + "step": 14050 + }, + { + "epoch": 29.22037422037422, + "grad_norm": 0.00045676474110223353, + "learning_rate": 0.21753113034555244, + "loss": 0.2497, + "num_input_tokens_seen": 10721344, + "step": 14055 + }, + { + "epoch": 29.23076923076923, + "grad_norm": 0.00011392116721253842, + "learning_rate": 0.2174785275777367, + "loss": 0.3248, + "num_input_tokens_seen": 10725312, + "step": 14060 + }, + { + "epoch": 29.241164241164242, + "grad_norm": 0.00035922828828915954, + "learning_rate": 0.21742591440388404, + "loss": 0.2263, + "num_input_tokens_seen": 10728864, + "step": 14065 + }, + { + "epoch": 29.25155925155925, + "grad_norm": 0.00013645851868204772, + "learning_rate": 0.21737329083210802, + "loss": 0.2233, + "num_input_tokens_seen": 10732640, + "step": 14070 + }, + { + "epoch": 29.261954261954262, + "grad_norm": 0.0001173399796243757, + "learning_rate": 0.2173206568705239, + "loss": 0.2297, + "num_input_tokens_seen": 10736416, + "step": 14075 + }, + { + "epoch": 29.272349272349274, + "grad_norm": 0.00047181517584249377, + "learning_rate": 0.2172680125272485, + "loss": 0.2609, + "num_input_tokens_seen": 10740096, + "step": 14080 + }, + { + "epoch": 29.282744282744282, + "grad_norm": 0.00013942638179287314, + "learning_rate": 0.2172153578104002, + "loss": 0.299, + "num_input_tokens_seen": 10743936, + "step": 14085 + }, + { + "epoch": 29.293139293139294, + "grad_norm": 0.00021279761858750135, + "learning_rate": 0.21716269272809902, + "loss": 0.2727, + "num_input_tokens_seen": 10747680, + "step": 14090 + }, + { + "epoch": 29.303534303534303, + "grad_norm": 0.00019477716705296189, + "learning_rate": 0.21711001728846666, + "loss": 0.2689, + "num_input_tokens_seen": 10751392, + "step": 14095 + }, + { + "epoch": 29.313929313929314, + "grad_norm": 0.0002950727939605713, + "learning_rate": 0.21705733149962628, + "loss": 0.2491, + "num_input_tokens_seen": 10755232, + "step": 14100 + }, + { + "epoch": 29.324324324324323, + "grad_norm": 0.000299810926662758, + "learning_rate": 0.21700463536970263, + "loss": 0.2718, + "num_input_tokens_seen": 10759232, + "step": 14105 + }, + { + "epoch": 29.334719334719335, + "grad_norm": 0.0007824341882951558, + "learning_rate": 0.21695192890682222, + "loss": 0.252, + "num_input_tokens_seen": 10763136, + "step": 14110 + }, + { + "epoch": 29.345114345114347, + "grad_norm": 0.0002583378809504211, + "learning_rate": 0.21689921211911298, + "loss": 0.2369, + "num_input_tokens_seen": 10766912, + "step": 14115 + }, + { + "epoch": 29.355509355509355, + "grad_norm": 7.649027975276113e-05, + "learning_rate": 0.21684648501470452, + "loss": 0.2736, + "num_input_tokens_seen": 10770880, + "step": 14120 + }, + { + "epoch": 29.365904365904367, + "grad_norm": 0.00036276126047596335, + "learning_rate": 0.216793747601728, + "loss": 0.2902, + "num_input_tokens_seen": 10774688, + "step": 14125 + }, + { + "epoch": 29.376299376299375, + "grad_norm": 0.0001871962595032528, + "learning_rate": 0.21674099988831627, + "loss": 0.2616, + "num_input_tokens_seen": 10778560, + "step": 14130 + }, + { + "epoch": 29.386694386694387, + "grad_norm": 0.0005337464972399175, + "learning_rate": 0.21668824188260363, + "loss": 0.2632, + "num_input_tokens_seen": 10782304, + "step": 14135 + }, + { + "epoch": 29.397089397089395, + "grad_norm": 0.0007402925402857363, + "learning_rate": 0.21663547359272606, + "loss": 0.2654, + "num_input_tokens_seen": 10786304, + "step": 14140 + }, + { + "epoch": 29.407484407484407, + "grad_norm": 0.0001457414764445275, + "learning_rate": 0.216582695026821, + "loss": 0.2546, + "num_input_tokens_seen": 10790112, + "step": 14145 + }, + { + "epoch": 29.41787941787942, + "grad_norm": 0.00019454641733318567, + "learning_rate": 0.21652990619302767, + "loss": 0.2586, + "num_input_tokens_seen": 10793984, + "step": 14150 + }, + { + "epoch": 29.428274428274428, + "grad_norm": 0.0005052004707977176, + "learning_rate": 0.21647710709948673, + "loss": 0.2815, + "num_input_tokens_seen": 10797632, + "step": 14155 + }, + { + "epoch": 29.43866943866944, + "grad_norm": 0.00010229657345917076, + "learning_rate": 0.2164242977543405, + "loss": 0.2479, + "num_input_tokens_seen": 10801344, + "step": 14160 + }, + { + "epoch": 29.449064449064448, + "grad_norm": 0.0007540743099525571, + "learning_rate": 0.21637147816573277, + "loss": 0.313, + "num_input_tokens_seen": 10805152, + "step": 14165 + }, + { + "epoch": 29.45945945945946, + "grad_norm": 0.0003524782950989902, + "learning_rate": 0.21631864834180908, + "loss": 0.2595, + "num_input_tokens_seen": 10809024, + "step": 14170 + }, + { + "epoch": 29.46985446985447, + "grad_norm": 6.010181095916778e-05, + "learning_rate": 0.21626580829071637, + "loss": 0.2634, + "num_input_tokens_seen": 10812672, + "step": 14175 + }, + { + "epoch": 29.48024948024948, + "grad_norm": 0.0005448845331557095, + "learning_rate": 0.21621295802060328, + "loss": 0.282, + "num_input_tokens_seen": 10816480, + "step": 14180 + }, + { + "epoch": 29.490644490644492, + "grad_norm": 0.0006666610715910792, + "learning_rate": 0.21616009753961996, + "loss": 0.262, + "num_input_tokens_seen": 10820224, + "step": 14185 + }, + { + "epoch": 29.5010395010395, + "grad_norm": 0.0002556899271439761, + "learning_rate": 0.2161072268559182, + "loss": 0.2724, + "num_input_tokens_seen": 10823872, + "step": 14190 + }, + { + "epoch": 29.511434511434512, + "grad_norm": 0.0002947613247670233, + "learning_rate": 0.21605434597765133, + "loss": 0.2601, + "num_input_tokens_seen": 10827648, + "step": 14195 + }, + { + "epoch": 29.52182952182952, + "grad_norm": 5.8358047681394964e-05, + "learning_rate": 0.21600145491297418, + "loss": 0.2554, + "num_input_tokens_seen": 10831520, + "step": 14200 + }, + { + "epoch": 29.52182952182952, + "eval_loss": 0.24808000028133392, + "eval_runtime": 13.4332, + "eval_samples_per_second": 63.723, + "eval_steps_per_second": 15.931, + "num_input_tokens_seen": 10831520, + "step": 14200 + }, + { + "epoch": 29.532224532224532, + "grad_norm": 0.0004730260989163071, + "learning_rate": 0.21594855367004326, + "loss": 0.2488, + "num_input_tokens_seen": 10835584, + "step": 14205 + }, + { + "epoch": 29.542619542619544, + "grad_norm": 0.0008272915729321539, + "learning_rate": 0.21589564225701663, + "loss": 0.2317, + "num_input_tokens_seen": 10839392, + "step": 14210 + }, + { + "epoch": 29.553014553014552, + "grad_norm": 6.040620428393595e-05, + "learning_rate": 0.21584272068205385, + "loss": 0.2575, + "num_input_tokens_seen": 10843200, + "step": 14215 + }, + { + "epoch": 29.563409563409564, + "grad_norm": 0.00018372329941485077, + "learning_rate": 0.2157897889533161, + "loss": 0.2745, + "num_input_tokens_seen": 10846944, + "step": 14220 + }, + { + "epoch": 29.573804573804573, + "grad_norm": 0.00029755596187897027, + "learning_rate": 0.21573684707896612, + "loss": 0.2656, + "num_input_tokens_seen": 10850816, + "step": 14225 + }, + { + "epoch": 29.584199584199585, + "grad_norm": 0.0002608706708997488, + "learning_rate": 0.21568389506716826, + "loss": 0.2686, + "num_input_tokens_seen": 10854592, + "step": 14230 + }, + { + "epoch": 29.594594594594593, + "grad_norm": 0.00029237096896395087, + "learning_rate": 0.21563093292608831, + "loss": 0.2798, + "num_input_tokens_seen": 10858368, + "step": 14235 + }, + { + "epoch": 29.604989604989605, + "grad_norm": 0.0002047140005743131, + "learning_rate": 0.21557796066389376, + "loss": 0.2774, + "num_input_tokens_seen": 10862272, + "step": 14240 + }, + { + "epoch": 29.615384615384617, + "grad_norm": 0.0006563168717548251, + "learning_rate": 0.21552497828875353, + "loss": 0.273, + "num_input_tokens_seen": 10865984, + "step": 14245 + }, + { + "epoch": 29.625779625779625, + "grad_norm": 0.0001679753913776949, + "learning_rate": 0.21547198580883828, + "loss": 0.2607, + "num_input_tokens_seen": 10869856, + "step": 14250 + }, + { + "epoch": 29.636174636174637, + "grad_norm": 0.00012551013787742704, + "learning_rate": 0.21541898323232, + "loss": 0.2215, + "num_input_tokens_seen": 10873792, + "step": 14255 + }, + { + "epoch": 29.646569646569645, + "grad_norm": 0.0003813078219536692, + "learning_rate": 0.2153659705673724, + "loss": 0.3193, + "num_input_tokens_seen": 10877632, + "step": 14260 + }, + { + "epoch": 29.656964656964657, + "grad_norm": 0.0005680934991687536, + "learning_rate": 0.2153129478221707, + "loss": 0.2757, + "num_input_tokens_seen": 10881536, + "step": 14265 + }, + { + "epoch": 29.66735966735967, + "grad_norm": 0.00011374261521268636, + "learning_rate": 0.21525991500489164, + "loss": 0.2805, + "num_input_tokens_seen": 10885312, + "step": 14270 + }, + { + "epoch": 29.677754677754677, + "grad_norm": 3.348291284055449e-05, + "learning_rate": 0.21520687212371362, + "loss": 0.2726, + "num_input_tokens_seen": 10889056, + "step": 14275 + }, + { + "epoch": 29.68814968814969, + "grad_norm": 5.475520811160095e-05, + "learning_rate": 0.21515381918681648, + "loss": 0.2594, + "num_input_tokens_seen": 10892928, + "step": 14280 + }, + { + "epoch": 29.698544698544698, + "grad_norm": 9.333314665127546e-05, + "learning_rate": 0.21510075620238167, + "loss": 0.2632, + "num_input_tokens_seen": 10896960, + "step": 14285 + }, + { + "epoch": 29.70893970893971, + "grad_norm": 0.00015351705951616168, + "learning_rate": 0.21504768317859208, + "loss": 0.2798, + "num_input_tokens_seen": 10900704, + "step": 14290 + }, + { + "epoch": 29.719334719334718, + "grad_norm": 0.00033292261650785804, + "learning_rate": 0.2149946001236323, + "loss": 0.2461, + "num_input_tokens_seen": 10904544, + "step": 14295 + }, + { + "epoch": 29.72972972972973, + "grad_norm": 0.00029600097332149744, + "learning_rate": 0.21494150704568848, + "loss": 0.2871, + "num_input_tokens_seen": 10908096, + "step": 14300 + }, + { + "epoch": 29.74012474012474, + "grad_norm": 0.0004019803600385785, + "learning_rate": 0.21488840395294811, + "loss": 0.2626, + "num_input_tokens_seen": 10911936, + "step": 14305 + }, + { + "epoch": 29.75051975051975, + "grad_norm": 0.0001952204038389027, + "learning_rate": 0.21483529085360042, + "loss": 0.2774, + "num_input_tokens_seen": 10915776, + "step": 14310 + }, + { + "epoch": 29.760914760914762, + "grad_norm": 0.00013958119961898774, + "learning_rate": 0.2147821677558361, + "loss": 0.2566, + "num_input_tokens_seen": 10919680, + "step": 14315 + }, + { + "epoch": 29.77130977130977, + "grad_norm": 0.0002875577483791858, + "learning_rate": 0.2147290346678475, + "loss": 0.2941, + "num_input_tokens_seen": 10923488, + "step": 14320 + }, + { + "epoch": 29.781704781704782, + "grad_norm": 0.0005459499661810696, + "learning_rate": 0.21467589159782827, + "loss": 0.2386, + "num_input_tokens_seen": 10927392, + "step": 14325 + }, + { + "epoch": 29.79209979209979, + "grad_norm": 0.00017256855790037662, + "learning_rate": 0.21462273855397374, + "loss": 0.2661, + "num_input_tokens_seen": 10931168, + "step": 14330 + }, + { + "epoch": 29.802494802494802, + "grad_norm": 0.0001086880947696045, + "learning_rate": 0.21456957554448083, + "loss": 0.2901, + "num_input_tokens_seen": 10934912, + "step": 14335 + }, + { + "epoch": 29.812889812889814, + "grad_norm": 0.00013465310621540993, + "learning_rate": 0.21451640257754795, + "loss": 0.2663, + "num_input_tokens_seen": 10938656, + "step": 14340 + }, + { + "epoch": 29.823284823284823, + "grad_norm": 0.0007124594412744045, + "learning_rate": 0.21446321966137508, + "loss": 0.2719, + "num_input_tokens_seen": 10942304, + "step": 14345 + }, + { + "epoch": 29.833679833679835, + "grad_norm": 0.00045124292955733836, + "learning_rate": 0.21441002680416354, + "loss": 0.2463, + "num_input_tokens_seen": 10945952, + "step": 14350 + }, + { + "epoch": 29.844074844074843, + "grad_norm": 0.0005798981292173266, + "learning_rate": 0.21435682401411654, + "loss": 0.2653, + "num_input_tokens_seen": 10949536, + "step": 14355 + }, + { + "epoch": 29.854469854469855, + "grad_norm": 9.99642361421138e-05, + "learning_rate": 0.2143036112994385, + "loss": 0.2816, + "num_input_tokens_seen": 10953408, + "step": 14360 + }, + { + "epoch": 29.864864864864863, + "grad_norm": 0.00011309548426652327, + "learning_rate": 0.21425038866833548, + "loss": 0.2679, + "num_input_tokens_seen": 10957280, + "step": 14365 + }, + { + "epoch": 29.875259875259875, + "grad_norm": 9.759383829077706e-05, + "learning_rate": 0.21419715612901508, + "loss": 0.2794, + "num_input_tokens_seen": 10961216, + "step": 14370 + }, + { + "epoch": 29.885654885654887, + "grad_norm": 0.00030324768158607185, + "learning_rate": 0.21414391368968652, + "loss": 0.2731, + "num_input_tokens_seen": 10965056, + "step": 14375 + }, + { + "epoch": 29.896049896049895, + "grad_norm": 0.00017458086949773133, + "learning_rate": 0.21409066135856034, + "loss": 0.2837, + "num_input_tokens_seen": 10968960, + "step": 14380 + }, + { + "epoch": 29.906444906444907, + "grad_norm": 0.000135528709506616, + "learning_rate": 0.21403739914384878, + "loss": 0.2751, + "num_input_tokens_seen": 10972864, + "step": 14385 + }, + { + "epoch": 29.916839916839916, + "grad_norm": 0.00046915566781535745, + "learning_rate": 0.21398412705376554, + "loss": 0.2817, + "num_input_tokens_seen": 10976736, + "step": 14390 + }, + { + "epoch": 29.927234927234927, + "grad_norm": 0.00025480909971520305, + "learning_rate": 0.2139308450965258, + "loss": 0.2656, + "num_input_tokens_seen": 10980512, + "step": 14395 + }, + { + "epoch": 29.93762993762994, + "grad_norm": 0.0002910270995926112, + "learning_rate": 0.21387755328034638, + "loss": 0.2605, + "num_input_tokens_seen": 10984224, + "step": 14400 + }, + { + "epoch": 29.93762993762994, + "eval_loss": 0.24851953983306885, + "eval_runtime": 13.4113, + "eval_samples_per_second": 63.827, + "eval_steps_per_second": 15.957, + "num_input_tokens_seen": 10984224, + "step": 14400 + }, + { + "epoch": 29.948024948024948, + "grad_norm": 0.00014919557725079358, + "learning_rate": 0.2138242516134455, + "loss": 0.2688, + "num_input_tokens_seen": 10988000, + "step": 14405 + }, + { + "epoch": 29.95841995841996, + "grad_norm": 0.0001066965633071959, + "learning_rate": 0.2137709401040429, + "loss": 0.2667, + "num_input_tokens_seen": 10991808, + "step": 14410 + }, + { + "epoch": 29.968814968814968, + "grad_norm": 0.00021898998238611966, + "learning_rate": 0.21371761876036, + "loss": 0.2726, + "num_input_tokens_seen": 10995616, + "step": 14415 + }, + { + "epoch": 29.97920997920998, + "grad_norm": 0.0002709201944526285, + "learning_rate": 0.21366428759061956, + "loss": 0.2842, + "num_input_tokens_seen": 10999360, + "step": 14420 + }, + { + "epoch": 29.989604989604988, + "grad_norm": 0.0007954153697937727, + "learning_rate": 0.2136109466030459, + "loss": 0.2774, + "num_input_tokens_seen": 11003104, + "step": 14425 + }, + { + "epoch": 30.0, + "grad_norm": 0.0003685402625706047, + "learning_rate": 0.2135575958058649, + "loss": 0.2779, + "num_input_tokens_seen": 11006712, + "step": 14430 + }, + { + "epoch": 30.010395010395012, + "grad_norm": 0.00014657160500064492, + "learning_rate": 0.2135042352073039, + "loss": 0.2699, + "num_input_tokens_seen": 11010584, + "step": 14435 + }, + { + "epoch": 30.02079002079002, + "grad_norm": 3.5852124710800126e-05, + "learning_rate": 0.2134508648155918, + "loss": 0.2682, + "num_input_tokens_seen": 11014328, + "step": 14440 + }, + { + "epoch": 30.031185031185032, + "grad_norm": 0.0004663629806600511, + "learning_rate": 0.213397484638959, + "loss": 0.262, + "num_input_tokens_seen": 11017944, + "step": 14445 + }, + { + "epoch": 30.04158004158004, + "grad_norm": 8.454402268398553e-05, + "learning_rate": 0.21334409468563728, + "loss": 0.2558, + "num_input_tokens_seen": 11021720, + "step": 14450 + }, + { + "epoch": 30.051975051975052, + "grad_norm": 0.00010607652802718803, + "learning_rate": 0.2132906949638602, + "loss": 0.2653, + "num_input_tokens_seen": 11025688, + "step": 14455 + }, + { + "epoch": 30.06237006237006, + "grad_norm": 0.00010131805174751207, + "learning_rate": 0.21323728548186255, + "loss": 0.2648, + "num_input_tokens_seen": 11029464, + "step": 14460 + }, + { + "epoch": 30.072765072765073, + "grad_norm": 0.00019035505829378963, + "learning_rate": 0.21318386624788088, + "loss": 0.2554, + "num_input_tokens_seen": 11033368, + "step": 14465 + }, + { + "epoch": 30.083160083160084, + "grad_norm": 0.00047763189650140703, + "learning_rate": 0.21313043727015288, + "loss": 0.2415, + "num_input_tokens_seen": 11037112, + "step": 14470 + }, + { + "epoch": 30.093555093555093, + "grad_norm": 0.0001339707669103518, + "learning_rate": 0.2130769985569182, + "loss": 0.2781, + "num_input_tokens_seen": 11040952, + "step": 14475 + }, + { + "epoch": 30.103950103950105, + "grad_norm": 0.00038092880276963115, + "learning_rate": 0.21302355011641766, + "loss": 0.2668, + "num_input_tokens_seen": 11044888, + "step": 14480 + }, + { + "epoch": 30.114345114345113, + "grad_norm": 0.0001321235322393477, + "learning_rate": 0.21297009195689365, + "loss": 0.2725, + "num_input_tokens_seen": 11048760, + "step": 14485 + }, + { + "epoch": 30.124740124740125, + "grad_norm": 0.00017710465181153268, + "learning_rate": 0.21291662408659015, + "loss": 0.2488, + "num_input_tokens_seen": 11052472, + "step": 14490 + }, + { + "epoch": 30.135135135135137, + "grad_norm": 0.0002573048113845289, + "learning_rate": 0.21286314651375254, + "loss": 0.2469, + "num_input_tokens_seen": 11056216, + "step": 14495 + }, + { + "epoch": 30.145530145530145, + "grad_norm": 0.0006025422480888665, + "learning_rate": 0.2128096592466278, + "loss": 0.2795, + "num_input_tokens_seen": 11059928, + "step": 14500 + }, + { + "epoch": 30.155925155925157, + "grad_norm": 0.00015312687901314348, + "learning_rate": 0.21275616229346428, + "loss": 0.272, + "num_input_tokens_seen": 11063768, + "step": 14505 + }, + { + "epoch": 30.166320166320165, + "grad_norm": 0.00030205590883269906, + "learning_rate": 0.21270265566251184, + "loss": 0.279, + "num_input_tokens_seen": 11067512, + "step": 14510 + }, + { + "epoch": 30.176715176715177, + "grad_norm": 1.925172909977846e-05, + "learning_rate": 0.21264913936202193, + "loss": 0.2766, + "num_input_tokens_seen": 11071416, + "step": 14515 + }, + { + "epoch": 30.187110187110186, + "grad_norm": 0.00011313604773022234, + "learning_rate": 0.2125956134002475, + "loss": 0.2699, + "num_input_tokens_seen": 11075160, + "step": 14520 + }, + { + "epoch": 30.197505197505198, + "grad_norm": 0.00010148439469048753, + "learning_rate": 0.2125420777854428, + "loss": 0.2596, + "num_input_tokens_seen": 11079064, + "step": 14525 + }, + { + "epoch": 30.20790020790021, + "grad_norm": 0.0004043885273858905, + "learning_rate": 0.21248853252586372, + "loss": 0.291, + "num_input_tokens_seen": 11082808, + "step": 14530 + }, + { + "epoch": 30.218295218295218, + "grad_norm": 0.00021684229432139546, + "learning_rate": 0.21243497762976774, + "loss": 0.2237, + "num_input_tokens_seen": 11086680, + "step": 14535 + }, + { + "epoch": 30.22869022869023, + "grad_norm": 0.0008704495849087834, + "learning_rate": 0.21238141310541356, + "loss": 0.2942, + "num_input_tokens_seen": 11090456, + "step": 14540 + }, + { + "epoch": 30.239085239085238, + "grad_norm": 0.0005876815412193537, + "learning_rate": 0.21232783896106153, + "loss": 0.2703, + "num_input_tokens_seen": 11094264, + "step": 14545 + }, + { + "epoch": 30.24948024948025, + "grad_norm": 0.0007548096473328769, + "learning_rate": 0.21227425520497345, + "loss": 0.271, + "num_input_tokens_seen": 11098104, + "step": 14550 + }, + { + "epoch": 30.25987525987526, + "grad_norm": 0.00036280223866924644, + "learning_rate": 0.2122206618454127, + "loss": 0.2861, + "num_input_tokens_seen": 11102008, + "step": 14555 + }, + { + "epoch": 30.27027027027027, + "grad_norm": 0.00014951849880162627, + "learning_rate": 0.2121670588906439, + "loss": 0.2715, + "num_input_tokens_seen": 11105784, + "step": 14560 + }, + { + "epoch": 30.280665280665282, + "grad_norm": 5.922588024986908e-05, + "learning_rate": 0.21211344634893345, + "loss": 0.2498, + "num_input_tokens_seen": 11109528, + "step": 14565 + }, + { + "epoch": 30.29106029106029, + "grad_norm": 0.00031049034441821277, + "learning_rate": 0.21205982422854897, + "loss": 0.2563, + "num_input_tokens_seen": 11113272, + "step": 14570 + }, + { + "epoch": 30.301455301455302, + "grad_norm": 0.00013043484068475664, + "learning_rate": 0.21200619253775974, + "loss": 0.2424, + "num_input_tokens_seen": 11117048, + "step": 14575 + }, + { + "epoch": 30.31185031185031, + "grad_norm": 0.00012856299872510135, + "learning_rate": 0.21195255128483637, + "loss": 0.2657, + "num_input_tokens_seen": 11120792, + "step": 14580 + }, + { + "epoch": 30.322245322245323, + "grad_norm": 0.00017962511628866196, + "learning_rate": 0.21189890047805102, + "loss": 0.2805, + "num_input_tokens_seen": 11124600, + "step": 14585 + }, + { + "epoch": 30.33264033264033, + "grad_norm": 0.00015870440984144807, + "learning_rate": 0.21184524012567735, + "loss": 0.2646, + "num_input_tokens_seen": 11128408, + "step": 14590 + }, + { + "epoch": 30.343035343035343, + "grad_norm": 0.0002069677138933912, + "learning_rate": 0.2117915702359905, + "loss": 0.2592, + "num_input_tokens_seen": 11132120, + "step": 14595 + }, + { + "epoch": 30.353430353430355, + "grad_norm": 0.000353456474840641, + "learning_rate": 0.211737890817267, + "loss": 0.2916, + "num_input_tokens_seen": 11135896, + "step": 14600 + }, + { + "epoch": 30.353430353430355, + "eval_loss": 0.2486879974603653, + "eval_runtime": 13.4204, + "eval_samples_per_second": 63.783, + "eval_steps_per_second": 15.946, + "num_input_tokens_seen": 11135896, + "step": 14600 + }, + { + "epoch": 30.363825363825363, + "grad_norm": 0.0002123817102983594, + "learning_rate": 0.21168420187778483, + "loss": 0.2508, + "num_input_tokens_seen": 11139672, + "step": 14605 + }, + { + "epoch": 30.374220374220375, + "grad_norm": 0.00018537460709922016, + "learning_rate": 0.21163050342582362, + "loss": 0.2418, + "num_input_tokens_seen": 11143480, + "step": 14610 + }, + { + "epoch": 30.384615384615383, + "grad_norm": 0.00010919541091425344, + "learning_rate": 0.21157679546966426, + "loss": 0.2786, + "num_input_tokens_seen": 11147416, + "step": 14615 + }, + { + "epoch": 30.395010395010395, + "grad_norm": 0.0005122404545545578, + "learning_rate": 0.2115230780175892, + "loss": 0.2537, + "num_input_tokens_seen": 11151128, + "step": 14620 + }, + { + "epoch": 30.405405405405407, + "grad_norm": 0.0005644683260470629, + "learning_rate": 0.21146935107788237, + "loss": 0.2852, + "num_input_tokens_seen": 11154840, + "step": 14625 + }, + { + "epoch": 30.415800415800415, + "grad_norm": 0.0005030115135014057, + "learning_rate": 0.21141561465882916, + "loss": 0.2782, + "num_input_tokens_seen": 11158584, + "step": 14630 + }, + { + "epoch": 30.426195426195427, + "grad_norm": 0.00010961908265016973, + "learning_rate": 0.21136186876871635, + "loss": 0.2733, + "num_input_tokens_seen": 11162712, + "step": 14635 + }, + { + "epoch": 30.436590436590436, + "grad_norm": 5.42264933756087e-05, + "learning_rate": 0.21130811341583225, + "loss": 0.2578, + "num_input_tokens_seen": 11166520, + "step": 14640 + }, + { + "epoch": 30.446985446985448, + "grad_norm": 0.0004980212543159723, + "learning_rate": 0.21125434860846667, + "loss": 0.2776, + "num_input_tokens_seen": 11170264, + "step": 14645 + }, + { + "epoch": 30.457380457380456, + "grad_norm": 0.0004620563704520464, + "learning_rate": 0.2112005743549107, + "loss": 0.2673, + "num_input_tokens_seen": 11174072, + "step": 14650 + }, + { + "epoch": 30.467775467775468, + "grad_norm": 0.0003817227843683213, + "learning_rate": 0.21114679066345707, + "loss": 0.283, + "num_input_tokens_seen": 11177880, + "step": 14655 + }, + { + "epoch": 30.47817047817048, + "grad_norm": 0.00012710904411505908, + "learning_rate": 0.21109299754239993, + "loss": 0.2833, + "num_input_tokens_seen": 11181720, + "step": 14660 + }, + { + "epoch": 30.488565488565488, + "grad_norm": 0.00015433464432135224, + "learning_rate": 0.21103919500003482, + "loss": 0.2677, + "num_input_tokens_seen": 11185560, + "step": 14665 + }, + { + "epoch": 30.4989604989605, + "grad_norm": 0.0003585397789720446, + "learning_rate": 0.21098538304465872, + "loss": 0.2797, + "num_input_tokens_seen": 11189368, + "step": 14670 + }, + { + "epoch": 30.509355509355508, + "grad_norm": 0.0005723215290345252, + "learning_rate": 0.2109315616845702, + "loss": 0.2888, + "num_input_tokens_seen": 11193112, + "step": 14675 + }, + { + "epoch": 30.51975051975052, + "grad_norm": 0.0002040408580796793, + "learning_rate": 0.21087773092806925, + "loss": 0.2808, + "num_input_tokens_seen": 11197016, + "step": 14680 + }, + { + "epoch": 30.53014553014553, + "grad_norm": 8.161534060491249e-05, + "learning_rate": 0.21082389078345704, + "loss": 0.2717, + "num_input_tokens_seen": 11200888, + "step": 14685 + }, + { + "epoch": 30.54054054054054, + "grad_norm": 0.00043451960664242506, + "learning_rate": 0.2107700412590365, + "loss": 0.277, + "num_input_tokens_seen": 11204632, + "step": 14690 + }, + { + "epoch": 30.550935550935552, + "grad_norm": 0.0005331505089998245, + "learning_rate": 0.210716182363112, + "loss": 0.2622, + "num_input_tokens_seen": 11208536, + "step": 14695 + }, + { + "epoch": 30.56133056133056, + "grad_norm": 0.00018044031457975507, + "learning_rate": 0.2106623141039891, + "loss": 0.2398, + "num_input_tokens_seen": 11212440, + "step": 14700 + }, + { + "epoch": 30.571725571725572, + "grad_norm": 0.00010524292156333104, + "learning_rate": 0.21060843648997507, + "loss": 0.2758, + "num_input_tokens_seen": 11216376, + "step": 14705 + }, + { + "epoch": 30.58212058212058, + "grad_norm": 0.0001528459251858294, + "learning_rate": 0.21055454952937844, + "loss": 0.2693, + "num_input_tokens_seen": 11220184, + "step": 14710 + }, + { + "epoch": 30.592515592515593, + "grad_norm": 0.00018082439783029258, + "learning_rate": 0.21050065323050937, + "loss": 0.2674, + "num_input_tokens_seen": 11223928, + "step": 14715 + }, + { + "epoch": 30.602910602910605, + "grad_norm": 0.00011943643767153844, + "learning_rate": 0.21044674760167928, + "loss": 0.2776, + "num_input_tokens_seen": 11227736, + "step": 14720 + }, + { + "epoch": 30.613305613305613, + "grad_norm": 0.0004058046324644238, + "learning_rate": 0.210392832651201, + "loss": 0.2646, + "num_input_tokens_seen": 11231736, + "step": 14725 + }, + { + "epoch": 30.623700623700625, + "grad_norm": 0.0002450610918458551, + "learning_rate": 0.210338908387389, + "loss": 0.2446, + "num_input_tokens_seen": 11235480, + "step": 14730 + }, + { + "epoch": 30.634095634095633, + "grad_norm": 0.0001955005864147097, + "learning_rate": 0.21028497481855912, + "loss": 0.2746, + "num_input_tokens_seen": 11239160, + "step": 14735 + }, + { + "epoch": 30.644490644490645, + "grad_norm": 0.0002759649360086769, + "learning_rate": 0.21023103195302847, + "loss": 0.2292, + "num_input_tokens_seen": 11242968, + "step": 14740 + }, + { + "epoch": 30.654885654885653, + "grad_norm": 0.0003272598551120609, + "learning_rate": 0.21017707979911582, + "loss": 0.2753, + "num_input_tokens_seen": 11246840, + "step": 14745 + }, + { + "epoch": 30.665280665280665, + "grad_norm": 0.00018998980522155762, + "learning_rate": 0.21012311836514122, + "loss": 0.243, + "num_input_tokens_seen": 11250648, + "step": 14750 + }, + { + "epoch": 30.675675675675677, + "grad_norm": 0.0002959202975034714, + "learning_rate": 0.21006914765942622, + "loss": 0.2692, + "num_input_tokens_seen": 11254424, + "step": 14755 + }, + { + "epoch": 30.686070686070686, + "grad_norm": 0.0002520527341403067, + "learning_rate": 0.2100151676902938, + "loss": 0.2704, + "num_input_tokens_seen": 11258136, + "step": 14760 + }, + { + "epoch": 30.696465696465697, + "grad_norm": 0.0004911919822916389, + "learning_rate": 0.2099611784660683, + "loss": 0.2802, + "num_input_tokens_seen": 11262104, + "step": 14765 + }, + { + "epoch": 30.706860706860706, + "grad_norm": 0.00017487841250840575, + "learning_rate": 0.20990717999507552, + "loss": 0.2659, + "num_input_tokens_seen": 11265816, + "step": 14770 + }, + { + "epoch": 30.717255717255718, + "grad_norm": 0.0005485157016664743, + "learning_rate": 0.20985317228564276, + "loss": 0.2616, + "num_input_tokens_seen": 11269656, + "step": 14775 + }, + { + "epoch": 30.727650727650726, + "grad_norm": 0.0004643997235689312, + "learning_rate": 0.20979915534609872, + "loss": 0.3221, + "num_input_tokens_seen": 11273400, + "step": 14780 + }, + { + "epoch": 30.738045738045738, + "grad_norm": 7.572733011329547e-05, + "learning_rate": 0.20974512918477342, + "loss": 0.2686, + "num_input_tokens_seen": 11277240, + "step": 14785 + }, + { + "epoch": 30.74844074844075, + "grad_norm": 0.0007108752615749836, + "learning_rate": 0.2096910938099984, + "loss": 0.2893, + "num_input_tokens_seen": 11281080, + "step": 14790 + }, + { + "epoch": 30.758835758835758, + "grad_norm": 9.185164526570588e-05, + "learning_rate": 0.2096370492301066, + "loss": 0.2728, + "num_input_tokens_seen": 11284920, + "step": 14795 + }, + { + "epoch": 30.76923076923077, + "grad_norm": 0.00011459348752396181, + "learning_rate": 0.2095829954534323, + "loss": 0.2517, + "num_input_tokens_seen": 11288728, + "step": 14800 + }, + { + "epoch": 30.76923076923077, + "eval_loss": 0.25030773878097534, + "eval_runtime": 13.5245, + "eval_samples_per_second": 63.292, + "eval_steps_per_second": 15.823, + "num_input_tokens_seen": 11288728, + "step": 14800 + }, + { + "epoch": 30.77962577962578, + "grad_norm": 5.75693447899539e-05, + "learning_rate": 0.2095289324883114, + "loss": 0.2984, + "num_input_tokens_seen": 11292728, + "step": 14805 + }, + { + "epoch": 30.79002079002079, + "grad_norm": 0.00034482585033401847, + "learning_rate": 0.20947486034308097, + "loss": 0.2827, + "num_input_tokens_seen": 11296440, + "step": 14810 + }, + { + "epoch": 30.8004158004158, + "grad_norm": 0.00020812460570596159, + "learning_rate": 0.2094207790260797, + "loss": 0.2655, + "num_input_tokens_seen": 11300120, + "step": 14815 + }, + { + "epoch": 30.81081081081081, + "grad_norm": 0.00023784405493643135, + "learning_rate": 0.20936668854564758, + "loss": 0.2549, + "num_input_tokens_seen": 11303928, + "step": 14820 + }, + { + "epoch": 30.821205821205822, + "grad_norm": 0.0002435097994748503, + "learning_rate": 0.20931258891012602, + "loss": 0.2623, + "num_input_tokens_seen": 11307672, + "step": 14825 + }, + { + "epoch": 30.83160083160083, + "grad_norm": 0.00035023706732317805, + "learning_rate": 0.20925848012785792, + "loss": 0.2715, + "num_input_tokens_seen": 11311448, + "step": 14830 + }, + { + "epoch": 30.841995841995843, + "grad_norm": 0.0004942963132634759, + "learning_rate": 0.20920436220718747, + "loss": 0.2658, + "num_input_tokens_seen": 11315192, + "step": 14835 + }, + { + "epoch": 30.85239085239085, + "grad_norm": 0.00016547893756069243, + "learning_rate": 0.20915023515646033, + "loss": 0.2668, + "num_input_tokens_seen": 11319128, + "step": 14840 + }, + { + "epoch": 30.862785862785863, + "grad_norm": 0.0004132789617870003, + "learning_rate": 0.20909609898402368, + "loss": 0.243, + "num_input_tokens_seen": 11322840, + "step": 14845 + }, + { + "epoch": 30.873180873180875, + "grad_norm": 0.00033745967084541917, + "learning_rate": 0.2090419536982258, + "loss": 0.2712, + "num_input_tokens_seen": 11326584, + "step": 14850 + }, + { + "epoch": 30.883575883575883, + "grad_norm": 0.000425832491600886, + "learning_rate": 0.2089877993074168, + "loss": 0.2436, + "num_input_tokens_seen": 11330616, + "step": 14855 + }, + { + "epoch": 30.893970893970895, + "grad_norm": 7.139249646570534e-05, + "learning_rate": 0.20893363581994784, + "loss": 0.2432, + "num_input_tokens_seen": 11334488, + "step": 14860 + }, + { + "epoch": 30.904365904365903, + "grad_norm": 0.000588972819969058, + "learning_rate": 0.2088794632441716, + "loss": 0.2519, + "num_input_tokens_seen": 11338200, + "step": 14865 + }, + { + "epoch": 30.914760914760915, + "grad_norm": 6.970312097109854e-05, + "learning_rate": 0.20882528158844219, + "loss": 0.276, + "num_input_tokens_seen": 11342072, + "step": 14870 + }, + { + "epoch": 30.925155925155924, + "grad_norm": 0.0005155346007086337, + "learning_rate": 0.20877109086111514, + "loss": 0.2696, + "num_input_tokens_seen": 11345688, + "step": 14875 + }, + { + "epoch": 30.935550935550935, + "grad_norm": 0.000572209304664284, + "learning_rate": 0.2087168910705473, + "loss": 0.2827, + "num_input_tokens_seen": 11349528, + "step": 14880 + }, + { + "epoch": 30.945945945945947, + "grad_norm": 5.021830293117091e-05, + "learning_rate": 0.208662682225097, + "loss": 0.2746, + "num_input_tokens_seen": 11353272, + "step": 14885 + }, + { + "epoch": 30.956340956340956, + "grad_norm": 0.0005160231376066804, + "learning_rate": 0.2086084643331239, + "loss": 0.2774, + "num_input_tokens_seen": 11357080, + "step": 14890 + }, + { + "epoch": 30.966735966735968, + "grad_norm": 0.00018559713498689234, + "learning_rate": 0.20855423740298906, + "loss": 0.2839, + "num_input_tokens_seen": 11360984, + "step": 14895 + }, + { + "epoch": 30.977130977130976, + "grad_norm": 7.044704398140311e-05, + "learning_rate": 0.208500001443055, + "loss": 0.2645, + "num_input_tokens_seen": 11364696, + "step": 14900 + }, + { + "epoch": 30.987525987525988, + "grad_norm": 0.00016494252486154437, + "learning_rate": 0.20844575646168553, + "loss": 0.2448, + "num_input_tokens_seen": 11368568, + "step": 14905 + }, + { + "epoch": 30.997920997921, + "grad_norm": 0.00010968758579110727, + "learning_rate": 0.20839150246724594, + "loss": 0.2737, + "num_input_tokens_seen": 11372376, + "step": 14910 + }, + { + "epoch": 31.008316008316008, + "grad_norm": 0.00016372599930036813, + "learning_rate": 0.20833723946810287, + "loss": 0.2649, + "num_input_tokens_seen": 11376080, + "step": 14915 + }, + { + "epoch": 31.01871101871102, + "grad_norm": 0.0007427406380884349, + "learning_rate": 0.20828296747262437, + "loss": 0.2638, + "num_input_tokens_seen": 11379856, + "step": 14920 + }, + { + "epoch": 31.02910602910603, + "grad_norm": 0.0005963361472822726, + "learning_rate": 0.20822868648917986, + "loss": 0.2714, + "num_input_tokens_seen": 11383664, + "step": 14925 + }, + { + "epoch": 31.03950103950104, + "grad_norm": 0.0003358885005582124, + "learning_rate": 0.20817439652614017, + "loss": 0.2511, + "num_input_tokens_seen": 11387440, + "step": 14930 + }, + { + "epoch": 31.04989604989605, + "grad_norm": 0.00011668734805425629, + "learning_rate": 0.20812009759187744, + "loss": 0.2909, + "num_input_tokens_seen": 11391280, + "step": 14935 + }, + { + "epoch": 31.06029106029106, + "grad_norm": 0.00025887053925544024, + "learning_rate": 0.2080657896947653, + "loss": 0.2524, + "num_input_tokens_seen": 11395152, + "step": 14940 + }, + { + "epoch": 31.070686070686072, + "grad_norm": 0.00021860796550754458, + "learning_rate": 0.2080114728431787, + "loss": 0.2712, + "num_input_tokens_seen": 11399024, + "step": 14945 + }, + { + "epoch": 31.08108108108108, + "grad_norm": 0.0001636134402360767, + "learning_rate": 0.20795714704549392, + "loss": 0.2585, + "num_input_tokens_seen": 11402960, + "step": 14950 + }, + { + "epoch": 31.091476091476093, + "grad_norm": 0.000590287905652076, + "learning_rate": 0.20790281231008875, + "loss": 0.2704, + "num_input_tokens_seen": 11406736, + "step": 14955 + }, + { + "epoch": 31.1018711018711, + "grad_norm": 0.000144617966725491, + "learning_rate": 0.20784846864534226, + "loss": 0.2705, + "num_input_tokens_seen": 11410704, + "step": 14960 + }, + { + "epoch": 31.112266112266113, + "grad_norm": 0.0011210811790078878, + "learning_rate": 0.20779411605963496, + "loss": 0.2793, + "num_input_tokens_seen": 11414480, + "step": 14965 + }, + { + "epoch": 31.12266112266112, + "grad_norm": 6.359563121804968e-05, + "learning_rate": 0.2077397545613487, + "loss": 0.2526, + "num_input_tokens_seen": 11418320, + "step": 14970 + }, + { + "epoch": 31.133056133056133, + "grad_norm": 0.0005163986934348941, + "learning_rate": 0.20768538415886661, + "loss": 0.272, + "num_input_tokens_seen": 11422128, + "step": 14975 + }, + { + "epoch": 31.143451143451145, + "grad_norm": 0.00028234333149157465, + "learning_rate": 0.20763100486057343, + "loss": 0.2513, + "num_input_tokens_seen": 11426096, + "step": 14980 + }, + { + "epoch": 31.153846153846153, + "grad_norm": 0.00027606377261690795, + "learning_rate": 0.20757661667485502, + "loss": 0.2627, + "num_input_tokens_seen": 11429808, + "step": 14985 + }, + { + "epoch": 31.164241164241165, + "grad_norm": 0.0005827684071846306, + "learning_rate": 0.2075222196100988, + "loss": 0.2672, + "num_input_tokens_seen": 11433520, + "step": 14990 + }, + { + "epoch": 31.174636174636174, + "grad_norm": 0.00015686712868046016, + "learning_rate": 0.20746781367469344, + "loss": 0.2604, + "num_input_tokens_seen": 11437264, + "step": 14995 + }, + { + "epoch": 31.185031185031185, + "grad_norm": 0.00039666733937337995, + "learning_rate": 0.207413398877029, + "loss": 0.2847, + "num_input_tokens_seen": 11441040, + "step": 15000 + }, + { + "epoch": 31.185031185031185, + "eval_loss": 0.24743162095546722, + "eval_runtime": 13.4157, + "eval_samples_per_second": 63.806, + "eval_steps_per_second": 15.951, + "num_input_tokens_seen": 11441040, + "step": 15000 + }, + { + "epoch": 31.195426195426194, + "grad_norm": 0.00020095422223675996, + "learning_rate": 0.20735897522549698, + "loss": 0.2329, + "num_input_tokens_seen": 11444976, + "step": 15005 + }, + { + "epoch": 31.205821205821206, + "grad_norm": 0.00011427220306359231, + "learning_rate": 0.2073045427284902, + "loss": 0.278, + "num_input_tokens_seen": 11448720, + "step": 15010 + }, + { + "epoch": 31.216216216216218, + "grad_norm": 0.0003669828874990344, + "learning_rate": 0.2072501013944027, + "loss": 0.2476, + "num_input_tokens_seen": 11452624, + "step": 15015 + }, + { + "epoch": 31.226611226611226, + "grad_norm": 0.0003313622728455812, + "learning_rate": 0.20719565123163017, + "loss": 0.2786, + "num_input_tokens_seen": 11456400, + "step": 15020 + }, + { + "epoch": 31.237006237006238, + "grad_norm": 0.0005790401482954621, + "learning_rate": 0.20714119224856944, + "loss": 0.2588, + "num_input_tokens_seen": 11460208, + "step": 15025 + }, + { + "epoch": 31.247401247401246, + "grad_norm": 0.00020884226250927895, + "learning_rate": 0.2070867244536188, + "loss": 0.2597, + "num_input_tokens_seen": 11464112, + "step": 15030 + }, + { + "epoch": 31.257796257796258, + "grad_norm": 0.00036518240813165903, + "learning_rate": 0.20703224785517785, + "loss": 0.2772, + "num_input_tokens_seen": 11467856, + "step": 15035 + }, + { + "epoch": 31.26819126819127, + "grad_norm": 0.0004984860424883664, + "learning_rate": 0.20697776246164754, + "loss": 0.2683, + "num_input_tokens_seen": 11471632, + "step": 15040 + }, + { + "epoch": 31.27858627858628, + "grad_norm": 0.00014108205505181104, + "learning_rate": 0.2069232682814303, + "loss": 0.2393, + "num_input_tokens_seen": 11475536, + "step": 15045 + }, + { + "epoch": 31.28898128898129, + "grad_norm": 0.0004306963237468153, + "learning_rate": 0.20686876532292972, + "loss": 0.2755, + "num_input_tokens_seen": 11479376, + "step": 15050 + }, + { + "epoch": 31.2993762993763, + "grad_norm": 0.0003535844443831593, + "learning_rate": 0.20681425359455083, + "loss": 0.2786, + "num_input_tokens_seen": 11483056, + "step": 15055 + }, + { + "epoch": 31.30977130977131, + "grad_norm": 0.0004925647517666221, + "learning_rate": 0.20675973310470008, + "loss": 0.2736, + "num_input_tokens_seen": 11486864, + "step": 15060 + }, + { + "epoch": 31.32016632016632, + "grad_norm": 0.0006180040654726326, + "learning_rate": 0.2067052038617852, + "loss": 0.2758, + "num_input_tokens_seen": 11490480, + "step": 15065 + }, + { + "epoch": 31.33056133056133, + "grad_norm": 0.0008966726600192487, + "learning_rate": 0.2066506658742153, + "loss": 0.285, + "num_input_tokens_seen": 11494288, + "step": 15070 + }, + { + "epoch": 31.340956340956343, + "grad_norm": 0.00029795506270602345, + "learning_rate": 0.20659611915040077, + "loss": 0.254, + "num_input_tokens_seen": 11498000, + "step": 15075 + }, + { + "epoch": 31.35135135135135, + "grad_norm": 0.0002672720293048769, + "learning_rate": 0.20654156369875348, + "loss": 0.2572, + "num_input_tokens_seen": 11501744, + "step": 15080 + }, + { + "epoch": 31.361746361746363, + "grad_norm": 0.0005981008871458471, + "learning_rate": 0.20648699952768648, + "loss": 0.2419, + "num_input_tokens_seen": 11505520, + "step": 15085 + }, + { + "epoch": 31.37214137214137, + "grad_norm": 6.665072578471154e-05, + "learning_rate": 0.20643242664561437, + "loss": 0.26, + "num_input_tokens_seen": 11509360, + "step": 15090 + }, + { + "epoch": 31.382536382536383, + "grad_norm": 0.00021264624956529588, + "learning_rate": 0.20637784506095277, + "loss": 0.2713, + "num_input_tokens_seen": 11513200, + "step": 15095 + }, + { + "epoch": 31.39293139293139, + "grad_norm": 0.0002655001007951796, + "learning_rate": 0.20632325478211908, + "loss": 0.2741, + "num_input_tokens_seen": 11516880, + "step": 15100 + }, + { + "epoch": 31.403326403326403, + "grad_norm": 0.0005713010323233902, + "learning_rate": 0.20626865581753165, + "loss": 0.2606, + "num_input_tokens_seen": 11520560, + "step": 15105 + }, + { + "epoch": 31.413721413721415, + "grad_norm": 0.0005730713601224124, + "learning_rate": 0.2062140481756104, + "loss": 0.2967, + "num_input_tokens_seen": 11524368, + "step": 15110 + }, + { + "epoch": 31.424116424116423, + "grad_norm": 0.0005838734796270728, + "learning_rate": 0.20615943186477648, + "loss": 0.2826, + "num_input_tokens_seen": 11528176, + "step": 15115 + }, + { + "epoch": 31.434511434511435, + "grad_norm": 0.0001749437506077811, + "learning_rate": 0.20610480689345242, + "loss": 0.272, + "num_input_tokens_seen": 11532016, + "step": 15120 + }, + { + "epoch": 31.444906444906444, + "grad_norm": 0.0004875649174209684, + "learning_rate": 0.2060501732700621, + "loss": 0.2556, + "num_input_tokens_seen": 11535920, + "step": 15125 + }, + { + "epoch": 31.455301455301456, + "grad_norm": 0.00014699202438350767, + "learning_rate": 0.20599553100303067, + "loss": 0.2763, + "num_input_tokens_seen": 11539664, + "step": 15130 + }, + { + "epoch": 31.465696465696467, + "grad_norm": 0.001029477920383215, + "learning_rate": 0.20594088010078465, + "loss": 0.2565, + "num_input_tokens_seen": 11543568, + "step": 15135 + }, + { + "epoch": 31.476091476091476, + "grad_norm": 0.00029580388218164444, + "learning_rate": 0.20588622057175196, + "loss": 0.2474, + "num_input_tokens_seen": 11547472, + "step": 15140 + }, + { + "epoch": 31.486486486486488, + "grad_norm": 0.00011070615437347442, + "learning_rate": 0.20583155242436177, + "loss": 0.2729, + "num_input_tokens_seen": 11551312, + "step": 15145 + }, + { + "epoch": 31.496881496881496, + "grad_norm": 0.0001253844820894301, + "learning_rate": 0.20577687566704453, + "loss": 0.2581, + "num_input_tokens_seen": 11555184, + "step": 15150 + }, + { + "epoch": 31.507276507276508, + "grad_norm": 0.00015230536519084126, + "learning_rate": 0.20572219030823213, + "loss": 0.2554, + "num_input_tokens_seen": 11559056, + "step": 15155 + }, + { + "epoch": 31.517671517671516, + "grad_norm": 0.0002877760271076113, + "learning_rate": 0.20566749635635775, + "loss": 0.2568, + "num_input_tokens_seen": 11562864, + "step": 15160 + }, + { + "epoch": 31.528066528066528, + "grad_norm": 0.0002018535597017035, + "learning_rate": 0.20561279381985587, + "loss": 0.289, + "num_input_tokens_seen": 11566704, + "step": 15165 + }, + { + "epoch": 31.53846153846154, + "grad_norm": 0.00036334648029878736, + "learning_rate": 0.2055580827071623, + "loss": 0.2707, + "num_input_tokens_seen": 11570512, + "step": 15170 + }, + { + "epoch": 31.54885654885655, + "grad_norm": 0.0006616413593292236, + "learning_rate": 0.20550336302671418, + "loss": 0.2698, + "num_input_tokens_seen": 11574320, + "step": 15175 + }, + { + "epoch": 31.55925155925156, + "grad_norm": 0.00012892454105895013, + "learning_rate": 0.20544863478695, + "loss": 0.2399, + "num_input_tokens_seen": 11578160, + "step": 15180 + }, + { + "epoch": 31.56964656964657, + "grad_norm": 0.0002216225693700835, + "learning_rate": 0.20539389799630953, + "loss": 0.2881, + "num_input_tokens_seen": 11582000, + "step": 15185 + }, + { + "epoch": 31.58004158004158, + "grad_norm": 0.00015289570728782564, + "learning_rate": 0.20533915266323388, + "loss": 0.2751, + "num_input_tokens_seen": 11585776, + "step": 15190 + }, + { + "epoch": 31.59043659043659, + "grad_norm": 0.0011692427797243, + "learning_rate": 0.20528439879616542, + "loss": 0.2604, + "num_input_tokens_seen": 11589552, + "step": 15195 + }, + { + "epoch": 31.6008316008316, + "grad_norm": 0.000329432834405452, + "learning_rate": 0.20522963640354794, + "loss": 0.2556, + "num_input_tokens_seen": 11593456, + "step": 15200 + }, + { + "epoch": 31.6008316008316, + "eval_loss": 0.25349685549736023, + "eval_runtime": 13.4167, + "eval_samples_per_second": 63.801, + "eval_steps_per_second": 15.95, + "num_input_tokens_seen": 11593456, + "step": 15200 + }, + { + "epoch": 31.611226611226613, + "grad_norm": 0.00016098761989269406, + "learning_rate": 0.20517486549382644, + "loss": 0.2759, + "num_input_tokens_seen": 11597232, + "step": 15205 + }, + { + "epoch": 31.62162162162162, + "grad_norm": 0.0003965785726904869, + "learning_rate": 0.20512008607544735, + "loss": 0.2576, + "num_input_tokens_seen": 11600816, + "step": 15210 + }, + { + "epoch": 31.632016632016633, + "grad_norm": 0.00016998022329062223, + "learning_rate": 0.20506529815685826, + "loss": 0.285, + "num_input_tokens_seen": 11604688, + "step": 15215 + }, + { + "epoch": 31.64241164241164, + "grad_norm": 0.0002748713013716042, + "learning_rate": 0.2050105017465082, + "loss": 0.2781, + "num_input_tokens_seen": 11608528, + "step": 15220 + }, + { + "epoch": 31.652806652806653, + "grad_norm": 2.1838486645719968e-05, + "learning_rate": 0.20495569685284754, + "loss": 0.2684, + "num_input_tokens_seen": 11612304, + "step": 15225 + }, + { + "epoch": 31.66320166320166, + "grad_norm": 0.0005906136939302087, + "learning_rate": 0.20490088348432778, + "loss": 0.2442, + "num_input_tokens_seen": 11616272, + "step": 15230 + }, + { + "epoch": 31.673596673596673, + "grad_norm": 0.00036783976247534156, + "learning_rate": 0.2048460616494018, + "loss": 0.2841, + "num_input_tokens_seen": 11620272, + "step": 15235 + }, + { + "epoch": 31.683991683991685, + "grad_norm": 0.0005082294810563326, + "learning_rate": 0.2047912313565239, + "loss": 0.2736, + "num_input_tokens_seen": 11624144, + "step": 15240 + }, + { + "epoch": 31.694386694386694, + "grad_norm": 8.031754987314343e-05, + "learning_rate": 0.20473639261414958, + "loss": 0.2546, + "num_input_tokens_seen": 11627888, + "step": 15245 + }, + { + "epoch": 31.704781704781706, + "grad_norm": 0.000283773522824049, + "learning_rate": 0.2046815454307357, + "loss": 0.2678, + "num_input_tokens_seen": 11631696, + "step": 15250 + }, + { + "epoch": 31.715176715176714, + "grad_norm": 7.317338895518333e-05, + "learning_rate": 0.20462668981474028, + "loss": 0.2565, + "num_input_tokens_seen": 11635472, + "step": 15255 + }, + { + "epoch": 31.725571725571726, + "grad_norm": 0.00017082841077353805, + "learning_rate": 0.20457182577462288, + "loss": 0.2636, + "num_input_tokens_seen": 11639248, + "step": 15260 + }, + { + "epoch": 31.735966735966738, + "grad_norm": 6.815803499193862e-05, + "learning_rate": 0.2045169533188441, + "loss": 0.2543, + "num_input_tokens_seen": 11643024, + "step": 15265 + }, + { + "epoch": 31.746361746361746, + "grad_norm": 6.431232031900436e-05, + "learning_rate": 0.20446207245586603, + "loss": 0.2748, + "num_input_tokens_seen": 11646832, + "step": 15270 + }, + { + "epoch": 31.756756756756758, + "grad_norm": 0.00018241336510982364, + "learning_rate": 0.20440718319415196, + "loss": 0.2701, + "num_input_tokens_seen": 11650480, + "step": 15275 + }, + { + "epoch": 31.767151767151766, + "grad_norm": 0.0007018996984697878, + "learning_rate": 0.20435228554216653, + "loss": 0.27, + "num_input_tokens_seen": 11654480, + "step": 15280 + }, + { + "epoch": 31.777546777546778, + "grad_norm": 0.0004037163162138313, + "learning_rate": 0.20429737950837565, + "loss": 0.2568, + "num_input_tokens_seen": 11658224, + "step": 15285 + }, + { + "epoch": 31.787941787941786, + "grad_norm": 0.00017245157505385578, + "learning_rate": 0.20424246510124647, + "loss": 0.2476, + "num_input_tokens_seen": 11662160, + "step": 15290 + }, + { + "epoch": 31.7983367983368, + "grad_norm": 0.0002721804194152355, + "learning_rate": 0.20418754232924755, + "loss": 0.2747, + "num_input_tokens_seen": 11665904, + "step": 15295 + }, + { + "epoch": 31.80873180873181, + "grad_norm": 0.00017300549370702356, + "learning_rate": 0.20413261120084863, + "loss": 0.2964, + "num_input_tokens_seen": 11669616, + "step": 15300 + }, + { + "epoch": 31.81912681912682, + "grad_norm": 0.0009127103257924318, + "learning_rate": 0.2040776717245208, + "loss": 0.2895, + "num_input_tokens_seen": 11673392, + "step": 15305 + }, + { + "epoch": 31.82952182952183, + "grad_norm": 0.0006942368927411735, + "learning_rate": 0.2040227239087364, + "loss": 0.2778, + "num_input_tokens_seen": 11677264, + "step": 15310 + }, + { + "epoch": 31.83991683991684, + "grad_norm": 0.0005713171558454633, + "learning_rate": 0.20396776776196904, + "loss": 0.2931, + "num_input_tokens_seen": 11680912, + "step": 15315 + }, + { + "epoch": 31.85031185031185, + "grad_norm": 0.000258445244980976, + "learning_rate": 0.20391280329269373, + "loss": 0.2958, + "num_input_tokens_seen": 11684752, + "step": 15320 + }, + { + "epoch": 31.86070686070686, + "grad_norm": 0.0003681774251163006, + "learning_rate": 0.20385783050938663, + "loss": 0.2811, + "num_input_tokens_seen": 11688496, + "step": 15325 + }, + { + "epoch": 31.87110187110187, + "grad_norm": 0.00021278730127960443, + "learning_rate": 0.20380284942052526, + "loss": 0.28, + "num_input_tokens_seen": 11692368, + "step": 15330 + }, + { + "epoch": 31.881496881496883, + "grad_norm": 0.00010583264520391822, + "learning_rate": 0.2037478600345884, + "loss": 0.2759, + "num_input_tokens_seen": 11696304, + "step": 15335 + }, + { + "epoch": 31.89189189189189, + "grad_norm": 0.0002900429244618863, + "learning_rate": 0.20369286236005604, + "loss": 0.2822, + "num_input_tokens_seen": 11700048, + "step": 15340 + }, + { + "epoch": 31.902286902286903, + "grad_norm": 9.922471508616582e-05, + "learning_rate": 0.20363785640540957, + "loss": 0.2824, + "num_input_tokens_seen": 11703856, + "step": 15345 + }, + { + "epoch": 31.91268191268191, + "grad_norm": 0.0008042150875553489, + "learning_rate": 0.2035828421791316, + "loss": 0.2742, + "num_input_tokens_seen": 11707792, + "step": 15350 + }, + { + "epoch": 31.923076923076923, + "grad_norm": 0.0001235621457453817, + "learning_rate": 0.20352781968970599, + "loss": 0.2751, + "num_input_tokens_seen": 11711536, + "step": 15355 + }, + { + "epoch": 31.933471933471935, + "grad_norm": 0.00010859921167138964, + "learning_rate": 0.2034727889456179, + "loss": 0.2441, + "num_input_tokens_seen": 11715280, + "step": 15360 + }, + { + "epoch": 31.943866943866944, + "grad_norm": 0.0003521858307067305, + "learning_rate": 0.2034177499553538, + "loss": 0.227, + "num_input_tokens_seen": 11719120, + "step": 15365 + }, + { + "epoch": 31.954261954261955, + "grad_norm": 0.00040172928129322827, + "learning_rate": 0.2033627027274014, + "loss": 0.3021, + "num_input_tokens_seen": 11722960, + "step": 15370 + }, + { + "epoch": 31.964656964656964, + "grad_norm": 0.0002521440328564495, + "learning_rate": 0.20330764727024955, + "loss": 0.2668, + "num_input_tokens_seen": 11726928, + "step": 15375 + }, + { + "epoch": 31.975051975051976, + "grad_norm": 0.00024207847309298813, + "learning_rate": 0.20325258359238868, + "loss": 0.2853, + "num_input_tokens_seen": 11730768, + "step": 15380 + }, + { + "epoch": 31.985446985446984, + "grad_norm": 0.0007389273960143328, + "learning_rate": 0.20319751170231018, + "loss": 0.2853, + "num_input_tokens_seen": 11734608, + "step": 15385 + }, + { + "epoch": 31.995841995841996, + "grad_norm": 0.0003433914389461279, + "learning_rate": 0.2031424316085068, + "loss": 0.2678, + "num_input_tokens_seen": 11738192, + "step": 15390 + }, + { + "epoch": 32.00623700623701, + "grad_norm": 0.0002961848513223231, + "learning_rate": 0.20308734331947265, + "loss": 0.2892, + "num_input_tokens_seen": 11742032, + "step": 15395 + }, + { + "epoch": 32.016632016632016, + "grad_norm": 0.00013911539281252772, + "learning_rate": 0.20303224684370305, + "loss": 0.2735, + "num_input_tokens_seen": 11745744, + "step": 15400 + }, + { + "epoch": 32.016632016632016, + "eval_loss": 0.2534793019294739, + "eval_runtime": 13.4433, + "eval_samples_per_second": 63.675, + "eval_steps_per_second": 15.919, + "num_input_tokens_seen": 11745744, + "step": 15400 + }, + { + "epoch": 32.027027027027025, + "grad_norm": 0.0009543925989419222, + "learning_rate": 0.20297714218969456, + "loss": 0.248, + "num_input_tokens_seen": 11749392, + "step": 15405 + }, + { + "epoch": 32.03742203742204, + "grad_norm": 0.0006277076900005341, + "learning_rate": 0.20292202936594497, + "loss": 0.2533, + "num_input_tokens_seen": 11753168, + "step": 15410 + }, + { + "epoch": 32.04781704781705, + "grad_norm": 0.00030200681067071855, + "learning_rate": 0.2028669083809534, + "loss": 0.2815, + "num_input_tokens_seen": 11756912, + "step": 15415 + }, + { + "epoch": 32.05821205821206, + "grad_norm": 0.00011790819553425536, + "learning_rate": 0.20281177924322016, + "loss": 0.2587, + "num_input_tokens_seen": 11760688, + "step": 15420 + }, + { + "epoch": 32.06860706860707, + "grad_norm": 0.00010677181126084179, + "learning_rate": 0.2027566419612469, + "loss": 0.2824, + "num_input_tokens_seen": 11764656, + "step": 15425 + }, + { + "epoch": 32.07900207900208, + "grad_norm": 0.0005335345631465316, + "learning_rate": 0.20270149654353647, + "loss": 0.2707, + "num_input_tokens_seen": 11768592, + "step": 15430 + }, + { + "epoch": 32.08939708939709, + "grad_norm": 0.0002658010052982718, + "learning_rate": 0.202646342998593, + "loss": 0.271, + "num_input_tokens_seen": 11772400, + "step": 15435 + }, + { + "epoch": 32.0997920997921, + "grad_norm": 0.0003651408478617668, + "learning_rate": 0.20259118133492185, + "loss": 0.2264, + "num_input_tokens_seen": 11776336, + "step": 15440 + }, + { + "epoch": 32.11018711018711, + "grad_norm": 0.00028817771817557514, + "learning_rate": 0.20253601156102966, + "loss": 0.2816, + "num_input_tokens_seen": 11780080, + "step": 15445 + }, + { + "epoch": 32.12058212058212, + "grad_norm": 0.00010571320308372378, + "learning_rate": 0.20248083368542422, + "loss": 0.281, + "num_input_tokens_seen": 11783792, + "step": 15450 + }, + { + "epoch": 32.13097713097713, + "grad_norm": 5.7641031162347645e-05, + "learning_rate": 0.2024256477166147, + "loss": 0.2666, + "num_input_tokens_seen": 11787568, + "step": 15455 + }, + { + "epoch": 32.141372141372145, + "grad_norm": 0.00021898458362556994, + "learning_rate": 0.2023704536631115, + "loss": 0.24, + "num_input_tokens_seen": 11791440, + "step": 15460 + }, + { + "epoch": 32.15176715176715, + "grad_norm": 0.0003960388421546668, + "learning_rate": 0.20231525153342625, + "loss": 0.3096, + "num_input_tokens_seen": 11795216, + "step": 15465 + }, + { + "epoch": 32.16216216216216, + "grad_norm": 5.404434341471642e-05, + "learning_rate": 0.20226004133607173, + "loss": 0.2396, + "num_input_tokens_seen": 11799088, + "step": 15470 + }, + { + "epoch": 32.17255717255717, + "grad_norm": 0.00022807875939179212, + "learning_rate": 0.20220482307956214, + "loss": 0.2699, + "num_input_tokens_seen": 11802928, + "step": 15475 + }, + { + "epoch": 32.182952182952185, + "grad_norm": 0.000115324801299721, + "learning_rate": 0.20214959677241276, + "loss": 0.262, + "num_input_tokens_seen": 11806768, + "step": 15480 + }, + { + "epoch": 32.19334719334719, + "grad_norm": 0.00021686071704607457, + "learning_rate": 0.20209436242314022, + "loss": 0.271, + "num_input_tokens_seen": 11810640, + "step": 15485 + }, + { + "epoch": 32.2037422037422, + "grad_norm": 4.4735646952176467e-05, + "learning_rate": 0.2020391200402623, + "loss": 0.2635, + "num_input_tokens_seen": 11814352, + "step": 15490 + }, + { + "epoch": 32.21413721413722, + "grad_norm": 0.00015896240074653178, + "learning_rate": 0.2019838696322981, + "loss": 0.2733, + "num_input_tokens_seen": 11818160, + "step": 15495 + }, + { + "epoch": 32.224532224532226, + "grad_norm": 0.00027150404639542103, + "learning_rate": 0.20192861120776798, + "loss": 0.2821, + "num_input_tokens_seen": 11821904, + "step": 15500 + }, + { + "epoch": 32.234927234927234, + "grad_norm": 0.0004229968471918255, + "learning_rate": 0.20187334477519345, + "loss": 0.26, + "num_input_tokens_seen": 11825776, + "step": 15505 + }, + { + "epoch": 32.24532224532224, + "grad_norm": 3.722673864103854e-05, + "learning_rate": 0.20181807034309726, + "loss": 0.2624, + "num_input_tokens_seen": 11829584, + "step": 15510 + }, + { + "epoch": 32.25571725571726, + "grad_norm": 0.0002588228671811521, + "learning_rate": 0.2017627879200034, + "loss": 0.2364, + "num_input_tokens_seen": 11833360, + "step": 15515 + }, + { + "epoch": 32.266112266112266, + "grad_norm": 0.00022604198602493852, + "learning_rate": 0.2017074975144372, + "loss": 0.2469, + "num_input_tokens_seen": 11837264, + "step": 15520 + }, + { + "epoch": 32.276507276507274, + "grad_norm": 0.0003075867425650358, + "learning_rate": 0.20165219913492508, + "loss": 0.2455, + "num_input_tokens_seen": 11841168, + "step": 15525 + }, + { + "epoch": 32.28690228690229, + "grad_norm": 0.00024371856125071645, + "learning_rate": 0.20159689278999468, + "loss": 0.2517, + "num_input_tokens_seen": 11845168, + "step": 15530 + }, + { + "epoch": 32.2972972972973, + "grad_norm": 0.0008227110374718904, + "learning_rate": 0.20154157848817508, + "loss": 0.2804, + "num_input_tokens_seen": 11849104, + "step": 15535 + }, + { + "epoch": 32.30769230769231, + "grad_norm": 4.614240970113315e-05, + "learning_rate": 0.20148625623799632, + "loss": 0.2647, + "num_input_tokens_seen": 11852816, + "step": 15540 + }, + { + "epoch": 32.318087318087315, + "grad_norm": 5.6565393606433645e-05, + "learning_rate": 0.20143092604798984, + "loss": 0.288, + "num_input_tokens_seen": 11856816, + "step": 15545 + }, + { + "epoch": 32.32848232848233, + "grad_norm": 0.0008120402344502509, + "learning_rate": 0.2013755879266883, + "loss": 0.2788, + "num_input_tokens_seen": 11860496, + "step": 15550 + }, + { + "epoch": 32.33887733887734, + "grad_norm": 0.00021874764934182167, + "learning_rate": 0.20132024188262543, + "loss": 0.243, + "num_input_tokens_seen": 11864400, + "step": 15555 + }, + { + "epoch": 32.34927234927235, + "grad_norm": 0.0005347630358301103, + "learning_rate": 0.2012648879243363, + "loss": 0.3496, + "num_input_tokens_seen": 11868432, + "step": 15560 + }, + { + "epoch": 32.35966735966736, + "grad_norm": 0.0001643247960601002, + "learning_rate": 0.20120952606035725, + "loss": 0.2784, + "num_input_tokens_seen": 11872496, + "step": 15565 + }, + { + "epoch": 32.37006237006237, + "grad_norm": 0.00032321701291948557, + "learning_rate": 0.20115415629922576, + "loss": 0.278, + "num_input_tokens_seen": 11876240, + "step": 15570 + }, + { + "epoch": 32.38045738045738, + "grad_norm": 0.00019577352213673294, + "learning_rate": 0.20109877864948048, + "loss": 0.2826, + "num_input_tokens_seen": 11880144, + "step": 15575 + }, + { + "epoch": 32.39085239085239, + "grad_norm": 0.00010374561679782346, + "learning_rate": 0.20104339311966138, + "loss": 0.2743, + "num_input_tokens_seen": 11883920, + "step": 15580 + }, + { + "epoch": 32.4012474012474, + "grad_norm": 3.9047976315487176e-05, + "learning_rate": 0.2009879997183097, + "loss": 0.2534, + "num_input_tokens_seen": 11887664, + "step": 15585 + }, + { + "epoch": 32.41164241164241, + "grad_norm": 0.00010887360258493572, + "learning_rate": 0.20093259845396763, + "loss": 0.2867, + "num_input_tokens_seen": 11891312, + "step": 15590 + }, + { + "epoch": 32.42203742203742, + "grad_norm": 0.00020036347268614918, + "learning_rate": 0.20087718933517884, + "loss": 0.2589, + "num_input_tokens_seen": 11895024, + "step": 15595 + }, + { + "epoch": 32.432432432432435, + "grad_norm": 0.00023753606365062296, + "learning_rate": 0.20082177237048807, + "loss": 0.2402, + "num_input_tokens_seen": 11898672, + "step": 15600 + }, + { + "epoch": 32.432432432432435, + "eval_loss": 0.24875333905220032, + "eval_runtime": 13.4591, + "eval_samples_per_second": 63.6, + "eval_steps_per_second": 15.9, + "num_input_tokens_seen": 11898672, + "step": 15600 + }, + { + "epoch": 32.44282744282744, + "grad_norm": 0.0006049730582162738, + "learning_rate": 0.20076634756844133, + "loss": 0.273, + "num_input_tokens_seen": 11902576, + "step": 15605 + }, + { + "epoch": 32.45322245322245, + "grad_norm": 3.909133010893129e-05, + "learning_rate": 0.20071091493758586, + "loss": 0.2845, + "num_input_tokens_seen": 11906352, + "step": 15610 + }, + { + "epoch": 32.46361746361746, + "grad_norm": 0.00044083455577492714, + "learning_rate": 0.20065547448647003, + "loss": 0.2658, + "num_input_tokens_seen": 11910224, + "step": 15615 + }, + { + "epoch": 32.474012474012476, + "grad_norm": 0.0003770831972360611, + "learning_rate": 0.20060002622364348, + "loss": 0.2748, + "num_input_tokens_seen": 11914032, + "step": 15620 + }, + { + "epoch": 32.484407484407484, + "grad_norm": 9.661898366175592e-05, + "learning_rate": 0.20054457015765695, + "loss": 0.2807, + "num_input_tokens_seen": 11917872, + "step": 15625 + }, + { + "epoch": 32.49480249480249, + "grad_norm": 0.00025012672995217144, + "learning_rate": 0.20048910629706254, + "loss": 0.256, + "num_input_tokens_seen": 11921744, + "step": 15630 + }, + { + "epoch": 32.50519750519751, + "grad_norm": 0.00043226455454714596, + "learning_rate": 0.20043363465041347, + "loss": 0.2669, + "num_input_tokens_seen": 11925520, + "step": 15635 + }, + { + "epoch": 32.515592515592516, + "grad_norm": 0.0002737355825956911, + "learning_rate": 0.2003781552262641, + "loss": 0.2364, + "num_input_tokens_seen": 11929328, + "step": 15640 + }, + { + "epoch": 32.525987525987524, + "grad_norm": 5.1621307648019865e-05, + "learning_rate": 0.20032266803317014, + "loss": 0.2543, + "num_input_tokens_seen": 11932944, + "step": 15645 + }, + { + "epoch": 32.53638253638254, + "grad_norm": 0.0003747539594769478, + "learning_rate": 0.2002671730796884, + "loss": 0.2612, + "num_input_tokens_seen": 11936784, + "step": 15650 + }, + { + "epoch": 32.54677754677755, + "grad_norm": 0.0002851150056812912, + "learning_rate": 0.20021167037437684, + "loss": 0.2771, + "num_input_tokens_seen": 11940688, + "step": 15655 + }, + { + "epoch": 32.55717255717256, + "grad_norm": 0.00021103833569213748, + "learning_rate": 0.20015615992579472, + "loss": 0.256, + "num_input_tokens_seen": 11944464, + "step": 15660 + }, + { + "epoch": 32.567567567567565, + "grad_norm": 0.000189159472938627, + "learning_rate": 0.20010064174250244, + "loss": 0.2711, + "num_input_tokens_seen": 11948272, + "step": 15665 + }, + { + "epoch": 32.57796257796258, + "grad_norm": 0.00024098555149976164, + "learning_rate": 0.2000451158330616, + "loss": 0.2668, + "num_input_tokens_seen": 11952048, + "step": 15670 + }, + { + "epoch": 32.58835758835759, + "grad_norm": 0.0008048078743740916, + "learning_rate": 0.199989582206035, + "loss": 0.2857, + "num_input_tokens_seen": 11955952, + "step": 15675 + }, + { + "epoch": 32.5987525987526, + "grad_norm": 0.00024033478985074908, + "learning_rate": 0.1999340408699866, + "loss": 0.287, + "num_input_tokens_seen": 11959792, + "step": 15680 + }, + { + "epoch": 32.60914760914761, + "grad_norm": 0.00038019142812117934, + "learning_rate": 0.19987849183348155, + "loss": 0.3033, + "num_input_tokens_seen": 11963536, + "step": 15685 + }, + { + "epoch": 32.61954261954262, + "grad_norm": 9.447165939491242e-05, + "learning_rate": 0.19982293510508628, + "loss": 0.2773, + "num_input_tokens_seen": 11967280, + "step": 15690 + }, + { + "epoch": 32.62993762993763, + "grad_norm": 0.00020353558647911996, + "learning_rate": 0.19976737069336833, + "loss": 0.2607, + "num_input_tokens_seen": 11971152, + "step": 15695 + }, + { + "epoch": 32.64033264033264, + "grad_norm": 0.00034135454916395247, + "learning_rate": 0.1997117986068964, + "loss": 0.2845, + "num_input_tokens_seen": 11974896, + "step": 15700 + }, + { + "epoch": 32.65072765072765, + "grad_norm": 5.22316295246128e-05, + "learning_rate": 0.19965621885424037, + "loss": 0.2509, + "num_input_tokens_seen": 11978800, + "step": 15705 + }, + { + "epoch": 32.66112266112266, + "grad_norm": 0.00016469572437927127, + "learning_rate": 0.19960063144397142, + "loss": 0.2807, + "num_input_tokens_seen": 11982608, + "step": 15710 + }, + { + "epoch": 32.67151767151767, + "grad_norm": 0.00048100523417815566, + "learning_rate": 0.19954503638466176, + "loss": 0.2685, + "num_input_tokens_seen": 11986416, + "step": 15715 + }, + { + "epoch": 32.681912681912685, + "grad_norm": 0.00011516644735820591, + "learning_rate": 0.1994894336848848, + "loss": 0.2641, + "num_input_tokens_seen": 11990448, + "step": 15720 + }, + { + "epoch": 32.69230769230769, + "grad_norm": 9.864517778623849e-05, + "learning_rate": 0.1994338233532153, + "loss": 0.2601, + "num_input_tokens_seen": 11994288, + "step": 15725 + }, + { + "epoch": 32.7027027027027, + "grad_norm": 0.00024225196102634072, + "learning_rate": 0.19937820539822904, + "loss": 0.2788, + "num_input_tokens_seen": 11998064, + "step": 15730 + }, + { + "epoch": 32.71309771309771, + "grad_norm": 0.0001712695084279403, + "learning_rate": 0.199322579828503, + "loss": 0.2751, + "num_input_tokens_seen": 12001936, + "step": 15735 + }, + { + "epoch": 32.723492723492726, + "grad_norm": 9.223847882822156e-05, + "learning_rate": 0.19926694665261527, + "loss": 0.2713, + "num_input_tokens_seen": 12005648, + "step": 15740 + }, + { + "epoch": 32.733887733887734, + "grad_norm": 5.836924538016319e-05, + "learning_rate": 0.19921130587914526, + "loss": 0.2653, + "num_input_tokens_seen": 12009328, + "step": 15745 + }, + { + "epoch": 32.74428274428274, + "grad_norm": 7.56628141971305e-05, + "learning_rate": 0.19915565751667344, + "loss": 0.245, + "num_input_tokens_seen": 12013104, + "step": 15750 + }, + { + "epoch": 32.75467775467776, + "grad_norm": 0.0001046677862177603, + "learning_rate": 0.19910000157378152, + "loss": 0.2104, + "num_input_tokens_seen": 12017168, + "step": 15755 + }, + { + "epoch": 32.765072765072766, + "grad_norm": 0.0005254389834590256, + "learning_rate": 0.1990443380590523, + "loss": 0.302, + "num_input_tokens_seen": 12021040, + "step": 15760 + }, + { + "epoch": 32.775467775467774, + "grad_norm": 0.0003612090367823839, + "learning_rate": 0.19898866698106984, + "loss": 0.2889, + "num_input_tokens_seen": 12024720, + "step": 15765 + }, + { + "epoch": 32.78586278586278, + "grad_norm": 0.0006334806093946099, + "learning_rate": 0.19893298834841933, + "loss": 0.2735, + "num_input_tokens_seen": 12028528, + "step": 15770 + }, + { + "epoch": 32.7962577962578, + "grad_norm": 0.000236947278608568, + "learning_rate": 0.19887730216968705, + "loss": 0.2539, + "num_input_tokens_seen": 12032336, + "step": 15775 + }, + { + "epoch": 32.80665280665281, + "grad_norm": 0.00017690783715806901, + "learning_rate": 0.19882160845346053, + "loss": 0.2571, + "num_input_tokens_seen": 12036016, + "step": 15780 + }, + { + "epoch": 32.817047817047815, + "grad_norm": 5.5965025239856914e-05, + "learning_rate": 0.1987659072083285, + "loss": 0.2886, + "num_input_tokens_seen": 12039664, + "step": 15785 + }, + { + "epoch": 32.82744282744283, + "grad_norm": 0.0001633249339647591, + "learning_rate": 0.1987101984428807, + "loss": 0.2701, + "num_input_tokens_seen": 12043376, + "step": 15790 + }, + { + "epoch": 32.83783783783784, + "grad_norm": 0.0004956405027769506, + "learning_rate": 0.19865448216570822, + "loss": 0.2699, + "num_input_tokens_seen": 12047184, + "step": 15795 + }, + { + "epoch": 32.84823284823285, + "grad_norm": 0.00033626865479163826, + "learning_rate": 0.19859875838540317, + "loss": 0.2652, + "num_input_tokens_seen": 12050992, + "step": 15800 + }, + { + "epoch": 32.84823284823285, + "eval_loss": 0.25051096081733704, + "eval_runtime": 13.4256, + "eval_samples_per_second": 63.759, + "eval_steps_per_second": 15.94, + "num_input_tokens_seen": 12050992, + "step": 15800 + }, + { + "epoch": 32.858627858627855, + "grad_norm": 0.0002648533845786005, + "learning_rate": 0.1985430271105588, + "loss": 0.2834, + "num_input_tokens_seen": 12054768, + "step": 15805 + }, + { + "epoch": 32.86902286902287, + "grad_norm": 0.0002831656893249601, + "learning_rate": 0.19848728834976961, + "loss": 0.271, + "num_input_tokens_seen": 12058576, + "step": 15810 + }, + { + "epoch": 32.87941787941788, + "grad_norm": 0.00017496984219178557, + "learning_rate": 0.19843154211163128, + "loss": 0.2614, + "num_input_tokens_seen": 12062352, + "step": 15815 + }, + { + "epoch": 32.88981288981289, + "grad_norm": 0.0007717504049651325, + "learning_rate": 0.1983757884047405, + "loss": 0.2706, + "num_input_tokens_seen": 12066288, + "step": 15820 + }, + { + "epoch": 32.9002079002079, + "grad_norm": 0.0007676717359572649, + "learning_rate": 0.1983200272376952, + "loss": 0.2932, + "num_input_tokens_seen": 12070096, + "step": 15825 + }, + { + "epoch": 32.91060291060291, + "grad_norm": 0.0004217039095237851, + "learning_rate": 0.1982642586190945, + "loss": 0.2688, + "num_input_tokens_seen": 12073712, + "step": 15830 + }, + { + "epoch": 32.92099792099792, + "grad_norm": 0.0005239245947450399, + "learning_rate": 0.1982084825575386, + "loss": 0.2668, + "num_input_tokens_seen": 12077680, + "step": 15835 + }, + { + "epoch": 32.931392931392935, + "grad_norm": 5.8531808463158086e-05, + "learning_rate": 0.19815269906162883, + "loss": 0.2697, + "num_input_tokens_seen": 12081456, + "step": 15840 + }, + { + "epoch": 32.94178794178794, + "grad_norm": 0.0001866299717221409, + "learning_rate": 0.19809690813996775, + "loss": 0.2928, + "num_input_tokens_seen": 12085200, + "step": 15845 + }, + { + "epoch": 32.95218295218295, + "grad_norm": 0.00029696914134547114, + "learning_rate": 0.19804110980115905, + "loss": 0.2737, + "num_input_tokens_seen": 12088912, + "step": 15850 + }, + { + "epoch": 32.96257796257796, + "grad_norm": 0.00037159433122724295, + "learning_rate": 0.19798530405380746, + "loss": 0.2744, + "num_input_tokens_seen": 12092656, + "step": 15855 + }, + { + "epoch": 32.972972972972975, + "grad_norm": 0.00012550615065265447, + "learning_rate": 0.19792949090651893, + "loss": 0.2785, + "num_input_tokens_seen": 12096464, + "step": 15860 + }, + { + "epoch": 32.983367983367984, + "grad_norm": 0.00012656039325520396, + "learning_rate": 0.19787367036790066, + "loss": 0.2527, + "num_input_tokens_seen": 12100464, + "step": 15865 + }, + { + "epoch": 32.99376299376299, + "grad_norm": 0.00039187405491247773, + "learning_rate": 0.19781784244656075, + "loss": 0.2645, + "num_input_tokens_seen": 12104176, + "step": 15870 + }, + { + "epoch": 33.00415800415801, + "grad_norm": 0.00022895254369359463, + "learning_rate": 0.19776200715110864, + "loss": 0.2509, + "num_input_tokens_seen": 12108064, + "step": 15875 + }, + { + "epoch": 33.014553014553016, + "grad_norm": 0.0001250521745532751, + "learning_rate": 0.1977061644901548, + "loss": 0.2638, + "num_input_tokens_seen": 12111936, + "step": 15880 + }, + { + "epoch": 33.024948024948024, + "grad_norm": 0.0001927202829392627, + "learning_rate": 0.1976503144723109, + "loss": 0.2559, + "num_input_tokens_seen": 12115936, + "step": 15885 + }, + { + "epoch": 33.03534303534303, + "grad_norm": 0.0002640201710164547, + "learning_rate": 0.19759445710618967, + "loss": 0.2782, + "num_input_tokens_seen": 12119744, + "step": 15890 + }, + { + "epoch": 33.04573804573805, + "grad_norm": 6.730109453201294e-05, + "learning_rate": 0.19753859240040508, + "loss": 0.2609, + "num_input_tokens_seen": 12123456, + "step": 15895 + }, + { + "epoch": 33.056133056133056, + "grad_norm": 0.0004990263259969652, + "learning_rate": 0.1974827203635721, + "loss": 0.2742, + "num_input_tokens_seen": 12127328, + "step": 15900 + }, + { + "epoch": 33.066528066528065, + "grad_norm": 0.00037125975359231234, + "learning_rate": 0.19742684100430694, + "loss": 0.2644, + "num_input_tokens_seen": 12130944, + "step": 15905 + }, + { + "epoch": 33.07692307692308, + "grad_norm": 0.0002494606014806777, + "learning_rate": 0.19737095433122692, + "loss": 0.2755, + "num_input_tokens_seen": 12134880, + "step": 15910 + }, + { + "epoch": 33.08731808731809, + "grad_norm": 0.00015916707343421876, + "learning_rate": 0.19731506035295046, + "loss": 0.2821, + "num_input_tokens_seen": 12138752, + "step": 15915 + }, + { + "epoch": 33.0977130977131, + "grad_norm": 0.00013659498654305935, + "learning_rate": 0.19725915907809702, + "loss": 0.2696, + "num_input_tokens_seen": 12142784, + "step": 15920 + }, + { + "epoch": 33.108108108108105, + "grad_norm": 0.0005090643535368145, + "learning_rate": 0.1972032505152874, + "loss": 0.2726, + "num_input_tokens_seen": 12146592, + "step": 15925 + }, + { + "epoch": 33.11850311850312, + "grad_norm": 0.00010413699055789039, + "learning_rate": 0.19714733467314338, + "loss": 0.2501, + "num_input_tokens_seen": 12150496, + "step": 15930 + }, + { + "epoch": 33.12889812889813, + "grad_norm": 0.00035530258901417255, + "learning_rate": 0.19709141156028784, + "loss": 0.2694, + "num_input_tokens_seen": 12154304, + "step": 15935 + }, + { + "epoch": 33.13929313929314, + "grad_norm": 9.427600161870942e-05, + "learning_rate": 0.1970354811853448, + "loss": 0.2683, + "num_input_tokens_seen": 12158208, + "step": 15940 + }, + { + "epoch": 33.14968814968815, + "grad_norm": 9.188500553136691e-05, + "learning_rate": 0.19697954355693953, + "loss": 0.2702, + "num_input_tokens_seen": 12162048, + "step": 15945 + }, + { + "epoch": 33.16008316008316, + "grad_norm": 0.00042223732452839613, + "learning_rate": 0.19692359868369827, + "loss": 0.2654, + "num_input_tokens_seen": 12165824, + "step": 15950 + }, + { + "epoch": 33.17047817047817, + "grad_norm": 0.0004130545712541789, + "learning_rate": 0.1968676465742484, + "loss": 0.2708, + "num_input_tokens_seen": 12169696, + "step": 15955 + }, + { + "epoch": 33.18087318087318, + "grad_norm": 3.997397652710788e-05, + "learning_rate": 0.19681168723721845, + "loss": 0.2678, + "num_input_tokens_seen": 12173632, + "step": 15960 + }, + { + "epoch": 33.19126819126819, + "grad_norm": 0.0002635691489558667, + "learning_rate": 0.19675572068123803, + "loss": 0.2668, + "num_input_tokens_seen": 12177408, + "step": 15965 + }, + { + "epoch": 33.2016632016632, + "grad_norm": 0.00045208234223537147, + "learning_rate": 0.19669974691493794, + "loss": 0.2528, + "num_input_tokens_seen": 12181440, + "step": 15970 + }, + { + "epoch": 33.21205821205821, + "grad_norm": 0.0003312194312456995, + "learning_rate": 0.19664376594695002, + "loss": 0.27, + "num_input_tokens_seen": 12185312, + "step": 15975 + }, + { + "epoch": 33.222453222453225, + "grad_norm": 0.0004204189754091203, + "learning_rate": 0.19658777778590722, + "loss": 0.2541, + "num_input_tokens_seen": 12189120, + "step": 15980 + }, + { + "epoch": 33.232848232848234, + "grad_norm": 0.0001763259497238323, + "learning_rate": 0.19653178244044364, + "loss": 0.273, + "num_input_tokens_seen": 12192928, + "step": 15985 + }, + { + "epoch": 33.24324324324324, + "grad_norm": 0.00014383697998709977, + "learning_rate": 0.19647577991919443, + "loss": 0.2636, + "num_input_tokens_seen": 12196800, + "step": 15990 + }, + { + "epoch": 33.25363825363825, + "grad_norm": 0.00033120953594334424, + "learning_rate": 0.1964197702307959, + "loss": 0.2732, + "num_input_tokens_seen": 12200736, + "step": 15995 + }, + { + "epoch": 33.264033264033266, + "grad_norm": 0.0004515945620369166, + "learning_rate": 0.19636375338388545, + "loss": 0.246, + "num_input_tokens_seen": 12204352, + "step": 16000 + }, + { + "epoch": 33.264033264033266, + "eval_loss": 0.24759477376937866, + "eval_runtime": 13.413, + "eval_samples_per_second": 63.819, + "eval_steps_per_second": 15.955, + "num_input_tokens_seen": 12204352, + "step": 16000 + }, + { + "epoch": 33.274428274428274, + "grad_norm": 0.00024289079010486603, + "learning_rate": 0.1963077293871016, + "loss": 0.237, + "num_input_tokens_seen": 12208096, + "step": 16005 + }, + { + "epoch": 33.28482328482328, + "grad_norm": 0.00017072846821974963, + "learning_rate": 0.19625169824908395, + "loss": 0.3022, + "num_input_tokens_seen": 12211872, + "step": 16010 + }, + { + "epoch": 33.2952182952183, + "grad_norm": 0.0003880539152305573, + "learning_rate": 0.19619565997847319, + "loss": 0.2555, + "num_input_tokens_seen": 12215552, + "step": 16015 + }, + { + "epoch": 33.305613305613306, + "grad_norm": 0.0007254658848978579, + "learning_rate": 0.19613961458391113, + "loss": 0.2775, + "num_input_tokens_seen": 12219328, + "step": 16020 + }, + { + "epoch": 33.316008316008315, + "grad_norm": 0.00028293346986174583, + "learning_rate": 0.19608356207404065, + "loss": 0.2609, + "num_input_tokens_seen": 12223136, + "step": 16025 + }, + { + "epoch": 33.32640332640332, + "grad_norm": 0.0003068010846618563, + "learning_rate": 0.1960275024575058, + "loss": 0.2805, + "num_input_tokens_seen": 12226976, + "step": 16030 + }, + { + "epoch": 33.33679833679834, + "grad_norm": 0.00012259873619768769, + "learning_rate": 0.19597143574295164, + "loss": 0.2849, + "num_input_tokens_seen": 12230688, + "step": 16035 + }, + { + "epoch": 33.34719334719335, + "grad_norm": 0.00020578320254571736, + "learning_rate": 0.1959153619390244, + "loss": 0.2757, + "num_input_tokens_seen": 12234368, + "step": 16040 + }, + { + "epoch": 33.357588357588355, + "grad_norm": 0.0004884267109446228, + "learning_rate": 0.1958592810543713, + "loss": 0.2579, + "num_input_tokens_seen": 12238112, + "step": 16045 + }, + { + "epoch": 33.36798336798337, + "grad_norm": 0.00028216184000484645, + "learning_rate": 0.19580319309764077, + "loss": 0.2612, + "num_input_tokens_seen": 12241952, + "step": 16050 + }, + { + "epoch": 33.37837837837838, + "grad_norm": 0.0006488080834969878, + "learning_rate": 0.1957470980774823, + "loss": 0.3055, + "num_input_tokens_seen": 12245696, + "step": 16055 + }, + { + "epoch": 33.38877338877339, + "grad_norm": 0.000672056688927114, + "learning_rate": 0.19569099600254639, + "loss": 0.2836, + "num_input_tokens_seen": 12249408, + "step": 16060 + }, + { + "epoch": 33.3991683991684, + "grad_norm": 0.00028896002913825214, + "learning_rate": 0.1956348868814847, + "loss": 0.2745, + "num_input_tokens_seen": 12253280, + "step": 16065 + }, + { + "epoch": 33.40956340956341, + "grad_norm": 9.332145418738946e-05, + "learning_rate": 0.19557877072295, + "loss": 0.2799, + "num_input_tokens_seen": 12257184, + "step": 16070 + }, + { + "epoch": 33.41995841995842, + "grad_norm": 0.00045819731894880533, + "learning_rate": 0.19552264753559603, + "loss": 0.2628, + "num_input_tokens_seen": 12260864, + "step": 16075 + }, + { + "epoch": 33.43035343035343, + "grad_norm": 0.0001537467906018719, + "learning_rate": 0.19546651732807774, + "loss": 0.2539, + "num_input_tokens_seen": 12264672, + "step": 16080 + }, + { + "epoch": 33.44074844074844, + "grad_norm": 6.919514999026433e-05, + "learning_rate": 0.19541038010905112, + "loss": 0.2827, + "num_input_tokens_seen": 12268512, + "step": 16085 + }, + { + "epoch": 33.45114345114345, + "grad_norm": 0.0002537929394748062, + "learning_rate": 0.19535423588717324, + "loss": 0.2559, + "num_input_tokens_seen": 12272288, + "step": 16090 + }, + { + "epoch": 33.46153846153846, + "grad_norm": 0.00010120011575054377, + "learning_rate": 0.19529808467110224, + "loss": 0.2592, + "num_input_tokens_seen": 12276128, + "step": 16095 + }, + { + "epoch": 33.471933471933475, + "grad_norm": 0.0002596256381366402, + "learning_rate": 0.19524192646949734, + "loss": 0.2395, + "num_input_tokens_seen": 12279872, + "step": 16100 + }, + { + "epoch": 33.482328482328484, + "grad_norm": 0.0003985886287409812, + "learning_rate": 0.19518576129101878, + "loss": 0.2812, + "num_input_tokens_seen": 12283744, + "step": 16105 + }, + { + "epoch": 33.49272349272349, + "grad_norm": 0.0002144862082786858, + "learning_rate": 0.19512958914432804, + "loss": 0.2531, + "num_input_tokens_seen": 12287520, + "step": 16110 + }, + { + "epoch": 33.5031185031185, + "grad_norm": 6.764783756807446e-05, + "learning_rate": 0.1950734100380875, + "loss": 0.251, + "num_input_tokens_seen": 12291168, + "step": 16115 + }, + { + "epoch": 33.513513513513516, + "grad_norm": 0.0002308595139766112, + "learning_rate": 0.19501722398096066, + "loss": 0.2647, + "num_input_tokens_seen": 12295104, + "step": 16120 + }, + { + "epoch": 33.523908523908524, + "grad_norm": 0.00019253579375799745, + "learning_rate": 0.1949610309816122, + "loss": 0.2747, + "num_input_tokens_seen": 12299008, + "step": 16125 + }, + { + "epoch": 33.53430353430353, + "grad_norm": 0.00011218348663533106, + "learning_rate": 0.1949048310487078, + "loss": 0.2533, + "num_input_tokens_seen": 12302880, + "step": 16130 + }, + { + "epoch": 33.54469854469855, + "grad_norm": 0.00014726178778801113, + "learning_rate": 0.19484862419091406, + "loss": 0.2355, + "num_input_tokens_seen": 12306720, + "step": 16135 + }, + { + "epoch": 33.555093555093556, + "grad_norm": 7.352221291512251e-05, + "learning_rate": 0.19479241041689893, + "loss": 0.2741, + "num_input_tokens_seen": 12310560, + "step": 16140 + }, + { + "epoch": 33.565488565488565, + "grad_norm": 0.00012450353824533522, + "learning_rate": 0.19473618973533116, + "loss": 0.2783, + "num_input_tokens_seen": 12314400, + "step": 16145 + }, + { + "epoch": 33.57588357588357, + "grad_norm": 0.00011042161349905655, + "learning_rate": 0.19467996215488076, + "loss": 0.2704, + "num_input_tokens_seen": 12318240, + "step": 16150 + }, + { + "epoch": 33.58627858627859, + "grad_norm": 0.00021205050870776176, + "learning_rate": 0.1946237276842187, + "loss": 0.2718, + "num_input_tokens_seen": 12322016, + "step": 16155 + }, + { + "epoch": 33.5966735966736, + "grad_norm": 0.00013148492143955082, + "learning_rate": 0.19456748633201712, + "loss": 0.2787, + "num_input_tokens_seen": 12325824, + "step": 16160 + }, + { + "epoch": 33.607068607068605, + "grad_norm": 0.0007922688964754343, + "learning_rate": 0.194511238106949, + "loss": 0.2618, + "num_input_tokens_seen": 12329536, + "step": 16165 + }, + { + "epoch": 33.61746361746362, + "grad_norm": 4.1973460611188784e-05, + "learning_rate": 0.19445498301768863, + "loss": 0.2629, + "num_input_tokens_seen": 12333504, + "step": 16170 + }, + { + "epoch": 33.62785862785863, + "grad_norm": 0.00040032173274084926, + "learning_rate": 0.19439872107291126, + "loss": 0.2779, + "num_input_tokens_seen": 12337280, + "step": 16175 + }, + { + "epoch": 33.63825363825364, + "grad_norm": 0.00044166078441776335, + "learning_rate": 0.1943424522812931, + "loss": 0.2551, + "num_input_tokens_seen": 12341248, + "step": 16180 + }, + { + "epoch": 33.648648648648646, + "grad_norm": 0.0006496984860859811, + "learning_rate": 0.19428617665151157, + "loss": 0.2751, + "num_input_tokens_seen": 12345088, + "step": 16185 + }, + { + "epoch": 33.65904365904366, + "grad_norm": 0.0002717013703659177, + "learning_rate": 0.19422989419224507, + "loss": 0.2737, + "num_input_tokens_seen": 12348832, + "step": 16190 + }, + { + "epoch": 33.66943866943867, + "grad_norm": 0.0003359794500283897, + "learning_rate": 0.19417360491217303, + "loss": 0.2846, + "num_input_tokens_seen": 12352512, + "step": 16195 + }, + { + "epoch": 33.67983367983368, + "grad_norm": 0.00025236851070076227, + "learning_rate": 0.19411730881997605, + "loss": 0.2549, + "num_input_tokens_seen": 12356224, + "step": 16200 + }, + { + "epoch": 33.67983367983368, + "eval_loss": 0.24902480840682983, + "eval_runtime": 13.407, + "eval_samples_per_second": 63.847, + "eval_steps_per_second": 15.962, + "num_input_tokens_seen": 12356224, + "step": 16200 + }, + { + "epoch": 33.69022869022869, + "grad_norm": 0.00020736752776429057, + "learning_rate": 0.1940610059243356, + "loss": 0.2754, + "num_input_tokens_seen": 12360064, + "step": 16205 + }, + { + "epoch": 33.7006237006237, + "grad_norm": 7.55802757339552e-05, + "learning_rate": 0.19400469623393435, + "loss": 0.2654, + "num_input_tokens_seen": 12363552, + "step": 16210 + }, + { + "epoch": 33.71101871101871, + "grad_norm": 0.00015491966041736305, + "learning_rate": 0.1939483797574559, + "loss": 0.2691, + "num_input_tokens_seen": 12367424, + "step": 16215 + }, + { + "epoch": 33.72141372141372, + "grad_norm": 0.0004810387035831809, + "learning_rate": 0.19389205650358504, + "loss": 0.2451, + "num_input_tokens_seen": 12371008, + "step": 16220 + }, + { + "epoch": 33.731808731808734, + "grad_norm": 0.0003228721907362342, + "learning_rate": 0.19383572648100747, + "loss": 0.2646, + "num_input_tokens_seen": 12374816, + "step": 16225 + }, + { + "epoch": 33.74220374220374, + "grad_norm": 0.00026009109569713473, + "learning_rate": 0.19377938969841, + "loss": 0.2606, + "num_input_tokens_seen": 12378560, + "step": 16230 + }, + { + "epoch": 33.75259875259875, + "grad_norm": 0.0005737391184084117, + "learning_rate": 0.1937230461644805, + "loss": 0.2765, + "num_input_tokens_seen": 12382304, + "step": 16235 + }, + { + "epoch": 33.762993762993766, + "grad_norm": 0.0009987344965338707, + "learning_rate": 0.19366669588790777, + "loss": 0.2743, + "num_input_tokens_seen": 12386176, + "step": 16240 + }, + { + "epoch": 33.773388773388774, + "grad_norm": 0.0007878433098085225, + "learning_rate": 0.19361033887738185, + "loss": 0.2658, + "num_input_tokens_seen": 12389984, + "step": 16245 + }, + { + "epoch": 33.78378378378378, + "grad_norm": 0.00023748567036818713, + "learning_rate": 0.19355397514159361, + "loss": 0.2733, + "num_input_tokens_seen": 12393728, + "step": 16250 + }, + { + "epoch": 33.79417879417879, + "grad_norm": 0.00047703864402137697, + "learning_rate": 0.19349760468923508, + "loss": 0.2902, + "num_input_tokens_seen": 12397504, + "step": 16255 + }, + { + "epoch": 33.804573804573806, + "grad_norm": 0.00029075733618810773, + "learning_rate": 0.19344122752899925, + "loss": 0.2684, + "num_input_tokens_seen": 12401408, + "step": 16260 + }, + { + "epoch": 33.814968814968815, + "grad_norm": 0.00013549902359955013, + "learning_rate": 0.1933848436695802, + "loss": 0.2684, + "num_input_tokens_seen": 12405152, + "step": 16265 + }, + { + "epoch": 33.82536382536382, + "grad_norm": 0.0005686871591024101, + "learning_rate": 0.1933284531196731, + "loss": 0.2695, + "num_input_tokens_seen": 12409088, + "step": 16270 + }, + { + "epoch": 33.83575883575884, + "grad_norm": 0.0006737210787832737, + "learning_rate": 0.19327205588797403, + "loss": 0.2731, + "num_input_tokens_seen": 12412896, + "step": 16275 + }, + { + "epoch": 33.84615384615385, + "grad_norm": 7.488125993404537e-05, + "learning_rate": 0.19321565198318014, + "loss": 0.2744, + "num_input_tokens_seen": 12416704, + "step": 16280 + }, + { + "epoch": 33.856548856548855, + "grad_norm": 0.0001727318303892389, + "learning_rate": 0.1931592414139896, + "loss": 0.2702, + "num_input_tokens_seen": 12420608, + "step": 16285 + }, + { + "epoch": 33.86694386694387, + "grad_norm": 0.00054693385027349, + "learning_rate": 0.19310282418910169, + "loss": 0.2769, + "num_input_tokens_seen": 12424416, + "step": 16290 + }, + { + "epoch": 33.87733887733888, + "grad_norm": 0.00027967069763690233, + "learning_rate": 0.1930464003172166, + "loss": 0.2691, + "num_input_tokens_seen": 12428288, + "step": 16295 + }, + { + "epoch": 33.88773388773389, + "grad_norm": 0.0005294065340422094, + "learning_rate": 0.19298996980703567, + "loss": 0.2305, + "num_input_tokens_seen": 12432128, + "step": 16300 + }, + { + "epoch": 33.898128898128896, + "grad_norm": 9.152175334747881e-05, + "learning_rate": 0.19293353266726113, + "loss": 0.2522, + "num_input_tokens_seen": 12435872, + "step": 16305 + }, + { + "epoch": 33.90852390852391, + "grad_norm": 0.0005343279335647821, + "learning_rate": 0.19287708890659633, + "loss": 0.3056, + "num_input_tokens_seen": 12439744, + "step": 16310 + }, + { + "epoch": 33.91891891891892, + "grad_norm": 0.0006683728424832225, + "learning_rate": 0.19282063853374556, + "loss": 0.2509, + "num_input_tokens_seen": 12443584, + "step": 16315 + }, + { + "epoch": 33.92931392931393, + "grad_norm": 0.0003453294048085809, + "learning_rate": 0.19276418155741423, + "loss": 0.2641, + "num_input_tokens_seen": 12447456, + "step": 16320 + }, + { + "epoch": 33.93970893970894, + "grad_norm": 0.0001162965054390952, + "learning_rate": 0.19270771798630867, + "loss": 0.272, + "num_input_tokens_seen": 12451168, + "step": 16325 + }, + { + "epoch": 33.95010395010395, + "grad_norm": 0.00010587320866761729, + "learning_rate": 0.1926512478291363, + "loss": 0.2643, + "num_input_tokens_seen": 12454976, + "step": 16330 + }, + { + "epoch": 33.96049896049896, + "grad_norm": 0.00019101159705314785, + "learning_rate": 0.19259477109460557, + "loss": 0.2562, + "num_input_tokens_seen": 12458752, + "step": 16335 + }, + { + "epoch": 33.97089397089397, + "grad_norm": 0.00017973108333535492, + "learning_rate": 0.19253828779142584, + "loss": 0.2439, + "num_input_tokens_seen": 12462432, + "step": 16340 + }, + { + "epoch": 33.981288981288984, + "grad_norm": 0.00026924567646346986, + "learning_rate": 0.19248179792830755, + "loss": 0.2518, + "num_input_tokens_seen": 12466176, + "step": 16345 + }, + { + "epoch": 33.99168399168399, + "grad_norm": 0.0005582004669122398, + "learning_rate": 0.19242530151396217, + "loss": 0.3083, + "num_input_tokens_seen": 12469984, + "step": 16350 + }, + { + "epoch": 34.002079002079, + "grad_norm": 0.0005495326477102935, + "learning_rate": 0.19236879855710215, + "loss": 0.2918, + "num_input_tokens_seen": 12473720, + "step": 16355 + }, + { + "epoch": 34.012474012474016, + "grad_norm": 0.00022900123440194875, + "learning_rate": 0.19231228906644096, + "loss": 0.2733, + "num_input_tokens_seen": 12477592, + "step": 16360 + }, + { + "epoch": 34.022869022869024, + "grad_norm": 0.0001909696584334597, + "learning_rate": 0.19225577305069302, + "loss": 0.2685, + "num_input_tokens_seen": 12481496, + "step": 16365 + }, + { + "epoch": 34.03326403326403, + "grad_norm": 0.00014395166363101453, + "learning_rate": 0.1921992505185739, + "loss": 0.2554, + "num_input_tokens_seen": 12485304, + "step": 16370 + }, + { + "epoch": 34.04365904365904, + "grad_norm": 0.000253660895396024, + "learning_rate": 0.19214272147880004, + "loss": 0.291, + "num_input_tokens_seen": 12489144, + "step": 16375 + }, + { + "epoch": 34.054054054054056, + "grad_norm": 0.0006493324763141572, + "learning_rate": 0.19208618594008892, + "loss": 0.2526, + "num_input_tokens_seen": 12492824, + "step": 16380 + }, + { + "epoch": 34.064449064449065, + "grad_norm": 0.00018270219152327627, + "learning_rate": 0.19202964391115904, + "loss": 0.2479, + "num_input_tokens_seen": 12496696, + "step": 16385 + }, + { + "epoch": 34.07484407484407, + "grad_norm": 7.166177238104865e-05, + "learning_rate": 0.1919730954007299, + "loss": 0.2932, + "num_input_tokens_seen": 12500440, + "step": 16390 + }, + { + "epoch": 34.08523908523909, + "grad_norm": 0.0004000802000518888, + "learning_rate": 0.19191654041752199, + "loss": 0.2493, + "num_input_tokens_seen": 12504216, + "step": 16395 + }, + { + "epoch": 34.0956340956341, + "grad_norm": 0.0001537659263703972, + "learning_rate": 0.19185997897025678, + "loss": 0.2714, + "num_input_tokens_seen": 12507960, + "step": 16400 + }, + { + "epoch": 34.0956340956341, + "eval_loss": 0.2524295747280121, + "eval_runtime": 13.406, + "eval_samples_per_second": 63.852, + "eval_steps_per_second": 15.963, + "num_input_tokens_seen": 12507960, + "step": 16400 + }, + { + "epoch": 34.106029106029105, + "grad_norm": 0.00023281040193978697, + "learning_rate": 0.19180341106765672, + "loss": 0.2728, + "num_input_tokens_seen": 12511768, + "step": 16405 + }, + { + "epoch": 34.11642411642411, + "grad_norm": 0.00026537213125266135, + "learning_rate": 0.19174683671844536, + "loss": 0.2591, + "num_input_tokens_seen": 12515608, + "step": 16410 + }, + { + "epoch": 34.12681912681913, + "grad_norm": 0.00015339165111072361, + "learning_rate": 0.19169025593134717, + "loss": 0.2317, + "num_input_tokens_seen": 12519416, + "step": 16415 + }, + { + "epoch": 34.13721413721414, + "grad_norm": 0.0006768080638721585, + "learning_rate": 0.19163366871508764, + "loss": 0.268, + "num_input_tokens_seen": 12523256, + "step": 16420 + }, + { + "epoch": 34.147609147609145, + "grad_norm": 0.0002964687591884285, + "learning_rate": 0.19157707507839317, + "loss": 0.2544, + "num_input_tokens_seen": 12527064, + "step": 16425 + }, + { + "epoch": 34.15800415800416, + "grad_norm": 0.0001422977657057345, + "learning_rate": 0.19152047502999123, + "loss": 0.2576, + "num_input_tokens_seen": 12530808, + "step": 16430 + }, + { + "epoch": 34.16839916839917, + "grad_norm": 0.0005182041204534471, + "learning_rate": 0.19146386857861025, + "loss": 0.2613, + "num_input_tokens_seen": 12534648, + "step": 16435 + }, + { + "epoch": 34.17879417879418, + "grad_norm": 0.000290843250695616, + "learning_rate": 0.19140725573297968, + "loss": 0.262, + "num_input_tokens_seen": 12538680, + "step": 16440 + }, + { + "epoch": 34.189189189189186, + "grad_norm": 0.00018275891488883644, + "learning_rate": 0.19135063650182987, + "loss": 0.2645, + "num_input_tokens_seen": 12542712, + "step": 16445 + }, + { + "epoch": 34.1995841995842, + "grad_norm": 6.097853838582523e-05, + "learning_rate": 0.19129401089389234, + "loss": 0.2881, + "num_input_tokens_seen": 12546328, + "step": 16450 + }, + { + "epoch": 34.20997920997921, + "grad_norm": 6.842829316155985e-05, + "learning_rate": 0.19123737891789938, + "loss": 0.259, + "num_input_tokens_seen": 12550232, + "step": 16455 + }, + { + "epoch": 34.22037422037422, + "grad_norm": 6.534743442898616e-05, + "learning_rate": 0.19118074058258439, + "loss": 0.2771, + "num_input_tokens_seen": 12554232, + "step": 16460 + }, + { + "epoch": 34.23076923076923, + "grad_norm": 0.00034814924583770335, + "learning_rate": 0.1911240958966816, + "loss": 0.2845, + "num_input_tokens_seen": 12557944, + "step": 16465 + }, + { + "epoch": 34.24116424116424, + "grad_norm": 0.0008581766160205007, + "learning_rate": 0.19106744486892652, + "loss": 0.281, + "num_input_tokens_seen": 12561816, + "step": 16470 + }, + { + "epoch": 34.25155925155925, + "grad_norm": 0.00030134874396026134, + "learning_rate": 0.1910107875080553, + "loss": 0.2623, + "num_input_tokens_seen": 12565624, + "step": 16475 + }, + { + "epoch": 34.26195426195426, + "grad_norm": 4.233239087625407e-05, + "learning_rate": 0.19095412382280533, + "loss": 0.2928, + "num_input_tokens_seen": 12569496, + "step": 16480 + }, + { + "epoch": 34.272349272349274, + "grad_norm": 0.0003774478973355144, + "learning_rate": 0.19089745382191473, + "loss": 0.2662, + "num_input_tokens_seen": 12573240, + "step": 16485 + }, + { + "epoch": 34.28274428274428, + "grad_norm": 0.00011659821029752493, + "learning_rate": 0.19084077751412284, + "loss": 0.2777, + "num_input_tokens_seen": 12577016, + "step": 16490 + }, + { + "epoch": 34.29313929313929, + "grad_norm": 0.0007570674642920494, + "learning_rate": 0.19078409490816986, + "loss": 0.2725, + "num_input_tokens_seen": 12580856, + "step": 16495 + }, + { + "epoch": 34.303534303534306, + "grad_norm": 0.00016461701306980103, + "learning_rate": 0.19072740601279686, + "loss": 0.2654, + "num_input_tokens_seen": 12584568, + "step": 16500 + }, + { + "epoch": 34.313929313929314, + "grad_norm": 0.00023914442863315344, + "learning_rate": 0.19067071083674605, + "loss": 0.253, + "num_input_tokens_seen": 12588440, + "step": 16505 + }, + { + "epoch": 34.32432432432432, + "grad_norm": 0.0001310602092416957, + "learning_rate": 0.19061400938876052, + "loss": 0.2662, + "num_input_tokens_seen": 12592440, + "step": 16510 + }, + { + "epoch": 34.33471933471934, + "grad_norm": 0.00012074862752342597, + "learning_rate": 0.1905573016775844, + "loss": 0.2617, + "num_input_tokens_seen": 12596248, + "step": 16515 + }, + { + "epoch": 34.34511434511435, + "grad_norm": 0.00018509275105316192, + "learning_rate": 0.19050058771196263, + "loss": 0.2731, + "num_input_tokens_seen": 12600120, + "step": 16520 + }, + { + "epoch": 34.355509355509355, + "grad_norm": 0.00012660337961278856, + "learning_rate": 0.19044386750064132, + "loss": 0.2721, + "num_input_tokens_seen": 12603992, + "step": 16525 + }, + { + "epoch": 34.36590436590436, + "grad_norm": 0.00027666555251926184, + "learning_rate": 0.19038714105236737, + "loss": 0.2645, + "num_input_tokens_seen": 12607704, + "step": 16530 + }, + { + "epoch": 34.37629937629938, + "grad_norm": 0.0004880151245743036, + "learning_rate": 0.19033040837588874, + "loss": 0.2834, + "num_input_tokens_seen": 12611576, + "step": 16535 + }, + { + "epoch": 34.38669438669439, + "grad_norm": 0.00014753633877262473, + "learning_rate": 0.1902736694799543, + "loss": 0.2735, + "num_input_tokens_seen": 12615416, + "step": 16540 + }, + { + "epoch": 34.397089397089395, + "grad_norm": 0.00010072724398924038, + "learning_rate": 0.19021692437331392, + "loss": 0.2784, + "num_input_tokens_seen": 12619352, + "step": 16545 + }, + { + "epoch": 34.40748440748441, + "grad_norm": 0.00035135503276251256, + "learning_rate": 0.1901601730647184, + "loss": 0.2753, + "num_input_tokens_seen": 12623128, + "step": 16550 + }, + { + "epoch": 34.41787941787942, + "grad_norm": 3.114005448878743e-05, + "learning_rate": 0.19010341556291954, + "loss": 0.2651, + "num_input_tokens_seen": 12626904, + "step": 16555 + }, + { + "epoch": 34.42827442827443, + "grad_norm": 0.00038274715188890696, + "learning_rate": 0.19004665187667, + "loss": 0.2641, + "num_input_tokens_seen": 12630616, + "step": 16560 + }, + { + "epoch": 34.438669438669436, + "grad_norm": 0.0006404719897545874, + "learning_rate": 0.1899898820147235, + "loss": 0.2965, + "num_input_tokens_seen": 12634424, + "step": 16565 + }, + { + "epoch": 34.44906444906445, + "grad_norm": 0.0005553974770009518, + "learning_rate": 0.18993310598583465, + "loss": 0.2688, + "num_input_tokens_seen": 12638168, + "step": 16570 + }, + { + "epoch": 34.45945945945946, + "grad_norm": 0.0005463711568154395, + "learning_rate": 0.18987632379875904, + "loss": 0.2703, + "num_input_tokens_seen": 12641848, + "step": 16575 + }, + { + "epoch": 34.46985446985447, + "grad_norm": 3.965325959143229e-05, + "learning_rate": 0.18981953546225314, + "loss": 0.2487, + "num_input_tokens_seen": 12645528, + "step": 16580 + }, + { + "epoch": 34.48024948024948, + "grad_norm": 0.00010026699601439759, + "learning_rate": 0.18976274098507445, + "loss": 0.2767, + "num_input_tokens_seen": 12649464, + "step": 16585 + }, + { + "epoch": 34.49064449064449, + "grad_norm": 0.00048031070036813617, + "learning_rate": 0.18970594037598146, + "loss": 0.2474, + "num_input_tokens_seen": 12653208, + "step": 16590 + }, + { + "epoch": 34.5010395010395, + "grad_norm": 2.514071820769459e-05, + "learning_rate": 0.1896491336437335, + "loss": 0.2733, + "num_input_tokens_seen": 12656984, + "step": 16595 + }, + { + "epoch": 34.51143451143451, + "grad_norm": 0.00013399054296314716, + "learning_rate": 0.18959232079709085, + "loss": 0.2736, + "num_input_tokens_seen": 12660760, + "step": 16600 + }, + { + "epoch": 34.51143451143451, + "eval_loss": 0.2697230279445648, + "eval_runtime": 13.4038, + "eval_samples_per_second": 63.863, + "eval_steps_per_second": 15.966, + "num_input_tokens_seen": 12660760, + "step": 16600 + }, + { + "epoch": 34.521829521829524, + "grad_norm": 6.736852810718119e-05, + "learning_rate": 0.18953550184481477, + "loss": 0.2837, + "num_input_tokens_seen": 12664696, + "step": 16605 + }, + { + "epoch": 34.53222453222453, + "grad_norm": 0.00013937293260823935, + "learning_rate": 0.18947867679566752, + "loss": 0.2821, + "num_input_tokens_seen": 12668440, + "step": 16610 + }, + { + "epoch": 34.54261954261954, + "grad_norm": 0.00017785966338124126, + "learning_rate": 0.18942184565841216, + "loss": 0.2664, + "num_input_tokens_seen": 12672312, + "step": 16615 + }, + { + "epoch": 34.553014553014556, + "grad_norm": 5.3962445235811174e-05, + "learning_rate": 0.18936500844181278, + "loss": 0.2685, + "num_input_tokens_seen": 12676248, + "step": 16620 + }, + { + "epoch": 34.563409563409564, + "grad_norm": 0.0004858898464590311, + "learning_rate": 0.18930816515463436, + "loss": 0.2561, + "num_input_tokens_seen": 12680248, + "step": 16625 + }, + { + "epoch": 34.57380457380457, + "grad_norm": 0.0002345779212191701, + "learning_rate": 0.18925131580564297, + "loss": 0.2598, + "num_input_tokens_seen": 12684088, + "step": 16630 + }, + { + "epoch": 34.58419958419958, + "grad_norm": 0.0001271579967578873, + "learning_rate": 0.1891944604036054, + "loss": 0.2515, + "num_input_tokens_seen": 12687832, + "step": 16635 + }, + { + "epoch": 34.5945945945946, + "grad_norm": 4.2805033444892615e-05, + "learning_rate": 0.1891375989572895, + "loss": 0.2558, + "num_input_tokens_seen": 12691864, + "step": 16640 + }, + { + "epoch": 34.604989604989605, + "grad_norm": 0.00033412614720873535, + "learning_rate": 0.18908073147546398, + "loss": 0.2776, + "num_input_tokens_seen": 12695704, + "step": 16645 + }, + { + "epoch": 34.61538461538461, + "grad_norm": 0.00019627295841928571, + "learning_rate": 0.18902385796689858, + "loss": 0.2755, + "num_input_tokens_seen": 12699576, + "step": 16650 + }, + { + "epoch": 34.62577962577963, + "grad_norm": 0.00036488103796727955, + "learning_rate": 0.18896697844036384, + "loss": 0.2654, + "num_input_tokens_seen": 12703480, + "step": 16655 + }, + { + "epoch": 34.63617463617464, + "grad_norm": 0.00023361347848549485, + "learning_rate": 0.18891009290463137, + "loss": 0.2622, + "num_input_tokens_seen": 12707256, + "step": 16660 + }, + { + "epoch": 34.646569646569645, + "grad_norm": 0.0003277365758549422, + "learning_rate": 0.18885320136847353, + "loss": 0.2731, + "num_input_tokens_seen": 12711160, + "step": 16665 + }, + { + "epoch": 34.656964656964654, + "grad_norm": 0.0001076085027307272, + "learning_rate": 0.1887963038406639, + "loss": 0.2585, + "num_input_tokens_seen": 12714936, + "step": 16670 + }, + { + "epoch": 34.66735966735967, + "grad_norm": 0.00013675668742507696, + "learning_rate": 0.18873940032997658, + "loss": 0.2735, + "num_input_tokens_seen": 12718808, + "step": 16675 + }, + { + "epoch": 34.67775467775468, + "grad_norm": 0.00033890947815962136, + "learning_rate": 0.18868249084518693, + "loss": 0.2647, + "num_input_tokens_seen": 12722584, + "step": 16680 + }, + { + "epoch": 34.688149688149686, + "grad_norm": 0.0001570408494444564, + "learning_rate": 0.18862557539507102, + "loss": 0.2535, + "num_input_tokens_seen": 12726424, + "step": 16685 + }, + { + "epoch": 34.6985446985447, + "grad_norm": 0.00013297348050400615, + "learning_rate": 0.18856865398840605, + "loss": 0.2454, + "num_input_tokens_seen": 12730168, + "step": 16690 + }, + { + "epoch": 34.70893970893971, + "grad_norm": 5.4561322031077e-05, + "learning_rate": 0.18851172663396995, + "loss": 0.2616, + "num_input_tokens_seen": 12733848, + "step": 16695 + }, + { + "epoch": 34.71933471933472, + "grad_norm": 0.0005349882412701845, + "learning_rate": 0.1884547933405416, + "loss": 0.2754, + "num_input_tokens_seen": 12737560, + "step": 16700 + }, + { + "epoch": 34.729729729729726, + "grad_norm": 0.00012064033944625407, + "learning_rate": 0.1883978541169009, + "loss": 0.2677, + "num_input_tokens_seen": 12741432, + "step": 16705 + }, + { + "epoch": 34.74012474012474, + "grad_norm": 7.742404704913497e-05, + "learning_rate": 0.18834090897182854, + "loss": 0.2625, + "num_input_tokens_seen": 12745240, + "step": 16710 + }, + { + "epoch": 34.75051975051975, + "grad_norm": 0.0003235915210098028, + "learning_rate": 0.1882839579141062, + "loss": 0.2692, + "num_input_tokens_seen": 12749080, + "step": 16715 + }, + { + "epoch": 34.76091476091476, + "grad_norm": 0.0003752792254090309, + "learning_rate": 0.18822700095251646, + "loss": 0.2434, + "num_input_tokens_seen": 12752792, + "step": 16720 + }, + { + "epoch": 34.771309771309774, + "grad_norm": 0.0003952011466026306, + "learning_rate": 0.18817003809584273, + "loss": 0.2678, + "num_input_tokens_seen": 12756632, + "step": 16725 + }, + { + "epoch": 34.78170478170478, + "grad_norm": 8.917592640500516e-05, + "learning_rate": 0.1881130693528695, + "loss": 0.2819, + "num_input_tokens_seen": 12760536, + "step": 16730 + }, + { + "epoch": 34.79209979209979, + "grad_norm": 0.0004670246271416545, + "learning_rate": 0.18805609473238197, + "loss": 0.2826, + "num_input_tokens_seen": 12764248, + "step": 16735 + }, + { + "epoch": 34.802494802494806, + "grad_norm": 0.0005638967268168926, + "learning_rate": 0.18799911424316643, + "loss": 0.2786, + "num_input_tokens_seen": 12768056, + "step": 16740 + }, + { + "epoch": 34.812889812889814, + "grad_norm": 0.00013777521962765604, + "learning_rate": 0.18794212789400994, + "loss": 0.2675, + "num_input_tokens_seen": 12771896, + "step": 16745 + }, + { + "epoch": 34.82328482328482, + "grad_norm": 0.00013420867617242038, + "learning_rate": 0.18788513569370052, + "loss": 0.2589, + "num_input_tokens_seen": 12775576, + "step": 16750 + }, + { + "epoch": 34.83367983367983, + "grad_norm": 0.0007713295053690672, + "learning_rate": 0.1878281376510271, + "loss": 0.2974, + "num_input_tokens_seen": 12779288, + "step": 16755 + }, + { + "epoch": 34.84407484407485, + "grad_norm": 0.0002932752249762416, + "learning_rate": 0.18777113377477941, + "loss": 0.2837, + "num_input_tokens_seen": 12783128, + "step": 16760 + }, + { + "epoch": 34.854469854469855, + "grad_norm": 0.0009167591924779117, + "learning_rate": 0.1877141240737483, + "loss": 0.3014, + "num_input_tokens_seen": 12786904, + "step": 16765 + }, + { + "epoch": 34.86486486486486, + "grad_norm": 0.00026592848007567227, + "learning_rate": 0.18765710855672527, + "loss": 0.2812, + "num_input_tokens_seen": 12790712, + "step": 16770 + }, + { + "epoch": 34.87525987525988, + "grad_norm": 0.00012960981985088438, + "learning_rate": 0.18760008723250288, + "loss": 0.2517, + "num_input_tokens_seen": 12794328, + "step": 16775 + }, + { + "epoch": 34.88565488565489, + "grad_norm": 0.0004121332603972405, + "learning_rate": 0.18754306010987457, + "loss": 0.2611, + "num_input_tokens_seen": 12798200, + "step": 16780 + }, + { + "epoch": 34.896049896049895, + "grad_norm": 0.0001139149462687783, + "learning_rate": 0.18748602719763457, + "loss": 0.2674, + "num_input_tokens_seen": 12801976, + "step": 16785 + }, + { + "epoch": 34.906444906444904, + "grad_norm": 9.393378422828391e-05, + "learning_rate": 0.18742898850457804, + "loss": 0.2575, + "num_input_tokens_seen": 12805816, + "step": 16790 + }, + { + "epoch": 34.91683991683992, + "grad_norm": 0.0006696488708257675, + "learning_rate": 0.1873719440395012, + "loss": 0.2696, + "num_input_tokens_seen": 12809528, + "step": 16795 + }, + { + "epoch": 34.92723492723493, + "grad_norm": 0.0005263136699795723, + "learning_rate": 0.1873148938112009, + "loss": 0.2601, + "num_input_tokens_seen": 12813272, + "step": 16800 + }, + { + "epoch": 34.92723492723493, + "eval_loss": 0.2487134486436844, + "eval_runtime": 13.3944, + "eval_samples_per_second": 63.907, + "eval_steps_per_second": 15.977, + "num_input_tokens_seen": 12813272, + "step": 16800 + }, + { + "epoch": 34.937629937629936, + "grad_norm": 0.000244904775172472, + "learning_rate": 0.18725783782847508, + "loss": 0.2642, + "num_input_tokens_seen": 12816920, + "step": 16805 + }, + { + "epoch": 34.94802494802495, + "grad_norm": 7.562178507214412e-05, + "learning_rate": 0.1872007761001224, + "loss": 0.248, + "num_input_tokens_seen": 12820664, + "step": 16810 + }, + { + "epoch": 34.95841995841996, + "grad_norm": 0.00025312238722108305, + "learning_rate": 0.1871437086349426, + "loss": 0.2699, + "num_input_tokens_seen": 12824536, + "step": 16815 + }, + { + "epoch": 34.96881496881497, + "grad_norm": 0.00018691220611799508, + "learning_rate": 0.18708663544173615, + "loss": 0.2701, + "num_input_tokens_seen": 12828312, + "step": 16820 + }, + { + "epoch": 34.979209979209976, + "grad_norm": 6.183881487231702e-05, + "learning_rate": 0.18702955652930442, + "loss": 0.2694, + "num_input_tokens_seen": 12832120, + "step": 16825 + }, + { + "epoch": 34.98960498960499, + "grad_norm": 0.00010112955351360142, + "learning_rate": 0.18697247190644972, + "loss": 0.2454, + "num_input_tokens_seen": 12835992, + "step": 16830 + }, + { + "epoch": 35.0, + "grad_norm": 0.0004139370284974575, + "learning_rate": 0.18691538158197527, + "loss": 0.228, + "num_input_tokens_seen": 12839656, + "step": 16835 + }, + { + "epoch": 35.01039501039501, + "grad_norm": 8.809124119579792e-05, + "learning_rate": 0.1868582855646851, + "loss": 0.2882, + "num_input_tokens_seen": 12843528, + "step": 16840 + }, + { + "epoch": 35.020790020790024, + "grad_norm": 0.00020006607519462705, + "learning_rate": 0.18680118386338404, + "loss": 0.2757, + "num_input_tokens_seen": 12847272, + "step": 16845 + }, + { + "epoch": 35.03118503118503, + "grad_norm": 0.0004592253244481981, + "learning_rate": 0.18674407648687794, + "loss": 0.2698, + "num_input_tokens_seen": 12850984, + "step": 16850 + }, + { + "epoch": 35.04158004158004, + "grad_norm": 0.0007527959533035755, + "learning_rate": 0.1866869634439736, + "loss": 0.2361, + "num_input_tokens_seen": 12855048, + "step": 16855 + }, + { + "epoch": 35.05197505197505, + "grad_norm": 0.000490635575260967, + "learning_rate": 0.18662984474347838, + "loss": 0.2342, + "num_input_tokens_seen": 12858824, + "step": 16860 + }, + { + "epoch": 35.062370062370064, + "grad_norm": 7.042768993414938e-05, + "learning_rate": 0.1865727203942008, + "loss": 0.2212, + "num_input_tokens_seen": 12862664, + "step": 16865 + }, + { + "epoch": 35.07276507276507, + "grad_norm": 0.0005292988498695195, + "learning_rate": 0.1865155904049501, + "loss": 0.317, + "num_input_tokens_seen": 12866664, + "step": 16870 + }, + { + "epoch": 35.08316008316008, + "grad_norm": 0.00017502380069345236, + "learning_rate": 0.1864584547845365, + "loss": 0.2595, + "num_input_tokens_seen": 12870600, + "step": 16875 + }, + { + "epoch": 35.093555093555096, + "grad_norm": 0.00010429528629174456, + "learning_rate": 0.186401313541771, + "loss": 0.2744, + "num_input_tokens_seen": 12874440, + "step": 16880 + }, + { + "epoch": 35.103950103950105, + "grad_norm": 0.0002257017040392384, + "learning_rate": 0.18634416668546552, + "loss": 0.2791, + "num_input_tokens_seen": 12878248, + "step": 16885 + }, + { + "epoch": 35.11434511434511, + "grad_norm": 5.904418503632769e-05, + "learning_rate": 0.1862870142244328, + "loss": 0.2747, + "num_input_tokens_seen": 12882120, + "step": 16890 + }, + { + "epoch": 35.12474012474012, + "grad_norm": 0.00010746809130068868, + "learning_rate": 0.1862298561674865, + "loss": 0.2666, + "num_input_tokens_seen": 12885992, + "step": 16895 + }, + { + "epoch": 35.13513513513514, + "grad_norm": 0.000387133244657889, + "learning_rate": 0.18617269252344104, + "loss": 0.2563, + "num_input_tokens_seen": 12889864, + "step": 16900 + }, + { + "epoch": 35.145530145530145, + "grad_norm": 0.00039089430356398225, + "learning_rate": 0.18611552330111186, + "loss": 0.2775, + "num_input_tokens_seen": 12893768, + "step": 16905 + }, + { + "epoch": 35.15592515592515, + "grad_norm": 0.000233535174629651, + "learning_rate": 0.18605834850931507, + "loss": 0.2516, + "num_input_tokens_seen": 12897608, + "step": 16910 + }, + { + "epoch": 35.16632016632017, + "grad_norm": 0.00010665982699720189, + "learning_rate": 0.18600116815686787, + "loss": 0.2452, + "num_input_tokens_seen": 12901608, + "step": 16915 + }, + { + "epoch": 35.17671517671518, + "grad_norm": 0.0002713435678742826, + "learning_rate": 0.1859439822525881, + "loss": 0.2752, + "num_input_tokens_seen": 12905320, + "step": 16920 + }, + { + "epoch": 35.187110187110186, + "grad_norm": 0.0002450247120577842, + "learning_rate": 0.18588679080529455, + "loss": 0.262, + "num_input_tokens_seen": 12909224, + "step": 16925 + }, + { + "epoch": 35.197505197505194, + "grad_norm": 0.00023793573200237006, + "learning_rate": 0.1858295938238069, + "loss": 0.2721, + "num_input_tokens_seen": 12912936, + "step": 16930 + }, + { + "epoch": 35.20790020790021, + "grad_norm": 0.0001158507657237351, + "learning_rate": 0.18577239131694562, + "loss": 0.2855, + "num_input_tokens_seen": 12916648, + "step": 16935 + }, + { + "epoch": 35.21829521829522, + "grad_norm": 0.00017433053290005773, + "learning_rate": 0.18571518329353204, + "loss": 0.2489, + "num_input_tokens_seen": 12920424, + "step": 16940 + }, + { + "epoch": 35.228690228690226, + "grad_norm": 0.0002863065747078508, + "learning_rate": 0.18565796976238838, + "loss": 0.2398, + "num_input_tokens_seen": 12924136, + "step": 16945 + }, + { + "epoch": 35.23908523908524, + "grad_norm": 0.00020643460447899997, + "learning_rate": 0.18560075073233764, + "loss": 0.2636, + "num_input_tokens_seen": 12927912, + "step": 16950 + }, + { + "epoch": 35.24948024948025, + "grad_norm": 0.00025633437326177955, + "learning_rate": 0.18554352621220377, + "loss": 0.2922, + "num_input_tokens_seen": 12931752, + "step": 16955 + }, + { + "epoch": 35.25987525987526, + "grad_norm": 0.00010995643242495134, + "learning_rate": 0.18548629621081153, + "loss": 0.2664, + "num_input_tokens_seen": 12935432, + "step": 16960 + }, + { + "epoch": 35.270270270270274, + "grad_norm": 4.525436452240683e-05, + "learning_rate": 0.18542906073698645, + "loss": 0.2567, + "num_input_tokens_seen": 12939176, + "step": 16965 + }, + { + "epoch": 35.28066528066528, + "grad_norm": 0.00027432964998297393, + "learning_rate": 0.18537181979955494, + "loss": 0.271, + "num_input_tokens_seen": 12943112, + "step": 16970 + }, + { + "epoch": 35.29106029106029, + "grad_norm": 0.00011665987403830513, + "learning_rate": 0.18531457340734434, + "loss": 0.255, + "num_input_tokens_seen": 12946888, + "step": 16975 + }, + { + "epoch": 35.3014553014553, + "grad_norm": 0.0003307465522084385, + "learning_rate": 0.1852573215691827, + "loss": 0.2933, + "num_input_tokens_seen": 12950696, + "step": 16980 + }, + { + "epoch": 35.311850311850314, + "grad_norm": 0.0006522841867990792, + "learning_rate": 0.18520006429389904, + "loss": 0.2381, + "num_input_tokens_seen": 12954408, + "step": 16985 + }, + { + "epoch": 35.32224532224532, + "grad_norm": 0.0007473426521755755, + "learning_rate": 0.1851428015903231, + "loss": 0.278, + "num_input_tokens_seen": 12958216, + "step": 16990 + }, + { + "epoch": 35.33264033264033, + "grad_norm": 0.0001662546128500253, + "learning_rate": 0.1850855334672855, + "loss": 0.2428, + "num_input_tokens_seen": 12962024, + "step": 16995 + }, + { + "epoch": 35.343035343035346, + "grad_norm": 0.0002309860719833523, + "learning_rate": 0.1850282599336178, + "loss": 0.2638, + "num_input_tokens_seen": 12965896, + "step": 17000 + }, + { + "epoch": 35.343035343035346, + "eval_loss": 0.24889925122261047, + "eval_runtime": 13.3956, + "eval_samples_per_second": 63.901, + "eval_steps_per_second": 15.975, + "num_input_tokens_seen": 12965896, + "step": 17000 + }, + { + "epoch": 35.353430353430355, + "grad_norm": 0.0004286384501028806, + "learning_rate": 0.18497098099815215, + "loss": 0.2508, + "num_input_tokens_seen": 12969768, + "step": 17005 + }, + { + "epoch": 35.36382536382536, + "grad_norm": 2.1405317966127768e-05, + "learning_rate": 0.18491369666972174, + "loss": 0.2632, + "num_input_tokens_seen": 12973608, + "step": 17010 + }, + { + "epoch": 35.37422037422037, + "grad_norm": 7.8148914326448e-05, + "learning_rate": 0.1848564069571606, + "loss": 0.2618, + "num_input_tokens_seen": 12977288, + "step": 17015 + }, + { + "epoch": 35.38461538461539, + "grad_norm": 0.00011393125168979168, + "learning_rate": 0.18479911186930348, + "loss": 0.2816, + "num_input_tokens_seen": 12981000, + "step": 17020 + }, + { + "epoch": 35.395010395010395, + "grad_norm": 0.0001927412231452763, + "learning_rate": 0.18474181141498597, + "loss": 0.2891, + "num_input_tokens_seen": 12984744, + "step": 17025 + }, + { + "epoch": 35.4054054054054, + "grad_norm": 0.00037378535489551723, + "learning_rate": 0.18468450560304453, + "loss": 0.2839, + "num_input_tokens_seen": 12988648, + "step": 17030 + }, + { + "epoch": 35.41580041580042, + "grad_norm": 4.6136443415889516e-05, + "learning_rate": 0.1846271944423165, + "loss": 0.2726, + "num_input_tokens_seen": 12992488, + "step": 17035 + }, + { + "epoch": 35.42619542619543, + "grad_norm": 9.900882287183776e-05, + "learning_rate": 0.18456987794163993, + "loss": 0.2775, + "num_input_tokens_seen": 12996296, + "step": 17040 + }, + { + "epoch": 35.436590436590436, + "grad_norm": 0.00014404222019948065, + "learning_rate": 0.18451255610985373, + "loss": 0.2757, + "num_input_tokens_seen": 13000072, + "step": 17045 + }, + { + "epoch": 35.446985446985444, + "grad_norm": 0.00025201408425346017, + "learning_rate": 0.18445522895579766, + "loss": 0.267, + "num_input_tokens_seen": 13003976, + "step": 17050 + }, + { + "epoch": 35.45738045738046, + "grad_norm": 0.0001261031866306439, + "learning_rate": 0.1843978964883123, + "loss": 0.2571, + "num_input_tokens_seen": 13007848, + "step": 17055 + }, + { + "epoch": 35.46777546777547, + "grad_norm": 0.00010357849532738328, + "learning_rate": 0.18434055871623906, + "loss": 0.2723, + "num_input_tokens_seen": 13011624, + "step": 17060 + }, + { + "epoch": 35.478170478170476, + "grad_norm": 0.00022431333491113037, + "learning_rate": 0.18428321564842007, + "loss": 0.2557, + "num_input_tokens_seen": 13015432, + "step": 17065 + }, + { + "epoch": 35.48856548856549, + "grad_norm": 0.0003426216426305473, + "learning_rate": 0.18422586729369841, + "loss": 0.2555, + "num_input_tokens_seen": 13019144, + "step": 17070 + }, + { + "epoch": 35.4989604989605, + "grad_norm": 0.0005259495228528976, + "learning_rate": 0.1841685136609179, + "loss": 0.2831, + "num_input_tokens_seen": 13023176, + "step": 17075 + }, + { + "epoch": 35.50935550935551, + "grad_norm": 0.00013832398690283298, + "learning_rate": 0.18411115475892326, + "loss": 0.2593, + "num_input_tokens_seen": 13026952, + "step": 17080 + }, + { + "epoch": 35.51975051975052, + "grad_norm": 2.0800953279831447e-05, + "learning_rate": 0.18405379059655982, + "loss": 0.2628, + "num_input_tokens_seen": 13030856, + "step": 17085 + }, + { + "epoch": 35.53014553014553, + "grad_norm": 8.958084072219208e-05, + "learning_rate": 0.1839964211826739, + "loss": 0.2829, + "num_input_tokens_seen": 13034792, + "step": 17090 + }, + { + "epoch": 35.54054054054054, + "grad_norm": 0.0004163573030382395, + "learning_rate": 0.18393904652611265, + "loss": 0.2686, + "num_input_tokens_seen": 13038600, + "step": 17095 + }, + { + "epoch": 35.55093555093555, + "grad_norm": 0.00042450649198144674, + "learning_rate": 0.18388166663572392, + "loss": 0.264, + "num_input_tokens_seen": 13042504, + "step": 17100 + }, + { + "epoch": 35.561330561330564, + "grad_norm": 0.0001936911721713841, + "learning_rate": 0.18382428152035643, + "loss": 0.2406, + "num_input_tokens_seen": 13046376, + "step": 17105 + }, + { + "epoch": 35.57172557172557, + "grad_norm": 9.386425517732278e-05, + "learning_rate": 0.1837668911888596, + "loss": 0.2355, + "num_input_tokens_seen": 13050088, + "step": 17110 + }, + { + "epoch": 35.58212058212058, + "grad_norm": 0.00011518143583089113, + "learning_rate": 0.18370949565008388, + "loss": 0.2776, + "num_input_tokens_seen": 13053896, + "step": 17115 + }, + { + "epoch": 35.59251559251559, + "grad_norm": 6.083751213736832e-05, + "learning_rate": 0.1836520949128803, + "loss": 0.2569, + "num_input_tokens_seen": 13057736, + "step": 17120 + }, + { + "epoch": 35.602910602910605, + "grad_norm": 0.00035684771137312055, + "learning_rate": 0.18359468898610076, + "loss": 0.2829, + "num_input_tokens_seen": 13061672, + "step": 17125 + }, + { + "epoch": 35.61330561330561, + "grad_norm": 0.0005231992108747363, + "learning_rate": 0.18353727787859797, + "loss": 0.2725, + "num_input_tokens_seen": 13065448, + "step": 17130 + }, + { + "epoch": 35.62370062370062, + "grad_norm": 0.00024077462148852646, + "learning_rate": 0.18347986159922552, + "loss": 0.2711, + "num_input_tokens_seen": 13069224, + "step": 17135 + }, + { + "epoch": 35.63409563409564, + "grad_norm": 4.889236151939258e-05, + "learning_rate": 0.1834224401568377, + "loss": 0.2862, + "num_input_tokens_seen": 13073000, + "step": 17140 + }, + { + "epoch": 35.644490644490645, + "grad_norm": 0.00034011725801974535, + "learning_rate": 0.1833650135602896, + "loss": 0.2667, + "num_input_tokens_seen": 13076776, + "step": 17145 + }, + { + "epoch": 35.65488565488565, + "grad_norm": 0.00021094617841299623, + "learning_rate": 0.18330758181843707, + "loss": 0.2771, + "num_input_tokens_seen": 13080680, + "step": 17150 + }, + { + "epoch": 35.66528066528066, + "grad_norm": 0.0004294009995646775, + "learning_rate": 0.18325014494013686, + "loss": 0.2893, + "num_input_tokens_seen": 13084552, + "step": 17155 + }, + { + "epoch": 35.67567567567568, + "grad_norm": 0.000628128822427243, + "learning_rate": 0.18319270293424647, + "loss": 0.2836, + "num_input_tokens_seen": 13088360, + "step": 17160 + }, + { + "epoch": 35.686070686070686, + "grad_norm": 0.000327672460116446, + "learning_rate": 0.18313525580962417, + "loss": 0.271, + "num_input_tokens_seen": 13092104, + "step": 17165 + }, + { + "epoch": 35.696465696465694, + "grad_norm": 0.0004409045504871756, + "learning_rate": 0.18307780357512896, + "loss": 0.2593, + "num_input_tokens_seen": 13095848, + "step": 17170 + }, + { + "epoch": 35.70686070686071, + "grad_norm": 0.0002291866549057886, + "learning_rate": 0.1830203462396208, + "loss": 0.3165, + "num_input_tokens_seen": 13099688, + "step": 17175 + }, + { + "epoch": 35.71725571725572, + "grad_norm": 0.00036499035195447505, + "learning_rate": 0.18296288381196033, + "loss": 0.2498, + "num_input_tokens_seen": 13103656, + "step": 17180 + }, + { + "epoch": 35.727650727650726, + "grad_norm": 0.0004183951241429895, + "learning_rate": 0.1829054163010089, + "loss": 0.2848, + "num_input_tokens_seen": 13107528, + "step": 17185 + }, + { + "epoch": 35.73804573804574, + "grad_norm": 8.121300197672099e-05, + "learning_rate": 0.18284794371562874, + "loss": 0.2782, + "num_input_tokens_seen": 13111272, + "step": 17190 + }, + { + "epoch": 35.74844074844075, + "grad_norm": 0.00026350145344622433, + "learning_rate": 0.18279046606468288, + "loss": 0.2737, + "num_input_tokens_seen": 13115080, + "step": 17195 + }, + { + "epoch": 35.75883575883576, + "grad_norm": 0.00017929417663253844, + "learning_rate": 0.1827329833570351, + "loss": 0.2766, + "num_input_tokens_seen": 13118824, + "step": 17200 + }, + { + "epoch": 35.75883575883576, + "eval_loss": 0.25238293409347534, + "eval_runtime": 13.4257, + "eval_samples_per_second": 63.758, + "eval_steps_per_second": 15.94, + "num_input_tokens_seen": 13118824, + "step": 17200 + }, + { + "epoch": 35.76923076923077, + "grad_norm": 0.00010151242167921737, + "learning_rate": 0.18267549560154991, + "loss": 0.2499, + "num_input_tokens_seen": 13122632, + "step": 17205 + }, + { + "epoch": 35.77962577962578, + "grad_norm": 0.00030825851717963815, + "learning_rate": 0.18261800280709267, + "loss": 0.2719, + "num_input_tokens_seen": 13126536, + "step": 17210 + }, + { + "epoch": 35.79002079002079, + "grad_norm": 0.00035874691093340516, + "learning_rate": 0.18256050498252957, + "loss": 0.2659, + "num_input_tokens_seen": 13130376, + "step": 17215 + }, + { + "epoch": 35.8004158004158, + "grad_norm": 0.00014793197624385357, + "learning_rate": 0.18250300213672735, + "loss": 0.2887, + "num_input_tokens_seen": 13134088, + "step": 17220 + }, + { + "epoch": 35.810810810810814, + "grad_norm": 0.0006373718497343361, + "learning_rate": 0.18244549427855378, + "loss": 0.269, + "num_input_tokens_seen": 13137832, + "step": 17225 + }, + { + "epoch": 35.82120582120582, + "grad_norm": 0.0006245653494261205, + "learning_rate": 0.1823879814168772, + "loss": 0.2459, + "num_input_tokens_seen": 13141576, + "step": 17230 + }, + { + "epoch": 35.83160083160083, + "grad_norm": 0.00022500309569295496, + "learning_rate": 0.18233046356056692, + "loss": 0.2446, + "num_input_tokens_seen": 13145480, + "step": 17235 + }, + { + "epoch": 35.84199584199584, + "grad_norm": 0.00011551700299605727, + "learning_rate": 0.18227294071849284, + "loss": 0.3236, + "num_input_tokens_seen": 13149384, + "step": 17240 + }, + { + "epoch": 35.852390852390855, + "grad_norm": 0.0004671224160119891, + "learning_rate": 0.18221541289952578, + "loss": 0.2341, + "num_input_tokens_seen": 13153320, + "step": 17245 + }, + { + "epoch": 35.86278586278586, + "grad_norm": 0.0003103829803876579, + "learning_rate": 0.18215788011253717, + "loss": 0.2534, + "num_input_tokens_seen": 13157032, + "step": 17250 + }, + { + "epoch": 35.87318087318087, + "grad_norm": 0.000261551613220945, + "learning_rate": 0.18210034236639935, + "loss": 0.2583, + "num_input_tokens_seen": 13160968, + "step": 17255 + }, + { + "epoch": 35.88357588357589, + "grad_norm": 0.0004658151010517031, + "learning_rate": 0.1820427996699853, + "loss": 0.2515, + "num_input_tokens_seen": 13164776, + "step": 17260 + }, + { + "epoch": 35.893970893970895, + "grad_norm": 0.00015330652240663767, + "learning_rate": 0.1819852520321689, + "loss": 0.2559, + "num_input_tokens_seen": 13168456, + "step": 17265 + }, + { + "epoch": 35.9043659043659, + "grad_norm": 0.00028016703436151147, + "learning_rate": 0.18192769946182466, + "loss": 0.2674, + "num_input_tokens_seen": 13172168, + "step": 17270 + }, + { + "epoch": 35.91476091476091, + "grad_norm": 0.0004740901349578053, + "learning_rate": 0.18187014196782794, + "loss": 0.2638, + "num_input_tokens_seen": 13176040, + "step": 17275 + }, + { + "epoch": 35.92515592515593, + "grad_norm": 0.00027310673613101244, + "learning_rate": 0.18181257955905486, + "loss": 0.2402, + "num_input_tokens_seen": 13179976, + "step": 17280 + }, + { + "epoch": 35.935550935550935, + "grad_norm": 0.0001899885101011023, + "learning_rate": 0.18175501224438217, + "loss": 0.2603, + "num_input_tokens_seen": 13183720, + "step": 17285 + }, + { + "epoch": 35.945945945945944, + "grad_norm": 0.00026490914751775563, + "learning_rate": 0.18169744003268756, + "loss": 0.2969, + "num_input_tokens_seen": 13187496, + "step": 17290 + }, + { + "epoch": 35.95634095634096, + "grad_norm": 0.00013785452756565064, + "learning_rate": 0.18163986293284937, + "loss": 0.2707, + "num_input_tokens_seen": 13191336, + "step": 17295 + }, + { + "epoch": 35.96673596673597, + "grad_norm": 0.00028141759685240686, + "learning_rate": 0.18158228095374673, + "loss": 0.2485, + "num_input_tokens_seen": 13195240, + "step": 17300 + }, + { + "epoch": 35.977130977130976, + "grad_norm": 0.00023315436556003988, + "learning_rate": 0.18152469410425945, + "loss": 0.2933, + "num_input_tokens_seen": 13198952, + "step": 17305 + }, + { + "epoch": 35.987525987525984, + "grad_norm": 0.00018487342458683997, + "learning_rate": 0.18146710239326813, + "loss": 0.2949, + "num_input_tokens_seen": 13202728, + "step": 17310 + }, + { + "epoch": 35.997920997921, + "grad_norm": 0.0006413618102669716, + "learning_rate": 0.18140950582965423, + "loss": 0.2792, + "num_input_tokens_seen": 13206440, + "step": 17315 + }, + { + "epoch": 36.00831600831601, + "grad_norm": 9.09929076442495e-05, + "learning_rate": 0.1813519044222998, + "loss": 0.2695, + "num_input_tokens_seen": 13210144, + "step": 17320 + }, + { + "epoch": 36.018711018711016, + "grad_norm": 0.0003249425790272653, + "learning_rate": 0.18129429818008772, + "loss": 0.2612, + "num_input_tokens_seen": 13213984, + "step": 17325 + }, + { + "epoch": 36.02910602910603, + "grad_norm": 0.00035491675953380764, + "learning_rate": 0.18123668711190163, + "loss": 0.2658, + "num_input_tokens_seen": 13217856, + "step": 17330 + }, + { + "epoch": 36.03950103950104, + "grad_norm": 0.00041443223017267883, + "learning_rate": 0.18117907122662583, + "loss": 0.2774, + "num_input_tokens_seen": 13221504, + "step": 17335 + }, + { + "epoch": 36.04989604989605, + "grad_norm": 0.000182196352398023, + "learning_rate": 0.1811214505331454, + "loss": 0.2567, + "num_input_tokens_seen": 13225312, + "step": 17340 + }, + { + "epoch": 36.06029106029106, + "grad_norm": 0.00018606791854836047, + "learning_rate": 0.1810638250403462, + "loss": 0.2725, + "num_input_tokens_seen": 13229248, + "step": 17345 + }, + { + "epoch": 36.07068607068607, + "grad_norm": 0.000773865613155067, + "learning_rate": 0.1810061947571148, + "loss": 0.2267, + "num_input_tokens_seen": 13233024, + "step": 17350 + }, + { + "epoch": 36.08108108108108, + "grad_norm": 0.00011020462261512876, + "learning_rate": 0.1809485596923385, + "loss": 0.2727, + "num_input_tokens_seen": 13237056, + "step": 17355 + }, + { + "epoch": 36.09147609147609, + "grad_norm": 0.000755861634388566, + "learning_rate": 0.18089091985490546, + "loss": 0.2635, + "num_input_tokens_seen": 13240832, + "step": 17360 + }, + { + "epoch": 36.101871101871104, + "grad_norm": 0.00044536785571835935, + "learning_rate": 0.18083327525370432, + "loss": 0.251, + "num_input_tokens_seen": 13244768, + "step": 17365 + }, + { + "epoch": 36.11226611226611, + "grad_norm": 0.00032161345006898046, + "learning_rate": 0.18077562589762464, + "loss": 0.2612, + "num_input_tokens_seen": 13248704, + "step": 17370 + }, + { + "epoch": 36.12266112266112, + "grad_norm": 7.415869185933843e-05, + "learning_rate": 0.1807179717955567, + "loss": 0.2774, + "num_input_tokens_seen": 13252672, + "step": 17375 + }, + { + "epoch": 36.13305613305613, + "grad_norm": 7.982958777574822e-05, + "learning_rate": 0.1806603129563915, + "loss": 0.2487, + "num_input_tokens_seen": 13256480, + "step": 17380 + }, + { + "epoch": 36.143451143451145, + "grad_norm": 0.0001779584272298962, + "learning_rate": 0.1806026493890208, + "loss": 0.2872, + "num_input_tokens_seen": 13260352, + "step": 17385 + }, + { + "epoch": 36.15384615384615, + "grad_norm": 9.448125638300553e-05, + "learning_rate": 0.18054498110233688, + "loss": 0.251, + "num_input_tokens_seen": 13264128, + "step": 17390 + }, + { + "epoch": 36.16424116424116, + "grad_norm": 0.00026116447406820953, + "learning_rate": 0.1804873081052331, + "loss": 0.2471, + "num_input_tokens_seen": 13268032, + "step": 17395 + }, + { + "epoch": 36.17463617463618, + "grad_norm": 0.00038582689012400806, + "learning_rate": 0.18042963040660326, + "loss": 0.2599, + "num_input_tokens_seen": 13271872, + "step": 17400 + }, + { + "epoch": 36.17463617463618, + "eval_loss": 0.2490328550338745, + "eval_runtime": 13.4571, + "eval_samples_per_second": 63.61, + "eval_steps_per_second": 15.902, + "num_input_tokens_seen": 13271872, + "step": 17400 + }, + { + "epoch": 36.185031185031185, + "grad_norm": 5.038334347773343e-05, + "learning_rate": 0.180371948015342, + "loss": 0.2377, + "num_input_tokens_seen": 13275808, + "step": 17405 + }, + { + "epoch": 36.195426195426194, + "grad_norm": 0.0003406234900467098, + "learning_rate": 0.18031426094034472, + "loss": 0.2603, + "num_input_tokens_seen": 13279584, + "step": 17410 + }, + { + "epoch": 36.20582120582121, + "grad_norm": 0.0002932468487415463, + "learning_rate": 0.18025656919050737, + "loss": 0.2597, + "num_input_tokens_seen": 13283392, + "step": 17415 + }, + { + "epoch": 36.21621621621622, + "grad_norm": 0.0002370987058384344, + "learning_rate": 0.18019887277472688, + "loss": 0.2543, + "num_input_tokens_seen": 13287040, + "step": 17420 + }, + { + "epoch": 36.226611226611226, + "grad_norm": 0.0003863436577375978, + "learning_rate": 0.18014117170190067, + "loss": 0.2766, + "num_input_tokens_seen": 13290976, + "step": 17425 + }, + { + "epoch": 36.237006237006234, + "grad_norm": 0.0001916103792609647, + "learning_rate": 0.18008346598092703, + "loss": 0.2679, + "num_input_tokens_seen": 13294944, + "step": 17430 + }, + { + "epoch": 36.24740124740125, + "grad_norm": 0.00019920420891139656, + "learning_rate": 0.18002575562070489, + "loss": 0.2588, + "num_input_tokens_seen": 13298720, + "step": 17435 + }, + { + "epoch": 36.25779625779626, + "grad_norm": 0.00015447613259311765, + "learning_rate": 0.1799680406301339, + "loss": 0.2726, + "num_input_tokens_seen": 13302560, + "step": 17440 + }, + { + "epoch": 36.268191268191266, + "grad_norm": 0.0005707088857889175, + "learning_rate": 0.17991032101811447, + "loss": 0.2667, + "num_input_tokens_seen": 13306336, + "step": 17445 + }, + { + "epoch": 36.27858627858628, + "grad_norm": 0.00012345347204245627, + "learning_rate": 0.1798525967935476, + "loss": 0.2776, + "num_input_tokens_seen": 13310080, + "step": 17450 + }, + { + "epoch": 36.28898128898129, + "grad_norm": 3.670494697871618e-05, + "learning_rate": 0.17979486796533517, + "loss": 0.2688, + "num_input_tokens_seen": 13313856, + "step": 17455 + }, + { + "epoch": 36.2993762993763, + "grad_norm": 0.00022341989097185433, + "learning_rate": 0.1797371345423797, + "loss": 0.261, + "num_input_tokens_seen": 13317728, + "step": 17460 + }, + { + "epoch": 36.30977130977131, + "grad_norm": 0.0005145045579411089, + "learning_rate": 0.17967939653358436, + "loss": 0.2612, + "num_input_tokens_seen": 13321536, + "step": 17465 + }, + { + "epoch": 36.32016632016632, + "grad_norm": 7.41965341148898e-05, + "learning_rate": 0.17962165394785315, + "loss": 0.2747, + "num_input_tokens_seen": 13325216, + "step": 17470 + }, + { + "epoch": 36.33056133056133, + "grad_norm": 0.0006184731610119343, + "learning_rate": 0.17956390679409057, + "loss": 0.2728, + "num_input_tokens_seen": 13328960, + "step": 17475 + }, + { + "epoch": 36.34095634095634, + "grad_norm": 0.00042669460526667535, + "learning_rate": 0.1795061550812021, + "loss": 0.2803, + "num_input_tokens_seen": 13332768, + "step": 17480 + }, + { + "epoch": 36.351351351351354, + "grad_norm": 0.00027450561174191535, + "learning_rate": 0.1794483988180937, + "loss": 0.2643, + "num_input_tokens_seen": 13336704, + "step": 17485 + }, + { + "epoch": 36.36174636174636, + "grad_norm": 0.00046788016334176064, + "learning_rate": 0.17939063801367214, + "loss": 0.2774, + "num_input_tokens_seen": 13340736, + "step": 17490 + }, + { + "epoch": 36.37214137214137, + "grad_norm": 0.0002466692531015724, + "learning_rate": 0.17933287267684483, + "loss": 0.2701, + "num_input_tokens_seen": 13344448, + "step": 17495 + }, + { + "epoch": 36.38253638253638, + "grad_norm": 0.00035465057590045035, + "learning_rate": 0.17927510281651995, + "loss": 0.2726, + "num_input_tokens_seen": 13348384, + "step": 17500 + }, + { + "epoch": 36.392931392931395, + "grad_norm": 0.0003793487267103046, + "learning_rate": 0.17921732844160634, + "loss": 0.2792, + "num_input_tokens_seen": 13352224, + "step": 17505 + }, + { + "epoch": 36.4033264033264, + "grad_norm": 9.518696606392041e-05, + "learning_rate": 0.17915954956101351, + "loss": 0.27, + "num_input_tokens_seen": 13355936, + "step": 17510 + }, + { + "epoch": 36.41372141372141, + "grad_norm": 0.0002673929266165942, + "learning_rate": 0.17910176618365165, + "loss": 0.2563, + "num_input_tokens_seen": 13359936, + "step": 17515 + }, + { + "epoch": 36.42411642411643, + "grad_norm": 0.0003463521134108305, + "learning_rate": 0.17904397831843177, + "loss": 0.2748, + "num_input_tokens_seen": 13363712, + "step": 17520 + }, + { + "epoch": 36.434511434511435, + "grad_norm": 0.00024095296976156533, + "learning_rate": 0.17898618597426547, + "loss": 0.2521, + "num_input_tokens_seen": 13367488, + "step": 17525 + }, + { + "epoch": 36.444906444906444, + "grad_norm": 8.075922232819721e-05, + "learning_rate": 0.17892838916006495, + "loss": 0.2503, + "num_input_tokens_seen": 13371264, + "step": 17530 + }, + { + "epoch": 36.45530145530145, + "grad_norm": 6.309789750957862e-05, + "learning_rate": 0.17887058788474333, + "loss": 0.2417, + "num_input_tokens_seen": 13374944, + "step": 17535 + }, + { + "epoch": 36.46569646569647, + "grad_norm": 0.000544725451618433, + "learning_rate": 0.17881278215721427, + "loss": 0.2756, + "num_input_tokens_seen": 13378752, + "step": 17540 + }, + { + "epoch": 36.476091476091476, + "grad_norm": 0.00040329451439902186, + "learning_rate": 0.1787549719863921, + "loss": 0.2791, + "num_input_tokens_seen": 13382496, + "step": 17545 + }, + { + "epoch": 36.486486486486484, + "grad_norm": 0.0004914184100925922, + "learning_rate": 0.17869715738119188, + "loss": 0.2762, + "num_input_tokens_seen": 13386304, + "step": 17550 + }, + { + "epoch": 36.4968814968815, + "grad_norm": 9.815453813644126e-05, + "learning_rate": 0.17863933835052936, + "loss": 0.257, + "num_input_tokens_seen": 13390176, + "step": 17555 + }, + { + "epoch": 36.50727650727651, + "grad_norm": 0.0001887131220428273, + "learning_rate": 0.17858151490332097, + "loss": 0.2882, + "num_input_tokens_seen": 13394048, + "step": 17560 + }, + { + "epoch": 36.517671517671516, + "grad_norm": 6.773704080842435e-05, + "learning_rate": 0.17852368704848381, + "loss": 0.2767, + "num_input_tokens_seen": 13397888, + "step": 17565 + }, + { + "epoch": 36.528066528066525, + "grad_norm": 0.0003230652364436537, + "learning_rate": 0.17846585479493565, + "loss": 0.2654, + "num_input_tokens_seen": 13401664, + "step": 17570 + }, + { + "epoch": 36.53846153846154, + "grad_norm": 0.00027159383171238005, + "learning_rate": 0.178408018151595, + "loss": 0.2478, + "num_input_tokens_seen": 13405472, + "step": 17575 + }, + { + "epoch": 36.54885654885655, + "grad_norm": 0.00012372738274279982, + "learning_rate": 0.17835017712738085, + "loss": 0.2714, + "num_input_tokens_seen": 13409152, + "step": 17580 + }, + { + "epoch": 36.55925155925156, + "grad_norm": 0.00015436962712556124, + "learning_rate": 0.17829233173121323, + "loss": 0.2584, + "num_input_tokens_seen": 13412960, + "step": 17585 + }, + { + "epoch": 36.56964656964657, + "grad_norm": 0.0001529431319795549, + "learning_rate": 0.17823448197201244, + "loss": 0.2454, + "num_input_tokens_seen": 13416672, + "step": 17590 + }, + { + "epoch": 36.58004158004158, + "grad_norm": 0.00039333649328909814, + "learning_rate": 0.1781766278586997, + "loss": 0.285, + "num_input_tokens_seen": 13420384, + "step": 17595 + }, + { + "epoch": 36.59043659043659, + "grad_norm": 0.00039363655378110707, + "learning_rate": 0.1781187694001969, + "loss": 0.2689, + "num_input_tokens_seen": 13424128, + "step": 17600 + }, + { + "epoch": 36.59043659043659, + "eval_loss": 0.2603258788585663, + "eval_runtime": 13.4013, + "eval_samples_per_second": 63.874, + "eval_steps_per_second": 15.969, + "num_input_tokens_seen": 13424128, + "step": 17600 + }, + { + "epoch": 36.6008316008316, + "grad_norm": 0.00012132203846704215, + "learning_rate": 0.1780609066054265, + "loss": 0.2846, + "num_input_tokens_seen": 13428064, + "step": 17605 + }, + { + "epoch": 36.61122661122661, + "grad_norm": 0.0001391533005516976, + "learning_rate": 0.17800303948331164, + "loss": 0.2743, + "num_input_tokens_seen": 13431968, + "step": 17610 + }, + { + "epoch": 36.62162162162162, + "grad_norm": 6.1971259128768e-05, + "learning_rate": 0.1779451680427762, + "loss": 0.2542, + "num_input_tokens_seen": 13435712, + "step": 17615 + }, + { + "epoch": 36.63201663201663, + "grad_norm": 0.0002711884444579482, + "learning_rate": 0.17788729229274464, + "loss": 0.2405, + "num_input_tokens_seen": 13439456, + "step": 17620 + }, + { + "epoch": 36.642411642411645, + "grad_norm": 9.95310110738501e-05, + "learning_rate": 0.17782941224214222, + "loss": 0.3031, + "num_input_tokens_seen": 13443232, + "step": 17625 + }, + { + "epoch": 36.65280665280665, + "grad_norm": 0.00027481201686896384, + "learning_rate": 0.17777152789989464, + "loss": 0.247, + "num_input_tokens_seen": 13447008, + "step": 17630 + }, + { + "epoch": 36.66320166320166, + "grad_norm": 0.0005391245940700173, + "learning_rate": 0.17771363927492845, + "loss": 0.2543, + "num_input_tokens_seen": 13450720, + "step": 17635 + }, + { + "epoch": 36.67359667359668, + "grad_norm": 0.0005165913025848567, + "learning_rate": 0.17765574637617085, + "loss": 0.2831, + "num_input_tokens_seen": 13454464, + "step": 17640 + }, + { + "epoch": 36.683991683991685, + "grad_norm": 0.0005055902292951941, + "learning_rate": 0.17759784921254962, + "loss": 0.259, + "num_input_tokens_seen": 13458272, + "step": 17645 + }, + { + "epoch": 36.694386694386694, + "grad_norm": 0.0004741275042761117, + "learning_rate": 0.1775399477929932, + "loss": 0.2678, + "num_input_tokens_seen": 13462016, + "step": 17650 + }, + { + "epoch": 36.7047817047817, + "grad_norm": 0.0002835779741872102, + "learning_rate": 0.17748204212643076, + "loss": 0.2712, + "num_input_tokens_seen": 13465728, + "step": 17655 + }, + { + "epoch": 36.71517671517672, + "grad_norm": 0.00030802603578194976, + "learning_rate": 0.17742413222179204, + "loss": 0.2549, + "num_input_tokens_seen": 13469664, + "step": 17660 + }, + { + "epoch": 36.725571725571726, + "grad_norm": 0.00016740414139349014, + "learning_rate": 0.17736621808800754, + "loss": 0.2492, + "num_input_tokens_seen": 13473376, + "step": 17665 + }, + { + "epoch": 36.735966735966734, + "grad_norm": 0.0004384123894851655, + "learning_rate": 0.17730829973400827, + "loss": 0.2867, + "num_input_tokens_seen": 13477248, + "step": 17670 + }, + { + "epoch": 36.74636174636175, + "grad_norm": 0.00034706469159573317, + "learning_rate": 0.17725037716872602, + "loss": 0.2554, + "num_input_tokens_seen": 13481216, + "step": 17675 + }, + { + "epoch": 36.75675675675676, + "grad_norm": 0.00011181944137206301, + "learning_rate": 0.17719245040109313, + "loss": 0.2712, + "num_input_tokens_seen": 13485024, + "step": 17680 + }, + { + "epoch": 36.767151767151766, + "grad_norm": 0.00016816981951706111, + "learning_rate": 0.17713451944004271, + "loss": 0.2437, + "num_input_tokens_seen": 13488800, + "step": 17685 + }, + { + "epoch": 36.777546777546775, + "grad_norm": 0.00011353805166436359, + "learning_rate": 0.17707658429450843, + "loss": 0.2882, + "num_input_tokens_seen": 13492480, + "step": 17690 + }, + { + "epoch": 36.78794178794179, + "grad_norm": 0.00016840647731442004, + "learning_rate": 0.1770186449734245, + "loss": 0.2569, + "num_input_tokens_seen": 13496288, + "step": 17695 + }, + { + "epoch": 36.7983367983368, + "grad_norm": 5.462926856125705e-05, + "learning_rate": 0.17696070148572599, + "loss": 0.2657, + "num_input_tokens_seen": 13500224, + "step": 17700 + }, + { + "epoch": 36.80873180873181, + "grad_norm": 0.0004431198467500508, + "learning_rate": 0.17690275384034856, + "loss": 0.2799, + "num_input_tokens_seen": 13504096, + "step": 17705 + }, + { + "epoch": 36.81912681912682, + "grad_norm": 0.00029139991966076195, + "learning_rate": 0.17684480204622835, + "loss": 0.2669, + "num_input_tokens_seen": 13507968, + "step": 17710 + }, + { + "epoch": 36.82952182952183, + "grad_norm": 0.0007000116747803986, + "learning_rate": 0.1767868461123023, + "loss": 0.2967, + "num_input_tokens_seen": 13511968, + "step": 17715 + }, + { + "epoch": 36.83991683991684, + "grad_norm": 9.667058475315571e-05, + "learning_rate": 0.176728886047508, + "loss": 0.277, + "num_input_tokens_seen": 13515808, + "step": 17720 + }, + { + "epoch": 36.85031185031185, + "grad_norm": 7.989753066794947e-05, + "learning_rate": 0.17667092186078362, + "loss": 0.2559, + "num_input_tokens_seen": 13519488, + "step": 17725 + }, + { + "epoch": 36.86070686070686, + "grad_norm": 0.0005120862042531371, + "learning_rate": 0.17661295356106785, + "loss": 0.2703, + "num_input_tokens_seen": 13523552, + "step": 17730 + }, + { + "epoch": 36.87110187110187, + "grad_norm": 0.00016344754840247333, + "learning_rate": 0.1765549811573002, + "loss": 0.2892, + "num_input_tokens_seen": 13527360, + "step": 17735 + }, + { + "epoch": 36.88149688149688, + "grad_norm": 0.00044120170059613883, + "learning_rate": 0.17649700465842078, + "loss": 0.2779, + "num_input_tokens_seen": 13531072, + "step": 17740 + }, + { + "epoch": 36.891891891891895, + "grad_norm": 8.828326099319384e-05, + "learning_rate": 0.17643902407337023, + "loss": 0.267, + "num_input_tokens_seen": 13534944, + "step": 17745 + }, + { + "epoch": 36.9022869022869, + "grad_norm": 0.00010145201667910442, + "learning_rate": 0.17638103941108993, + "loss": 0.27, + "num_input_tokens_seen": 13538720, + "step": 17750 + }, + { + "epoch": 36.91268191268191, + "grad_norm": 0.00012181472993688658, + "learning_rate": 0.1763230506805218, + "loss": 0.2643, + "num_input_tokens_seen": 13542560, + "step": 17755 + }, + { + "epoch": 36.92307692307692, + "grad_norm": 0.00019509706180542707, + "learning_rate": 0.1762650578906085, + "loss": 0.2939, + "num_input_tokens_seen": 13546400, + "step": 17760 + }, + { + "epoch": 36.933471933471935, + "grad_norm": 0.0002702220226638019, + "learning_rate": 0.1762070610502932, + "loss": 0.2613, + "num_input_tokens_seen": 13550176, + "step": 17765 + }, + { + "epoch": 36.943866943866944, + "grad_norm": 0.00010535618639551103, + "learning_rate": 0.17614906016851975, + "loss": 0.2758, + "num_input_tokens_seen": 13553856, + "step": 17770 + }, + { + "epoch": 36.95426195426195, + "grad_norm": 0.0003402753791306168, + "learning_rate": 0.17609105525423258, + "loss": 0.2585, + "num_input_tokens_seen": 13557632, + "step": 17775 + }, + { + "epoch": 36.96465696465697, + "grad_norm": 0.0002104471204802394, + "learning_rate": 0.1760330463163768, + "loss": 0.26, + "num_input_tokens_seen": 13561216, + "step": 17780 + }, + { + "epoch": 36.975051975051976, + "grad_norm": 6.175143789732829e-05, + "learning_rate": 0.17597503336389816, + "loss": 0.2654, + "num_input_tokens_seen": 13564832, + "step": 17785 + }, + { + "epoch": 36.985446985446984, + "grad_norm": 0.0004685286548919976, + "learning_rate": 0.17591701640574298, + "loss": 0.2647, + "num_input_tokens_seen": 13568640, + "step": 17790 + }, + { + "epoch": 36.99584199584199, + "grad_norm": 0.0005425813724286854, + "learning_rate": 0.17585899545085815, + "loss": 0.2816, + "num_input_tokens_seen": 13572448, + "step": 17795 + }, + { + "epoch": 37.00623700623701, + "grad_norm": 0.0003534541465342045, + "learning_rate": 0.17580097050819124, + "loss": 0.2689, + "num_input_tokens_seen": 13576056, + "step": 17800 + }, + { + "epoch": 37.00623700623701, + "eval_loss": 0.26566049456596375, + "eval_runtime": 13.4131, + "eval_samples_per_second": 63.818, + "eval_steps_per_second": 15.955, + "num_input_tokens_seen": 13576056, + "step": 17800 + }, + { + "epoch": 37.016632016632016, + "grad_norm": 0.0001308080245507881, + "learning_rate": 0.17574294158669046, + "loss": 0.261, + "num_input_tokens_seen": 13579832, + "step": 17805 + }, + { + "epoch": 37.027027027027025, + "grad_norm": 0.0005428889999166131, + "learning_rate": 0.17568490869530456, + "loss": 0.2906, + "num_input_tokens_seen": 13583608, + "step": 17810 + }, + { + "epoch": 37.03742203742204, + "grad_norm": 0.0002659168385434896, + "learning_rate": 0.17562687184298295, + "loss": 0.2611, + "num_input_tokens_seen": 13587448, + "step": 17815 + }, + { + "epoch": 37.04781704781705, + "grad_norm": 0.0005051145562902093, + "learning_rate": 0.1755688310386757, + "loss": 0.2487, + "num_input_tokens_seen": 13591288, + "step": 17820 + }, + { + "epoch": 37.05821205821206, + "grad_norm": 0.0001192485069623217, + "learning_rate": 0.17551078629133335, + "loss": 0.2655, + "num_input_tokens_seen": 13595032, + "step": 17825 + }, + { + "epoch": 37.06860706860707, + "grad_norm": 3.7393612728919834e-05, + "learning_rate": 0.17545273760990718, + "loss": 0.2545, + "num_input_tokens_seen": 13598712, + "step": 17830 + }, + { + "epoch": 37.07900207900208, + "grad_norm": 0.00014857096539344639, + "learning_rate": 0.17539468500334904, + "loss": 0.2684, + "num_input_tokens_seen": 13602552, + "step": 17835 + }, + { + "epoch": 37.08939708939709, + "grad_norm": 0.0002657573204487562, + "learning_rate": 0.17533662848061132, + "loss": 0.2356, + "num_input_tokens_seen": 13606264, + "step": 17840 + }, + { + "epoch": 37.0997920997921, + "grad_norm": 0.0003399352717678994, + "learning_rate": 0.1752785680506471, + "loss": 0.2545, + "num_input_tokens_seen": 13610040, + "step": 17845 + }, + { + "epoch": 37.11018711018711, + "grad_norm": 0.00019116760813631117, + "learning_rate": 0.17522050372241, + "loss": 0.224, + "num_input_tokens_seen": 13613880, + "step": 17850 + }, + { + "epoch": 37.12058212058212, + "grad_norm": 0.0004366971261333674, + "learning_rate": 0.17516243550485425, + "loss": 0.2543, + "num_input_tokens_seen": 13618008, + "step": 17855 + }, + { + "epoch": 37.13097713097713, + "grad_norm": 0.00046624906826764345, + "learning_rate": 0.17510436340693478, + "loss": 0.2406, + "num_input_tokens_seen": 13621816, + "step": 17860 + }, + { + "epoch": 37.141372141372145, + "grad_norm": 0.0006874442333355546, + "learning_rate": 0.175046287437607, + "loss": 0.2839, + "num_input_tokens_seen": 13625688, + "step": 17865 + }, + { + "epoch": 37.15176715176715, + "grad_norm": 9.917110583046451e-05, + "learning_rate": 0.17498820760582695, + "loss": 0.2619, + "num_input_tokens_seen": 13629688, + "step": 17870 + }, + { + "epoch": 37.16216216216216, + "grad_norm": 4.814877684111707e-05, + "learning_rate": 0.1749301239205512, + "loss": 0.2626, + "num_input_tokens_seen": 13633592, + "step": 17875 + }, + { + "epoch": 37.17255717255717, + "grad_norm": 8.301712659886107e-05, + "learning_rate": 0.1748720363907371, + "loss": 0.2737, + "num_input_tokens_seen": 13637400, + "step": 17880 + }, + { + "epoch": 37.182952182952185, + "grad_norm": 0.00011215340055059642, + "learning_rate": 0.17481394502534242, + "loss": 0.2646, + "num_input_tokens_seen": 13641144, + "step": 17885 + }, + { + "epoch": 37.19334719334719, + "grad_norm": 3.621155337896198e-05, + "learning_rate": 0.17475584983332562, + "loss": 0.2749, + "num_input_tokens_seen": 13644952, + "step": 17890 + }, + { + "epoch": 37.2037422037422, + "grad_norm": 0.0002319800405530259, + "learning_rate": 0.17469775082364558, + "loss": 0.2659, + "num_input_tokens_seen": 13648856, + "step": 17895 + }, + { + "epoch": 37.21413721413722, + "grad_norm": 0.00035933632170781493, + "learning_rate": 0.17463964800526205, + "loss": 0.2818, + "num_input_tokens_seen": 13652728, + "step": 17900 + }, + { + "epoch": 37.224532224532226, + "grad_norm": 0.0003076719294767827, + "learning_rate": 0.17458154138713522, + "loss": 0.2774, + "num_input_tokens_seen": 13656728, + "step": 17905 + }, + { + "epoch": 37.234927234927234, + "grad_norm": 0.000195908680325374, + "learning_rate": 0.17452343097822576, + "loss": 0.2915, + "num_input_tokens_seen": 13660504, + "step": 17910 + }, + { + "epoch": 37.24532224532224, + "grad_norm": 0.00023033608158584684, + "learning_rate": 0.17446531678749497, + "loss": 0.2638, + "num_input_tokens_seen": 13664216, + "step": 17915 + }, + { + "epoch": 37.25571725571726, + "grad_norm": 0.00047455556341446936, + "learning_rate": 0.17440719882390496, + "loss": 0.2756, + "num_input_tokens_seen": 13668088, + "step": 17920 + }, + { + "epoch": 37.266112266112266, + "grad_norm": 0.00016654242062941194, + "learning_rate": 0.17434907709641814, + "loss": 0.2704, + "num_input_tokens_seen": 13671928, + "step": 17925 + }, + { + "epoch": 37.276507276507274, + "grad_norm": 0.0005451777833513916, + "learning_rate": 0.17429095161399769, + "loss": 0.2444, + "num_input_tokens_seen": 13675736, + "step": 17930 + }, + { + "epoch": 37.28690228690229, + "grad_norm": 0.00013748749915976077, + "learning_rate": 0.1742328223856072, + "loss": 0.2783, + "num_input_tokens_seen": 13679608, + "step": 17935 + }, + { + "epoch": 37.2972972972973, + "grad_norm": 0.00015887801419012249, + "learning_rate": 0.174174689420211, + "loss": 0.2742, + "num_input_tokens_seen": 13683480, + "step": 17940 + }, + { + "epoch": 37.30769230769231, + "grad_norm": 6.824556476203725e-05, + "learning_rate": 0.1741165527267739, + "loss": 0.2809, + "num_input_tokens_seen": 13687320, + "step": 17945 + }, + { + "epoch": 37.318087318087315, + "grad_norm": 0.00013770848454441875, + "learning_rate": 0.17405841231426125, + "loss": 0.2508, + "num_input_tokens_seen": 13691000, + "step": 17950 + }, + { + "epoch": 37.32848232848233, + "grad_norm": 0.0002024432469625026, + "learning_rate": 0.1740002681916391, + "loss": 0.2667, + "num_input_tokens_seen": 13694808, + "step": 17955 + }, + { + "epoch": 37.33887733887734, + "grad_norm": 0.0001446391543140635, + "learning_rate": 0.17394212036787401, + "loss": 0.2661, + "num_input_tokens_seen": 13698584, + "step": 17960 + }, + { + "epoch": 37.34927234927235, + "grad_norm": 0.000454866501968354, + "learning_rate": 0.1738839688519331, + "loss": 0.2843, + "num_input_tokens_seen": 13702360, + "step": 17965 + }, + { + "epoch": 37.35966735966736, + "grad_norm": 6.412077345885336e-05, + "learning_rate": 0.17382581365278402, + "loss": 0.2808, + "num_input_tokens_seen": 13706104, + "step": 17970 + }, + { + "epoch": 37.37006237006237, + "grad_norm": 0.0001545573177281767, + "learning_rate": 0.17376765477939507, + "loss": 0.2788, + "num_input_tokens_seen": 13709880, + "step": 17975 + }, + { + "epoch": 37.38045738045738, + "grad_norm": 0.00026132227503694594, + "learning_rate": 0.1737094922407351, + "loss": 0.2689, + "num_input_tokens_seen": 13713688, + "step": 17980 + }, + { + "epoch": 37.39085239085239, + "grad_norm": 0.00041704546310938895, + "learning_rate": 0.1736513260457734, + "loss": 0.2739, + "num_input_tokens_seen": 13717464, + "step": 17985 + }, + { + "epoch": 37.4012474012474, + "grad_norm": 7.604605343658477e-05, + "learning_rate": 0.17359315620348006, + "loss": 0.2457, + "num_input_tokens_seen": 13721176, + "step": 17990 + }, + { + "epoch": 37.41164241164241, + "grad_norm": 0.00019248152966611087, + "learning_rate": 0.17353498272282547, + "loss": 0.2463, + "num_input_tokens_seen": 13724952, + "step": 17995 + }, + { + "epoch": 37.42203742203742, + "grad_norm": 0.00016061661881394684, + "learning_rate": 0.17347680561278087, + "loss": 0.2675, + "num_input_tokens_seen": 13728696, + "step": 18000 + }, + { + "epoch": 37.42203742203742, + "eval_loss": 0.24717116355895996, + "eval_runtime": 13.4109, + "eval_samples_per_second": 63.829, + "eval_steps_per_second": 15.957, + "num_input_tokens_seen": 13728696, + "step": 18000 + }, + { + "epoch": 37.432432432432435, + "grad_norm": 0.00032401090720668435, + "learning_rate": 0.1734186248823178, + "loss": 0.2756, + "num_input_tokens_seen": 13732408, + "step": 18005 + }, + { + "epoch": 37.44282744282744, + "grad_norm": 3.154646037728526e-05, + "learning_rate": 0.17336044054040844, + "loss": 0.2817, + "num_input_tokens_seen": 13736312, + "step": 18010 + }, + { + "epoch": 37.45322245322245, + "grad_norm": 0.00012456002878025174, + "learning_rate": 0.1733022525960256, + "loss": 0.2742, + "num_input_tokens_seen": 13740184, + "step": 18015 + }, + { + "epoch": 37.46361746361746, + "grad_norm": 0.0002284859656356275, + "learning_rate": 0.1732440610581426, + "loss": 0.2725, + "num_input_tokens_seen": 13743992, + "step": 18020 + }, + { + "epoch": 37.474012474012476, + "grad_norm": 0.00026105876895599067, + "learning_rate": 0.17318586593573326, + "loss": 0.2755, + "num_input_tokens_seen": 13747672, + "step": 18025 + }, + { + "epoch": 37.484407484407484, + "grad_norm": 0.00011538234684849158, + "learning_rate": 0.17312766723777204, + "loss": 0.2685, + "num_input_tokens_seen": 13751544, + "step": 18030 + }, + { + "epoch": 37.49480249480249, + "grad_norm": 0.00030437938403338194, + "learning_rate": 0.1730694649732339, + "loss": 0.264, + "num_input_tokens_seen": 13755320, + "step": 18035 + }, + { + "epoch": 37.50519750519751, + "grad_norm": 3.2187650504056364e-05, + "learning_rate": 0.17301125915109428, + "loss": 0.2754, + "num_input_tokens_seen": 13759288, + "step": 18040 + }, + { + "epoch": 37.515592515592516, + "grad_norm": 0.0005967272445559502, + "learning_rate": 0.17295304978032938, + "loss": 0.2377, + "num_input_tokens_seen": 13763128, + "step": 18045 + }, + { + "epoch": 37.525987525987524, + "grad_norm": 0.0003936718567274511, + "learning_rate": 0.17289483686991577, + "loss": 0.2502, + "num_input_tokens_seen": 13766904, + "step": 18050 + }, + { + "epoch": 37.53638253638254, + "grad_norm": 0.00026558240642771125, + "learning_rate": 0.1728366204288306, + "loss": 0.2691, + "num_input_tokens_seen": 13770776, + "step": 18055 + }, + { + "epoch": 37.54677754677755, + "grad_norm": 0.00019525873358361423, + "learning_rate": 0.17277840046605153, + "loss": 0.2795, + "num_input_tokens_seen": 13774712, + "step": 18060 + }, + { + "epoch": 37.55717255717256, + "grad_norm": 0.00030830211471766233, + "learning_rate": 0.17272017699055686, + "loss": 0.266, + "num_input_tokens_seen": 13778520, + "step": 18065 + }, + { + "epoch": 37.567567567567565, + "grad_norm": 3.836326868622564e-05, + "learning_rate": 0.17266195001132542, + "loss": 0.2665, + "num_input_tokens_seen": 13782200, + "step": 18070 + }, + { + "epoch": 37.57796257796258, + "grad_norm": 0.00025523462682031095, + "learning_rate": 0.17260371953733647, + "loss": 0.2709, + "num_input_tokens_seen": 13786104, + "step": 18075 + }, + { + "epoch": 37.58835758835759, + "grad_norm": 0.00011523273860802874, + "learning_rate": 0.1725454855775699, + "loss": 0.2586, + "num_input_tokens_seen": 13789784, + "step": 18080 + }, + { + "epoch": 37.5987525987526, + "grad_norm": 5.492265700013377e-05, + "learning_rate": 0.17248724814100616, + "loss": 0.2727, + "num_input_tokens_seen": 13793400, + "step": 18085 + }, + { + "epoch": 37.60914760914761, + "grad_norm": 0.00020145783491898328, + "learning_rate": 0.17242900723662619, + "loss": 0.2534, + "num_input_tokens_seen": 13797176, + "step": 18090 + }, + { + "epoch": 37.61954261954262, + "grad_norm": 0.00022032538254279643, + "learning_rate": 0.1723707628734114, + "loss": 0.2585, + "num_input_tokens_seen": 13800984, + "step": 18095 + }, + { + "epoch": 37.62993762993763, + "grad_norm": 0.00030590148526243865, + "learning_rate": 0.1723125150603438, + "loss": 0.2725, + "num_input_tokens_seen": 13804792, + "step": 18100 + }, + { + "epoch": 37.64033264033264, + "grad_norm": 0.00034034260897897184, + "learning_rate": 0.1722542638064061, + "loss": 0.2597, + "num_input_tokens_seen": 13808664, + "step": 18105 + }, + { + "epoch": 37.65072765072765, + "grad_norm": 9.597495227353647e-05, + "learning_rate": 0.17219600912058117, + "loss": 0.229, + "num_input_tokens_seen": 13812408, + "step": 18110 + }, + { + "epoch": 37.66112266112266, + "grad_norm": 0.0004968281136825681, + "learning_rate": 0.17213775101185272, + "loss": 0.2525, + "num_input_tokens_seen": 13816248, + "step": 18115 + }, + { + "epoch": 37.67151767151767, + "grad_norm": 0.00032090532477013767, + "learning_rate": 0.17207948948920485, + "loss": 0.3041, + "num_input_tokens_seen": 13820024, + "step": 18120 + }, + { + "epoch": 37.681912681912685, + "grad_norm": 0.00036634900607168674, + "learning_rate": 0.17202122456162228, + "loss": 0.2601, + "num_input_tokens_seen": 13823896, + "step": 18125 + }, + { + "epoch": 37.69230769230769, + "grad_norm": 0.00010695265518734232, + "learning_rate": 0.17196295623809013, + "loss": 0.2793, + "num_input_tokens_seen": 13827704, + "step": 18130 + }, + { + "epoch": 37.7027027027027, + "grad_norm": 0.00015288103895727545, + "learning_rate": 0.1719046845275941, + "loss": 0.2727, + "num_input_tokens_seen": 13831480, + "step": 18135 + }, + { + "epoch": 37.71309771309771, + "grad_norm": 5.0813450798159465e-05, + "learning_rate": 0.17184640943912044, + "loss": 0.2649, + "num_input_tokens_seen": 13835384, + "step": 18140 + }, + { + "epoch": 37.723492723492726, + "grad_norm": 9.579612378729507e-05, + "learning_rate": 0.1717881309816559, + "loss": 0.2863, + "num_input_tokens_seen": 13839128, + "step": 18145 + }, + { + "epoch": 37.733887733887734, + "grad_norm": 0.0004308338393457234, + "learning_rate": 0.1717298491641878, + "loss": 0.2754, + "num_input_tokens_seen": 13842968, + "step": 18150 + }, + { + "epoch": 37.74428274428274, + "grad_norm": 0.0004642101703211665, + "learning_rate": 0.17167156399570385, + "loss": 0.2555, + "num_input_tokens_seen": 13846776, + "step": 18155 + }, + { + "epoch": 37.75467775467776, + "grad_norm": 0.0002685229410417378, + "learning_rate": 0.17161327548519242, + "loss": 0.27, + "num_input_tokens_seen": 13850776, + "step": 18160 + }, + { + "epoch": 37.765072765072766, + "grad_norm": 0.00011166080366820097, + "learning_rate": 0.1715549836416423, + "loss": 0.2842, + "num_input_tokens_seen": 13854744, + "step": 18165 + }, + { + "epoch": 37.775467775467774, + "grad_norm": 0.0006886075716465712, + "learning_rate": 0.17149668847404279, + "loss": 0.2805, + "num_input_tokens_seen": 13858488, + "step": 18170 + }, + { + "epoch": 37.78586278586278, + "grad_norm": 9.7030853794422e-05, + "learning_rate": 0.1714383899913838, + "loss": 0.251, + "num_input_tokens_seen": 13862360, + "step": 18175 + }, + { + "epoch": 37.7962577962578, + "grad_norm": 0.0003226956177968532, + "learning_rate": 0.17138008820265563, + "loss": 0.2512, + "num_input_tokens_seen": 13866072, + "step": 18180 + }, + { + "epoch": 37.80665280665281, + "grad_norm": 0.0001526285195723176, + "learning_rate": 0.17132178311684917, + "loss": 0.251, + "num_input_tokens_seen": 13869816, + "step": 18185 + }, + { + "epoch": 37.817047817047815, + "grad_norm": 0.00013543125533033162, + "learning_rate": 0.1712634747429559, + "loss": 0.2374, + "num_input_tokens_seen": 13873624, + "step": 18190 + }, + { + "epoch": 37.82744282744283, + "grad_norm": 0.0001464757660869509, + "learning_rate": 0.17120516308996753, + "loss": 0.2742, + "num_input_tokens_seen": 13877496, + "step": 18195 + }, + { + "epoch": 37.83783783783784, + "grad_norm": 0.00010562594252405688, + "learning_rate": 0.17114684816687653, + "loss": 0.2572, + "num_input_tokens_seen": 13881368, + "step": 18200 + }, + { + "epoch": 37.83783783783784, + "eval_loss": 0.2514818608760834, + "eval_runtime": 13.4284, + "eval_samples_per_second": 63.746, + "eval_steps_per_second": 15.936, + "num_input_tokens_seen": 13881368, + "step": 18200 + }, + { + "epoch": 37.84823284823285, + "grad_norm": 9.730438614496961e-05, + "learning_rate": 0.17108852998267585, + "loss": 0.2931, + "num_input_tokens_seen": 13885176, + "step": 18205 + }, + { + "epoch": 37.858627858627855, + "grad_norm": 0.00015595727018080652, + "learning_rate": 0.17103020854635878, + "loss": 0.2795, + "num_input_tokens_seen": 13889048, + "step": 18210 + }, + { + "epoch": 37.86902286902287, + "grad_norm": 0.0006141007761470973, + "learning_rate": 0.1709718838669193, + "loss": 0.2826, + "num_input_tokens_seen": 13892792, + "step": 18215 + }, + { + "epoch": 37.87941787941788, + "grad_norm": 0.00013923853111919016, + "learning_rate": 0.17091355595335173, + "loss": 0.2505, + "num_input_tokens_seen": 13896696, + "step": 18220 + }, + { + "epoch": 37.88981288981289, + "grad_norm": 0.00017420900985598564, + "learning_rate": 0.17085522481465107, + "loss": 0.2526, + "num_input_tokens_seen": 13900504, + "step": 18225 + }, + { + "epoch": 37.9002079002079, + "grad_norm": 0.00016382864851038903, + "learning_rate": 0.17079689045981264, + "loss": 0.2481, + "num_input_tokens_seen": 13904312, + "step": 18230 + }, + { + "epoch": 37.91060291060291, + "grad_norm": 0.00036020600236952305, + "learning_rate": 0.17073855289783238, + "loss": 0.2947, + "num_input_tokens_seen": 13908152, + "step": 18235 + }, + { + "epoch": 37.92099792099792, + "grad_norm": 0.0004964638501405716, + "learning_rate": 0.1706802121377066, + "loss": 0.2735, + "num_input_tokens_seen": 13911832, + "step": 18240 + }, + { + "epoch": 37.931392931392935, + "grad_norm": 0.00011524632282089442, + "learning_rate": 0.17062186818843225, + "loss": 0.2824, + "num_input_tokens_seen": 13915544, + "step": 18245 + }, + { + "epoch": 37.94178794178794, + "grad_norm": 0.00013317755656316876, + "learning_rate": 0.17056352105900668, + "loss": 0.2726, + "num_input_tokens_seen": 13919288, + "step": 18250 + }, + { + "epoch": 37.95218295218295, + "grad_norm": 0.00017390323046129197, + "learning_rate": 0.17050517075842772, + "loss": 0.2711, + "num_input_tokens_seen": 13923256, + "step": 18255 + }, + { + "epoch": 37.96257796257796, + "grad_norm": 5.497239180840552e-05, + "learning_rate": 0.17044681729569375, + "loss": 0.2592, + "num_input_tokens_seen": 13926936, + "step": 18260 + }, + { + "epoch": 37.972972972972975, + "grad_norm": 0.00011403862299630418, + "learning_rate": 0.17038846067980365, + "loss": 0.2657, + "num_input_tokens_seen": 13930776, + "step": 18265 + }, + { + "epoch": 37.983367983367984, + "grad_norm": 0.0002356619806960225, + "learning_rate": 0.17033010091975664, + "loss": 0.2551, + "num_input_tokens_seen": 13934424, + "step": 18270 + }, + { + "epoch": 37.99376299376299, + "grad_norm": 0.00011265333159826696, + "learning_rate": 0.17027173802455262, + "loss": 0.2585, + "num_input_tokens_seen": 13938200, + "step": 18275 + }, + { + "epoch": 38.00415800415801, + "grad_norm": 0.00018976478895638138, + "learning_rate": 0.1702133720031918, + "loss": 0.2567, + "num_input_tokens_seen": 13942000, + "step": 18280 + }, + { + "epoch": 38.014553014553016, + "grad_norm": 8.800016075838357e-05, + "learning_rate": 0.17015500286467503, + "loss": 0.2158, + "num_input_tokens_seen": 13945744, + "step": 18285 + }, + { + "epoch": 38.024948024948024, + "grad_norm": 0.00018180751067120582, + "learning_rate": 0.17009663061800354, + "loss": 0.2659, + "num_input_tokens_seen": 13949520, + "step": 18290 + }, + { + "epoch": 38.03534303534303, + "grad_norm": 0.0003712528559844941, + "learning_rate": 0.17003825527217903, + "loss": 0.2681, + "num_input_tokens_seen": 13953328, + "step": 18295 + }, + { + "epoch": 38.04573804573805, + "grad_norm": 0.00019450522086117417, + "learning_rate": 0.16997987683620377, + "loss": 0.2645, + "num_input_tokens_seen": 13957264, + "step": 18300 + }, + { + "epoch": 38.056133056133056, + "grad_norm": 0.00012056219566147774, + "learning_rate": 0.16992149531908043, + "loss": 0.2691, + "num_input_tokens_seen": 13961136, + "step": 18305 + }, + { + "epoch": 38.066528066528065, + "grad_norm": 0.0002713422290980816, + "learning_rate": 0.16986311072981214, + "loss": 0.2804, + "num_input_tokens_seen": 13964848, + "step": 18310 + }, + { + "epoch": 38.07692307692308, + "grad_norm": 0.0002110823552357033, + "learning_rate": 0.16980472307740255, + "loss": 0.2649, + "num_input_tokens_seen": 13968560, + "step": 18315 + }, + { + "epoch": 38.08731808731809, + "grad_norm": 0.00011486118455650285, + "learning_rate": 0.1697463323708558, + "loss": 0.265, + "num_input_tokens_seen": 13972336, + "step": 18320 + }, + { + "epoch": 38.0977130977131, + "grad_norm": 0.0002966140164062381, + "learning_rate": 0.16968793861917641, + "loss": 0.2651, + "num_input_tokens_seen": 13976144, + "step": 18325 + }, + { + "epoch": 38.108108108108105, + "grad_norm": 0.0004858723550569266, + "learning_rate": 0.16962954183136952, + "loss": 0.282, + "num_input_tokens_seen": 13979984, + "step": 18330 + }, + { + "epoch": 38.11850311850312, + "grad_norm": 0.00017579954874236137, + "learning_rate": 0.16957114201644058, + "loss": 0.2501, + "num_input_tokens_seen": 13983920, + "step": 18335 + }, + { + "epoch": 38.12889812889813, + "grad_norm": 1.1696202818711754e-05, + "learning_rate": 0.16951273918339563, + "loss": 0.2559, + "num_input_tokens_seen": 13987792, + "step": 18340 + }, + { + "epoch": 38.13929313929314, + "grad_norm": 0.00018843628640752286, + "learning_rate": 0.16945433334124105, + "loss": 0.2736, + "num_input_tokens_seen": 13991472, + "step": 18345 + }, + { + "epoch": 38.14968814968815, + "grad_norm": 0.00021506579651031643, + "learning_rate": 0.1693959244989838, + "loss": 0.2422, + "num_input_tokens_seen": 13995216, + "step": 18350 + }, + { + "epoch": 38.16008316008316, + "grad_norm": 0.0003718194493558258, + "learning_rate": 0.16933751266563127, + "loss": 0.2792, + "num_input_tokens_seen": 13998992, + "step": 18355 + }, + { + "epoch": 38.17047817047817, + "grad_norm": 0.00024612279958091676, + "learning_rate": 0.16927909785019118, + "loss": 0.2697, + "num_input_tokens_seen": 14002896, + "step": 18360 + }, + { + "epoch": 38.18087318087318, + "grad_norm": 0.00029218298732303083, + "learning_rate": 0.169220680061672, + "loss": 0.2789, + "num_input_tokens_seen": 14006672, + "step": 18365 + }, + { + "epoch": 38.19126819126819, + "grad_norm": 0.0004712914233095944, + "learning_rate": 0.16916225930908244, + "loss": 0.247, + "num_input_tokens_seen": 14010320, + "step": 18370 + }, + { + "epoch": 38.2016632016632, + "grad_norm": 0.00022511265706270933, + "learning_rate": 0.16910383560143163, + "loss": 0.2564, + "num_input_tokens_seen": 14014160, + "step": 18375 + }, + { + "epoch": 38.21205821205821, + "grad_norm": 0.0003480313753243536, + "learning_rate": 0.16904540894772935, + "loss": 0.2743, + "num_input_tokens_seen": 14018032, + "step": 18380 + }, + { + "epoch": 38.222453222453225, + "grad_norm": 0.00028071028646081686, + "learning_rate": 0.16898697935698562, + "loss": 0.2459, + "num_input_tokens_seen": 14021872, + "step": 18385 + }, + { + "epoch": 38.232848232848234, + "grad_norm": 0.00021995829592924565, + "learning_rate": 0.1689285468382111, + "loss": 0.2616, + "num_input_tokens_seen": 14025872, + "step": 18390 + }, + { + "epoch": 38.24324324324324, + "grad_norm": 4.734143294626847e-05, + "learning_rate": 0.16887011140041677, + "loss": 0.2651, + "num_input_tokens_seen": 14029776, + "step": 18395 + }, + { + "epoch": 38.25363825363825, + "grad_norm": 0.0006571648991666734, + "learning_rate": 0.1688116730526141, + "loss": 0.2781, + "num_input_tokens_seen": 14033616, + "step": 18400 + }, + { + "epoch": 38.25363825363825, + "eval_loss": 0.26234912872314453, + "eval_runtime": 13.4397, + "eval_samples_per_second": 63.692, + "eval_steps_per_second": 15.923, + "num_input_tokens_seen": 14033616, + "step": 18400 + }, + { + "epoch": 38.264033264033266, + "grad_norm": 0.0002502047864254564, + "learning_rate": 0.1687532318038151, + "loss": 0.2762, + "num_input_tokens_seen": 14037392, + "step": 18405 + }, + { + "epoch": 38.274428274428274, + "grad_norm": 0.00025649057351984084, + "learning_rate": 0.16869478766303206, + "loss": 0.2533, + "num_input_tokens_seen": 14041264, + "step": 18410 + }, + { + "epoch": 38.28482328482328, + "grad_norm": 9.528319060336798e-05, + "learning_rate": 0.16863634063927788, + "loss": 0.2458, + "num_input_tokens_seen": 14045168, + "step": 18415 + }, + { + "epoch": 38.2952182952183, + "grad_norm": 4.3037463910877705e-05, + "learning_rate": 0.16857789074156568, + "loss": 0.2758, + "num_input_tokens_seen": 14048944, + "step": 18420 + }, + { + "epoch": 38.305613305613306, + "grad_norm": 8.081365376710892e-05, + "learning_rate": 0.16851943797890928, + "loss": 0.2961, + "num_input_tokens_seen": 14052784, + "step": 18425 + }, + { + "epoch": 38.316008316008315, + "grad_norm": 0.00034480728209018707, + "learning_rate": 0.16846098236032284, + "loss": 0.2726, + "num_input_tokens_seen": 14056592, + "step": 18430 + }, + { + "epoch": 38.32640332640332, + "grad_norm": 3.981667759944685e-05, + "learning_rate": 0.16840252389482097, + "loss": 0.2748, + "num_input_tokens_seen": 14060208, + "step": 18435 + }, + { + "epoch": 38.33679833679834, + "grad_norm": 0.00014087762974668294, + "learning_rate": 0.16834406259141857, + "loss": 0.2682, + "num_input_tokens_seen": 14063984, + "step": 18440 + }, + { + "epoch": 38.34719334719335, + "grad_norm": 2.8825385015807115e-05, + "learning_rate": 0.16828559845913124, + "loss": 0.2487, + "num_input_tokens_seen": 14067696, + "step": 18445 + }, + { + "epoch": 38.357588357588355, + "grad_norm": 9.348901949124411e-05, + "learning_rate": 0.16822713150697488, + "loss": 0.2818, + "num_input_tokens_seen": 14071504, + "step": 18450 + }, + { + "epoch": 38.36798336798337, + "grad_norm": 0.00020476299687288702, + "learning_rate": 0.16816866174396575, + "loss": 0.2428, + "num_input_tokens_seen": 14075344, + "step": 18455 + }, + { + "epoch": 38.37837837837838, + "grad_norm": 5.3525480325333774e-05, + "learning_rate": 0.16811018917912057, + "loss": 0.2606, + "num_input_tokens_seen": 14079152, + "step": 18460 + }, + { + "epoch": 38.38877338877339, + "grad_norm": 0.0002666098007466644, + "learning_rate": 0.16805171382145673, + "loss": 0.2703, + "num_input_tokens_seen": 14082928, + "step": 18465 + }, + { + "epoch": 38.3991683991684, + "grad_norm": 0.0005928517202846706, + "learning_rate": 0.16799323567999175, + "loss": 0.2723, + "num_input_tokens_seen": 14086704, + "step": 18470 + }, + { + "epoch": 38.40956340956341, + "grad_norm": 0.00010158977238461375, + "learning_rate": 0.16793475476374367, + "loss": 0.269, + "num_input_tokens_seen": 14090512, + "step": 18475 + }, + { + "epoch": 38.41995841995842, + "grad_norm": 0.0005319610936567187, + "learning_rate": 0.1678762710817311, + "loss": 0.2726, + "num_input_tokens_seen": 14094160, + "step": 18480 + }, + { + "epoch": 38.43035343035343, + "grad_norm": 5.5819586123107e-05, + "learning_rate": 0.1678177846429728, + "loss": 0.2588, + "num_input_tokens_seen": 14098000, + "step": 18485 + }, + { + "epoch": 38.44074844074844, + "grad_norm": 0.0001422505738446489, + "learning_rate": 0.16775929545648827, + "loss": 0.262, + "num_input_tokens_seen": 14101616, + "step": 18490 + }, + { + "epoch": 38.45114345114345, + "grad_norm": 9.092781692743301e-05, + "learning_rate": 0.16770080353129715, + "loss": 0.2825, + "num_input_tokens_seen": 14105328, + "step": 18495 + }, + { + "epoch": 38.46153846153846, + "grad_norm": 0.00020319767645560205, + "learning_rate": 0.16764230887641968, + "loss": 0.2583, + "num_input_tokens_seen": 14109168, + "step": 18500 + }, + { + "epoch": 38.471933471933475, + "grad_norm": 7.987016579136252e-05, + "learning_rate": 0.1675838115008765, + "loss": 0.2711, + "num_input_tokens_seen": 14113040, + "step": 18505 + }, + { + "epoch": 38.482328482328484, + "grad_norm": 3.809152985922992e-05, + "learning_rate": 0.1675253114136886, + "loss": 0.2376, + "num_input_tokens_seen": 14116944, + "step": 18510 + }, + { + "epoch": 38.49272349272349, + "grad_norm": 3.1968440453056246e-05, + "learning_rate": 0.16746680862387747, + "loss": 0.2897, + "num_input_tokens_seen": 14120688, + "step": 18515 + }, + { + "epoch": 38.5031185031185, + "grad_norm": 0.0002977905096486211, + "learning_rate": 0.16740830314046493, + "loss": 0.2704, + "num_input_tokens_seen": 14124560, + "step": 18520 + }, + { + "epoch": 38.513513513513516, + "grad_norm": 4.5603363105328754e-05, + "learning_rate": 0.1673497949724733, + "loss": 0.2873, + "num_input_tokens_seen": 14128400, + "step": 18525 + }, + { + "epoch": 38.523908523908524, + "grad_norm": 0.00021074705000501126, + "learning_rate": 0.16729128412892522, + "loss": 0.2898, + "num_input_tokens_seen": 14132208, + "step": 18530 + }, + { + "epoch": 38.53430353430353, + "grad_norm": 6.72453606966883e-05, + "learning_rate": 0.16723277061884384, + "loss": 0.2791, + "num_input_tokens_seen": 14136016, + "step": 18535 + }, + { + "epoch": 38.54469854469855, + "grad_norm": 0.00014155970711726695, + "learning_rate": 0.16717425445125267, + "loss": 0.2543, + "num_input_tokens_seen": 14139856, + "step": 18540 + }, + { + "epoch": 38.555093555093556, + "grad_norm": 2.937451245088596e-05, + "learning_rate": 0.16711573563517565, + "loss": 0.2714, + "num_input_tokens_seen": 14143696, + "step": 18545 + }, + { + "epoch": 38.565488565488565, + "grad_norm": 4.8365214752266183e-05, + "learning_rate": 0.1670572141796371, + "loss": 0.2567, + "num_input_tokens_seen": 14147600, + "step": 18550 + }, + { + "epoch": 38.57588357588357, + "grad_norm": 0.00045258383033797145, + "learning_rate": 0.16699869009366175, + "loss": 0.2781, + "num_input_tokens_seen": 14151440, + "step": 18555 + }, + { + "epoch": 38.58627858627859, + "grad_norm": 0.00012728873116429895, + "learning_rate": 0.1669401633862748, + "loss": 0.2709, + "num_input_tokens_seen": 14155312, + "step": 18560 + }, + { + "epoch": 38.5966735966736, + "grad_norm": 0.00024172550183720887, + "learning_rate": 0.16688163406650178, + "loss": 0.2724, + "num_input_tokens_seen": 14159152, + "step": 18565 + }, + { + "epoch": 38.607068607068605, + "grad_norm": 0.0006905028130859137, + "learning_rate": 0.1668231021433686, + "loss": 0.2677, + "num_input_tokens_seen": 14162992, + "step": 18570 + }, + { + "epoch": 38.61746361746362, + "grad_norm": 0.0003708757576532662, + "learning_rate": 0.1667645676259017, + "loss": 0.278, + "num_input_tokens_seen": 14166864, + "step": 18575 + }, + { + "epoch": 38.62785862785863, + "grad_norm": 0.00017842563102021813, + "learning_rate": 0.1667060305231277, + "loss": 0.2631, + "num_input_tokens_seen": 14170640, + "step": 18580 + }, + { + "epoch": 38.63825363825364, + "grad_norm": 0.0001081644804798998, + "learning_rate": 0.16664749084407396, + "loss": 0.2539, + "num_input_tokens_seen": 14174384, + "step": 18585 + }, + { + "epoch": 38.648648648648646, + "grad_norm": 0.00019887782400473952, + "learning_rate": 0.16658894859776788, + "loss": 0.247, + "num_input_tokens_seen": 14178000, + "step": 18590 + }, + { + "epoch": 38.65904365904366, + "grad_norm": 0.00016245538427028805, + "learning_rate": 0.16653040379323752, + "loss": 0.2697, + "num_input_tokens_seen": 14181840, + "step": 18595 + }, + { + "epoch": 38.66943866943867, + "grad_norm": 0.00024234774173237383, + "learning_rate": 0.16647185643951107, + "loss": 0.2692, + "num_input_tokens_seen": 14185616, + "step": 18600 + }, + { + "epoch": 38.66943866943867, + "eval_loss": 0.24956050515174866, + "eval_runtime": 13.4245, + "eval_samples_per_second": 63.764, + "eval_steps_per_second": 15.941, + "num_input_tokens_seen": 14185616, + "step": 18600 + }, + { + "epoch": 38.67983367983368, + "grad_norm": 0.0001595732319401577, + "learning_rate": 0.1664133065456174, + "loss": 0.2644, + "num_input_tokens_seen": 14189456, + "step": 18605 + }, + { + "epoch": 38.69022869022869, + "grad_norm": 0.00010833367559826002, + "learning_rate": 0.1663547541205856, + "loss": 0.2635, + "num_input_tokens_seen": 14193168, + "step": 18610 + }, + { + "epoch": 38.7006237006237, + "grad_norm": 0.0004917460610158741, + "learning_rate": 0.16629619917344518, + "loss": 0.2603, + "num_input_tokens_seen": 14197008, + "step": 18615 + }, + { + "epoch": 38.71101871101871, + "grad_norm": 0.00015672293375246227, + "learning_rate": 0.16623764171322605, + "loss": 0.2778, + "num_input_tokens_seen": 14200912, + "step": 18620 + }, + { + "epoch": 38.72141372141372, + "grad_norm": 0.00013964631943963468, + "learning_rate": 0.1661790817489585, + "loss": 0.2879, + "num_input_tokens_seen": 14204816, + "step": 18625 + }, + { + "epoch": 38.731808731808734, + "grad_norm": 0.0003688719589263201, + "learning_rate": 0.16612051928967328, + "loss": 0.2748, + "num_input_tokens_seen": 14208688, + "step": 18630 + }, + { + "epoch": 38.74220374220374, + "grad_norm": 0.00041695183608680964, + "learning_rate": 0.16606195434440138, + "loss": 0.2704, + "num_input_tokens_seen": 14212528, + "step": 18635 + }, + { + "epoch": 38.75259875259875, + "grad_norm": 0.00018519068544264883, + "learning_rate": 0.16600338692217426, + "loss": 0.2593, + "num_input_tokens_seen": 14216240, + "step": 18640 + }, + { + "epoch": 38.762993762993766, + "grad_norm": 0.000599623192101717, + "learning_rate": 0.16594481703202374, + "loss": 0.28, + "num_input_tokens_seen": 14220144, + "step": 18645 + }, + { + "epoch": 38.773388773388774, + "grad_norm": 0.0002322217624168843, + "learning_rate": 0.1658862446829821, + "loss": 0.2562, + "num_input_tokens_seen": 14223984, + "step": 18650 + }, + { + "epoch": 38.78378378378378, + "grad_norm": 0.0001424806978320703, + "learning_rate": 0.16582766988408187, + "loss": 0.2689, + "num_input_tokens_seen": 14227856, + "step": 18655 + }, + { + "epoch": 38.79417879417879, + "grad_norm": 3.8309521187329665e-05, + "learning_rate": 0.16576909264435608, + "loss": 0.2641, + "num_input_tokens_seen": 14231728, + "step": 18660 + }, + { + "epoch": 38.804573804573806, + "grad_norm": 6.855159881524742e-05, + "learning_rate": 0.16571051297283798, + "loss": 0.2592, + "num_input_tokens_seen": 14235664, + "step": 18665 + }, + { + "epoch": 38.814968814968815, + "grad_norm": 0.00028376287082210183, + "learning_rate": 0.16565193087856137, + "loss": 0.2727, + "num_input_tokens_seen": 14239408, + "step": 18670 + }, + { + "epoch": 38.82536382536382, + "grad_norm": 0.0004970021545886993, + "learning_rate": 0.16559334637056033, + "loss": 0.2695, + "num_input_tokens_seen": 14243216, + "step": 18675 + }, + { + "epoch": 38.83575883575884, + "grad_norm": 0.00023062255058903247, + "learning_rate": 0.16553475945786933, + "loss": 0.2775, + "num_input_tokens_seen": 14247312, + "step": 18680 + }, + { + "epoch": 38.84615384615385, + "grad_norm": 0.0006265403935685754, + "learning_rate": 0.16547617014952318, + "loss": 0.2647, + "num_input_tokens_seen": 14251056, + "step": 18685 + }, + { + "epoch": 38.856548856548855, + "grad_norm": 0.00014990029740147293, + "learning_rate": 0.1654175784545571, + "loss": 0.2425, + "num_input_tokens_seen": 14254800, + "step": 18690 + }, + { + "epoch": 38.86694386694387, + "grad_norm": 0.0006807952304370701, + "learning_rate": 0.1653589843820067, + "loss": 0.2853, + "num_input_tokens_seen": 14258960, + "step": 18695 + }, + { + "epoch": 38.87733887733888, + "grad_norm": 0.00014907633885741234, + "learning_rate": 0.1653003879409079, + "loss": 0.2678, + "num_input_tokens_seen": 14262736, + "step": 18700 + }, + { + "epoch": 38.88773388773389, + "grad_norm": 0.00019776716362684965, + "learning_rate": 0.165241789140297, + "loss": 0.266, + "num_input_tokens_seen": 14266512, + "step": 18705 + }, + { + "epoch": 38.898128898128896, + "grad_norm": 0.00017867582209873945, + "learning_rate": 0.16518318798921064, + "loss": 0.262, + "num_input_tokens_seen": 14270320, + "step": 18710 + }, + { + "epoch": 38.90852390852391, + "grad_norm": 0.00011465382704045624, + "learning_rate": 0.16512458449668593, + "loss": 0.2772, + "num_input_tokens_seen": 14274064, + "step": 18715 + }, + { + "epoch": 38.91891891891892, + "grad_norm": 9.577245509717613e-05, + "learning_rate": 0.1650659786717602, + "loss": 0.2488, + "num_input_tokens_seen": 14277968, + "step": 18720 + }, + { + "epoch": 38.92931392931393, + "grad_norm": 0.00030009375768713653, + "learning_rate": 0.1650073705234712, + "loss": 0.2529, + "num_input_tokens_seen": 14281904, + "step": 18725 + }, + { + "epoch": 38.93970893970894, + "grad_norm": 0.00014796375762671232, + "learning_rate": 0.16494876006085712, + "loss": 0.2595, + "num_input_tokens_seen": 14285712, + "step": 18730 + }, + { + "epoch": 38.95010395010395, + "grad_norm": 0.0006412433576770127, + "learning_rate": 0.16489014729295634, + "loss": 0.2824, + "num_input_tokens_seen": 14289488, + "step": 18735 + }, + { + "epoch": 38.96049896049896, + "grad_norm": 0.00011991824430879205, + "learning_rate": 0.16483153222880775, + "loss": 0.2706, + "num_input_tokens_seen": 14293264, + "step": 18740 + }, + { + "epoch": 38.97089397089397, + "grad_norm": 0.0005097811808809638, + "learning_rate": 0.16477291487745052, + "loss": 0.2733, + "num_input_tokens_seen": 14297040, + "step": 18745 + }, + { + "epoch": 38.981288981288984, + "grad_norm": 0.0003074663982260972, + "learning_rate": 0.16471429524792416, + "loss": 0.2727, + "num_input_tokens_seen": 14300688, + "step": 18750 + }, + { + "epoch": 38.99168399168399, + "grad_norm": 9.364308789372444e-05, + "learning_rate": 0.16465567334926856, + "loss": 0.2635, + "num_input_tokens_seen": 14304496, + "step": 18755 + }, + { + "epoch": 39.002079002079, + "grad_norm": 0.00017646928608883172, + "learning_rate": 0.16459704919052395, + "loss": 0.2446, + "num_input_tokens_seen": 14308320, + "step": 18760 + }, + { + "epoch": 39.012474012474016, + "grad_norm": 0.00029763428028672934, + "learning_rate": 0.16453842278073086, + "loss": 0.259, + "num_input_tokens_seen": 14312096, + "step": 18765 + }, + { + "epoch": 39.022869022869024, + "grad_norm": 0.00030581539613194764, + "learning_rate": 0.16447979412893038, + "loss": 0.2727, + "num_input_tokens_seen": 14315936, + "step": 18770 + }, + { + "epoch": 39.03326403326403, + "grad_norm": 0.00025723359431140125, + "learning_rate": 0.16442116324416367, + "loss": 0.2451, + "num_input_tokens_seen": 14319648, + "step": 18775 + }, + { + "epoch": 39.04365904365904, + "grad_norm": 9.510527888778597e-05, + "learning_rate": 0.1643625301354723, + "loss": 0.2866, + "num_input_tokens_seen": 14323456, + "step": 18780 + }, + { + "epoch": 39.054054054054056, + "grad_norm": 9.29101079236716e-05, + "learning_rate": 0.16430389481189828, + "loss": 0.2638, + "num_input_tokens_seen": 14327328, + "step": 18785 + }, + { + "epoch": 39.064449064449065, + "grad_norm": 0.00034269114257767797, + "learning_rate": 0.164245257282484, + "loss": 0.2647, + "num_input_tokens_seen": 14331136, + "step": 18790 + }, + { + "epoch": 39.07484407484407, + "grad_norm": 0.00016197112563531846, + "learning_rate": 0.16418661755627195, + "loss": 0.2804, + "num_input_tokens_seen": 14334944, + "step": 18795 + }, + { + "epoch": 39.08523908523909, + "grad_norm": 0.0004046610265504569, + "learning_rate": 0.16412797564230527, + "loss": 0.2311, + "num_input_tokens_seen": 14338720, + "step": 18800 + }, + { + "epoch": 39.08523908523909, + "eval_loss": 0.24795089662075043, + "eval_runtime": 13.4589, + "eval_samples_per_second": 63.601, + "eval_steps_per_second": 15.9, + "num_input_tokens_seen": 14338720, + "step": 18800 + }, + { + "epoch": 39.0956340956341, + "grad_norm": 0.00024159486929420382, + "learning_rate": 0.16406933154962713, + "loss": 0.2867, + "num_input_tokens_seen": 14342400, + "step": 18805 + }, + { + "epoch": 39.106029106029105, + "grad_norm": 6.994723662501201e-05, + "learning_rate": 0.16401068528728133, + "loss": 0.2723, + "num_input_tokens_seen": 14346016, + "step": 18810 + }, + { + "epoch": 39.11642411642411, + "grad_norm": 5.400833106250502e-05, + "learning_rate": 0.16395203686431173, + "loss": 0.2734, + "num_input_tokens_seen": 14349824, + "step": 18815 + }, + { + "epoch": 39.12681912681913, + "grad_norm": 0.0003098831803072244, + "learning_rate": 0.16389338628976277, + "loss": 0.2724, + "num_input_tokens_seen": 14353696, + "step": 18820 + }, + { + "epoch": 39.13721413721414, + "grad_norm": 0.0002838727959897369, + "learning_rate": 0.163834733572679, + "loss": 0.2744, + "num_input_tokens_seen": 14357312, + "step": 18825 + }, + { + "epoch": 39.147609147609145, + "grad_norm": 0.00011237752187298611, + "learning_rate": 0.16377607872210545, + "loss": 0.2661, + "num_input_tokens_seen": 14361184, + "step": 18830 + }, + { + "epoch": 39.15800415800416, + "grad_norm": 0.0003228596178814769, + "learning_rate": 0.16371742174708748, + "loss": 0.2209, + "num_input_tokens_seen": 14364896, + "step": 18835 + }, + { + "epoch": 39.16839916839917, + "grad_norm": 0.0002744507510215044, + "learning_rate": 0.16365876265667065, + "loss": 0.2532, + "num_input_tokens_seen": 14368544, + "step": 18840 + }, + { + "epoch": 39.17879417879418, + "grad_norm": 0.000546666793525219, + "learning_rate": 0.163600101459901, + "loss": 0.3089, + "num_input_tokens_seen": 14372288, + "step": 18845 + }, + { + "epoch": 39.189189189189186, + "grad_norm": 0.00024364182900171727, + "learning_rate": 0.16354143816582484, + "loss": 0.2577, + "num_input_tokens_seen": 14376096, + "step": 18850 + }, + { + "epoch": 39.1995841995842, + "grad_norm": 6.36070326436311e-05, + "learning_rate": 0.1634827727834887, + "loss": 0.2743, + "num_input_tokens_seen": 14379872, + "step": 18855 + }, + { + "epoch": 39.20997920997921, + "grad_norm": 0.00012105060886824504, + "learning_rate": 0.16342410532193954, + "loss": 0.2624, + "num_input_tokens_seen": 14383712, + "step": 18860 + }, + { + "epoch": 39.22037422037422, + "grad_norm": 0.000394036847865209, + "learning_rate": 0.16336543579022464, + "loss": 0.2792, + "num_input_tokens_seen": 14387648, + "step": 18865 + }, + { + "epoch": 39.23076923076923, + "grad_norm": 0.00014412598102353513, + "learning_rate": 0.16330676419739157, + "loss": 0.2391, + "num_input_tokens_seen": 14391488, + "step": 18870 + }, + { + "epoch": 39.24116424116424, + "grad_norm": 0.00044043874368071556, + "learning_rate": 0.1632480905524883, + "loss": 0.2876, + "num_input_tokens_seen": 14395168, + "step": 18875 + }, + { + "epoch": 39.25155925155925, + "grad_norm": 9.197332110488787e-05, + "learning_rate": 0.16318941486456293, + "loss": 0.2671, + "num_input_tokens_seen": 14398976, + "step": 18880 + }, + { + "epoch": 39.26195426195426, + "grad_norm": 0.0002466531877871603, + "learning_rate": 0.16313073714266405, + "loss": 0.2749, + "num_input_tokens_seen": 14402816, + "step": 18885 + }, + { + "epoch": 39.272349272349274, + "grad_norm": 0.000277255050605163, + "learning_rate": 0.16307205739584052, + "loss": 0.2639, + "num_input_tokens_seen": 14406496, + "step": 18890 + }, + { + "epoch": 39.28274428274428, + "grad_norm": 0.0004321524756960571, + "learning_rate": 0.16301337563314144, + "loss": 0.2454, + "num_input_tokens_seen": 14410400, + "step": 18895 + }, + { + "epoch": 39.29313929313929, + "grad_norm": 0.0004221178824082017, + "learning_rate": 0.1629546918636163, + "loss": 0.3007, + "num_input_tokens_seen": 14414144, + "step": 18900 + }, + { + "epoch": 39.303534303534306, + "grad_norm": 7.016614836174995e-05, + "learning_rate": 0.16289600609631485, + "loss": 0.2619, + "num_input_tokens_seen": 14417952, + "step": 18905 + }, + { + "epoch": 39.313929313929314, + "grad_norm": 0.00027923999004997313, + "learning_rate": 0.16283731834028722, + "loss": 0.2525, + "num_input_tokens_seen": 14421632, + "step": 18910 + }, + { + "epoch": 39.32432432432432, + "grad_norm": 0.0005079591064713895, + "learning_rate": 0.16277862860458378, + "loss": 0.289, + "num_input_tokens_seen": 14425504, + "step": 18915 + }, + { + "epoch": 39.33471933471934, + "grad_norm": 0.0005426101852208376, + "learning_rate": 0.16271993689825526, + "loss": 0.2808, + "num_input_tokens_seen": 14429472, + "step": 18920 + }, + { + "epoch": 39.34511434511435, + "grad_norm": 0.00028841980383731425, + "learning_rate": 0.1626612432303526, + "loss": 0.2728, + "num_input_tokens_seen": 14433120, + "step": 18925 + }, + { + "epoch": 39.355509355509355, + "grad_norm": 0.0003430134674999863, + "learning_rate": 0.1626025476099271, + "loss": 0.2733, + "num_input_tokens_seen": 14436896, + "step": 18930 + }, + { + "epoch": 39.36590436590436, + "grad_norm": 0.0004487436672206968, + "learning_rate": 0.1625438500460304, + "loss": 0.2576, + "num_input_tokens_seen": 14440672, + "step": 18935 + }, + { + "epoch": 39.37629937629938, + "grad_norm": 0.00038152706110849977, + "learning_rate": 0.16248515054771442, + "loss": 0.2201, + "num_input_tokens_seen": 14444448, + "step": 18940 + }, + { + "epoch": 39.38669438669439, + "grad_norm": 0.000663946324493736, + "learning_rate": 0.16242644912403123, + "loss": 0.3338, + "num_input_tokens_seen": 14448320, + "step": 18945 + }, + { + "epoch": 39.397089397089395, + "grad_norm": 0.00042350523290224373, + "learning_rate": 0.1623677457840335, + "loss": 0.3054, + "num_input_tokens_seen": 14452128, + "step": 18950 + }, + { + "epoch": 39.40748440748441, + "grad_norm": 0.00016630318714305758, + "learning_rate": 0.16230904053677397, + "loss": 0.2759, + "num_input_tokens_seen": 14455744, + "step": 18955 + }, + { + "epoch": 39.41787941787942, + "grad_norm": 0.00023443266400136054, + "learning_rate": 0.16225033339130568, + "loss": 0.2796, + "num_input_tokens_seen": 14459488, + "step": 18960 + }, + { + "epoch": 39.42827442827443, + "grad_norm": 0.0002817730710376054, + "learning_rate": 0.16219162435668197, + "loss": 0.2573, + "num_input_tokens_seen": 14463296, + "step": 18965 + }, + { + "epoch": 39.438669438669436, + "grad_norm": 0.0005313310539349914, + "learning_rate": 0.16213291344195666, + "loss": 0.2766, + "num_input_tokens_seen": 14467136, + "step": 18970 + }, + { + "epoch": 39.44906444906445, + "grad_norm": 0.0002998763229697943, + "learning_rate": 0.16207420065618358, + "loss": 0.283, + "num_input_tokens_seen": 14471008, + "step": 18975 + }, + { + "epoch": 39.45945945945946, + "grad_norm": 0.00015673093730583787, + "learning_rate": 0.16201548600841706, + "loss": 0.2741, + "num_input_tokens_seen": 14474848, + "step": 18980 + }, + { + "epoch": 39.46985446985447, + "grad_norm": 0.00018048730271402746, + "learning_rate": 0.16195676950771154, + "loss": 0.2759, + "num_input_tokens_seen": 14478720, + "step": 18985 + }, + { + "epoch": 39.48024948024948, + "grad_norm": 0.00031342243892140687, + "learning_rate": 0.16189805116312198, + "loss": 0.2647, + "num_input_tokens_seen": 14482720, + "step": 18990 + }, + { + "epoch": 39.49064449064449, + "grad_norm": 0.0005804307293146849, + "learning_rate": 0.16183933098370337, + "loss": 0.2656, + "num_input_tokens_seen": 14486560, + "step": 18995 + }, + { + "epoch": 39.5010395010395, + "grad_norm": 5.875601709703915e-05, + "learning_rate": 0.16178060897851115, + "loss": 0.2502, + "num_input_tokens_seen": 14490240, + "step": 19000 + }, + { + "epoch": 39.5010395010395, + "eval_loss": 0.2497202754020691, + "eval_runtime": 13.4861, + "eval_samples_per_second": 63.473, + "eval_steps_per_second": 15.868, + "num_input_tokens_seen": 14490240, + "step": 19000 + }, + { + "epoch": 39.51143451143451, + "grad_norm": 0.0005644356133416295, + "learning_rate": 0.16172188515660096, + "loss": 0.3014, + "num_input_tokens_seen": 14494176, + "step": 19005 + }, + { + "epoch": 39.521829521829524, + "grad_norm": 8.414949115831405e-05, + "learning_rate": 0.16166315952702878, + "loss": 0.2902, + "num_input_tokens_seen": 14497984, + "step": 19010 + }, + { + "epoch": 39.53222453222453, + "grad_norm": 0.00014315721637103707, + "learning_rate": 0.16160443209885084, + "loss": 0.28, + "num_input_tokens_seen": 14501792, + "step": 19015 + }, + { + "epoch": 39.54261954261954, + "grad_norm": 0.0004750539956148714, + "learning_rate": 0.16154570288112363, + "loss": 0.2801, + "num_input_tokens_seen": 14505600, + "step": 19020 + }, + { + "epoch": 39.553014553014556, + "grad_norm": 0.00038815807783976197, + "learning_rate": 0.16148697188290395, + "loss": 0.2677, + "num_input_tokens_seen": 14509376, + "step": 19025 + }, + { + "epoch": 39.563409563409564, + "grad_norm": 0.00015546979557257146, + "learning_rate": 0.16142823911324888, + "loss": 0.2516, + "num_input_tokens_seen": 14513248, + "step": 19030 + }, + { + "epoch": 39.57380457380457, + "grad_norm": 0.00020375453459564596, + "learning_rate": 0.16136950458121568, + "loss": 0.2514, + "num_input_tokens_seen": 14516864, + "step": 19035 + }, + { + "epoch": 39.58419958419958, + "grad_norm": 0.0001250451459782198, + "learning_rate": 0.16131076829586205, + "loss": 0.2462, + "num_input_tokens_seen": 14520736, + "step": 19040 + }, + { + "epoch": 39.5945945945946, + "grad_norm": 0.0005256778094917536, + "learning_rate": 0.1612520302662457, + "loss": 0.2896, + "num_input_tokens_seen": 14524480, + "step": 19045 + }, + { + "epoch": 39.604989604989605, + "grad_norm": 0.00018821176490746439, + "learning_rate": 0.16119329050142497, + "loss": 0.2544, + "num_input_tokens_seen": 14528032, + "step": 19050 + }, + { + "epoch": 39.61538461538461, + "grad_norm": 0.0003220030921511352, + "learning_rate": 0.16113454901045818, + "loss": 0.2605, + "num_input_tokens_seen": 14531904, + "step": 19055 + }, + { + "epoch": 39.62577962577963, + "grad_norm": 0.0002236302534583956, + "learning_rate": 0.16107580580240397, + "loss": 0.2516, + "num_input_tokens_seen": 14535648, + "step": 19060 + }, + { + "epoch": 39.63617463617464, + "grad_norm": 0.0001901189680211246, + "learning_rate": 0.16101706088632134, + "loss": 0.2267, + "num_input_tokens_seen": 14539520, + "step": 19065 + }, + { + "epoch": 39.646569646569645, + "grad_norm": 0.00013239476538728923, + "learning_rate": 0.16095831427126947, + "loss": 0.268, + "num_input_tokens_seen": 14543328, + "step": 19070 + }, + { + "epoch": 39.656964656964654, + "grad_norm": 0.0001788314402801916, + "learning_rate": 0.16089956596630783, + "loss": 0.2235, + "num_input_tokens_seen": 14547200, + "step": 19075 + }, + { + "epoch": 39.66735966735967, + "grad_norm": 0.0001133709738496691, + "learning_rate": 0.16084081598049618, + "loss": 0.2496, + "num_input_tokens_seen": 14550880, + "step": 19080 + }, + { + "epoch": 39.67775467775468, + "grad_norm": 0.00023295370920095593, + "learning_rate": 0.1607820643228944, + "loss": 0.2882, + "num_input_tokens_seen": 14554720, + "step": 19085 + }, + { + "epoch": 39.688149688149686, + "grad_norm": 0.0003464298788458109, + "learning_rate": 0.16072331100256285, + "loss": 0.2773, + "num_input_tokens_seen": 14558560, + "step": 19090 + }, + { + "epoch": 39.6985446985447, + "grad_norm": 0.00013704427692573518, + "learning_rate": 0.16066455602856197, + "loss": 0.2767, + "num_input_tokens_seen": 14562336, + "step": 19095 + }, + { + "epoch": 39.70893970893971, + "grad_norm": 0.0004995701019652188, + "learning_rate": 0.16060579940995257, + "loss": 0.2791, + "num_input_tokens_seen": 14566208, + "step": 19100 + }, + { + "epoch": 39.71933471933472, + "grad_norm": 0.00016199800302274525, + "learning_rate": 0.16054704115579557, + "loss": 0.2679, + "num_input_tokens_seen": 14569984, + "step": 19105 + }, + { + "epoch": 39.729729729729726, + "grad_norm": 0.00030695617897436023, + "learning_rate": 0.1604882812751523, + "loss": 0.2535, + "num_input_tokens_seen": 14573856, + "step": 19110 + }, + { + "epoch": 39.74012474012474, + "grad_norm": 0.0001370328536722809, + "learning_rate": 0.16042951977708425, + "loss": 0.2657, + "num_input_tokens_seen": 14577696, + "step": 19115 + }, + { + "epoch": 39.75051975051975, + "grad_norm": 8.185114711523056e-05, + "learning_rate": 0.16037075667065318, + "loss": 0.2793, + "num_input_tokens_seen": 14581536, + "step": 19120 + }, + { + "epoch": 39.76091476091476, + "grad_norm": 7.148168515414e-05, + "learning_rate": 0.1603119919649211, + "loss": 0.2797, + "num_input_tokens_seen": 14585280, + "step": 19125 + }, + { + "epoch": 39.771309771309774, + "grad_norm": 0.0004562731191981584, + "learning_rate": 0.16025322566895028, + "loss": 0.2664, + "num_input_tokens_seen": 14589088, + "step": 19130 + }, + { + "epoch": 39.78170478170478, + "grad_norm": 3.264590122853406e-05, + "learning_rate": 0.16019445779180322, + "loss": 0.253, + "num_input_tokens_seen": 14592896, + "step": 19135 + }, + { + "epoch": 39.79209979209979, + "grad_norm": 0.00019897542370017618, + "learning_rate": 0.16013568834254271, + "loss": 0.2898, + "num_input_tokens_seen": 14596576, + "step": 19140 + }, + { + "epoch": 39.802494802494806, + "grad_norm": 5.010885070078075e-05, + "learning_rate": 0.1600769173302316, + "loss": 0.2262, + "num_input_tokens_seen": 14600608, + "step": 19145 + }, + { + "epoch": 39.812889812889814, + "grad_norm": 0.00017095061775762588, + "learning_rate": 0.16001814476393322, + "loss": 0.2563, + "num_input_tokens_seen": 14604576, + "step": 19150 + }, + { + "epoch": 39.82328482328482, + "grad_norm": 7.988956349436194e-05, + "learning_rate": 0.15995937065271104, + "loss": 0.2814, + "num_input_tokens_seen": 14608448, + "step": 19155 + }, + { + "epoch": 39.83367983367983, + "grad_norm": 0.00030251420685090125, + "learning_rate": 0.15990059500562873, + "loss": 0.2508, + "num_input_tokens_seen": 14612256, + "step": 19160 + }, + { + "epoch": 39.84407484407485, + "grad_norm": 0.0003945770440623164, + "learning_rate": 0.15984181783175025, + "loss": 0.2511, + "num_input_tokens_seen": 14616032, + "step": 19165 + }, + { + "epoch": 39.854469854469855, + "grad_norm": 0.0002816040941979736, + "learning_rate": 0.1597830391401398, + "loss": 0.2826, + "num_input_tokens_seen": 14619968, + "step": 19170 + }, + { + "epoch": 39.86486486486486, + "grad_norm": 0.00021585570357274264, + "learning_rate": 0.15972425893986178, + "loss": 0.2756, + "num_input_tokens_seen": 14623872, + "step": 19175 + }, + { + "epoch": 39.87525987525988, + "grad_norm": 0.00045787045382894576, + "learning_rate": 0.15966547723998084, + "loss": 0.2725, + "num_input_tokens_seen": 14627680, + "step": 19180 + }, + { + "epoch": 39.88565488565489, + "grad_norm": 0.00048087825416587293, + "learning_rate": 0.15960669404956176, + "loss": 0.2785, + "num_input_tokens_seen": 14631520, + "step": 19185 + }, + { + "epoch": 39.896049896049895, + "grad_norm": 7.098748756106943e-05, + "learning_rate": 0.1595479093776698, + "loss": 0.2774, + "num_input_tokens_seen": 14635264, + "step": 19190 + }, + { + "epoch": 39.906444906444904, + "grad_norm": 0.00028822338208556175, + "learning_rate": 0.15948912323337022, + "loss": 0.2638, + "num_input_tokens_seen": 14639136, + "step": 19195 + }, + { + "epoch": 39.91683991683992, + "grad_norm": 0.000355742871761322, + "learning_rate": 0.1594303356257286, + "loss": 0.2788, + "num_input_tokens_seen": 14643072, + "step": 19200 + }, + { + "epoch": 39.91683991683992, + "eval_loss": 0.24827949702739716, + "eval_runtime": 13.4098, + "eval_samples_per_second": 63.834, + "eval_steps_per_second": 15.958, + "num_input_tokens_seen": 14643072, + "step": 19200 + }, + { + "epoch": 39.92723492723493, + "grad_norm": 0.00035336625296622515, + "learning_rate": 0.15937154656381072, + "loss": 0.2656, + "num_input_tokens_seen": 14646944, + "step": 19205 + }, + { + "epoch": 39.937629937629936, + "grad_norm": 0.0004528326098807156, + "learning_rate": 0.15931275605668258, + "loss": 0.2557, + "num_input_tokens_seen": 14650784, + "step": 19210 + }, + { + "epoch": 39.94802494802495, + "grad_norm": 7.95606174506247e-05, + "learning_rate": 0.1592539641134104, + "loss": 0.2628, + "num_input_tokens_seen": 14654688, + "step": 19215 + }, + { + "epoch": 39.95841995841996, + "grad_norm": 0.00010925047536147758, + "learning_rate": 0.1591951707430607, + "loss": 0.2683, + "num_input_tokens_seen": 14658432, + "step": 19220 + }, + { + "epoch": 39.96881496881497, + "grad_norm": 0.00040998036274686456, + "learning_rate": 0.15913637595470007, + "loss": 0.2549, + "num_input_tokens_seen": 14662144, + "step": 19225 + }, + { + "epoch": 39.979209979209976, + "grad_norm": 0.00030653062276542187, + "learning_rate": 0.15907757975739548, + "loss": 0.2563, + "num_input_tokens_seen": 14666080, + "step": 19230 + }, + { + "epoch": 39.98960498960499, + "grad_norm": 0.00036579446168616414, + "learning_rate": 0.159018782160214, + "loss": 0.2465, + "num_input_tokens_seen": 14669920, + "step": 19235 + }, + { + "epoch": 40.0, + "grad_norm": 0.0003128277894575149, + "learning_rate": 0.158959983172223, + "loss": 0.2569, + "num_input_tokens_seen": 14673744, + "step": 19240 + }, + { + "epoch": 40.01039501039501, + "grad_norm": 0.0006778359529562294, + "learning_rate": 0.15890118280249, + "loss": 0.2748, + "num_input_tokens_seen": 14677424, + "step": 19245 + }, + { + "epoch": 40.020790020790024, + "grad_norm": 0.00021710197324864566, + "learning_rate": 0.15884238106008275, + "loss": 0.2487, + "num_input_tokens_seen": 14681200, + "step": 19250 + }, + { + "epoch": 40.03118503118503, + "grad_norm": 0.00010688927432056516, + "learning_rate": 0.15878357795406922, + "loss": 0.26, + "num_input_tokens_seen": 14684976, + "step": 19255 + }, + { + "epoch": 40.04158004158004, + "grad_norm": 0.0005350644350983202, + "learning_rate": 0.15872477349351757, + "loss": 0.2642, + "num_input_tokens_seen": 14688720, + "step": 19260 + }, + { + "epoch": 40.05197505197505, + "grad_norm": 0.00046306231524795294, + "learning_rate": 0.15866596768749622, + "loss": 0.2921, + "num_input_tokens_seen": 14692368, + "step": 19265 + }, + { + "epoch": 40.062370062370064, + "grad_norm": 0.0008180579752661288, + "learning_rate": 0.15860716054507373, + "loss": 0.2617, + "num_input_tokens_seen": 14696080, + "step": 19270 + }, + { + "epoch": 40.07276507276507, + "grad_norm": 0.00015625995001755655, + "learning_rate": 0.1585483520753189, + "loss": 0.2713, + "num_input_tokens_seen": 14699760, + "step": 19275 + }, + { + "epoch": 40.08316008316008, + "grad_norm": 0.0004341186722740531, + "learning_rate": 0.1584895422873008, + "loss": 0.2685, + "num_input_tokens_seen": 14703696, + "step": 19280 + }, + { + "epoch": 40.093555093555096, + "grad_norm": 0.0003843106678687036, + "learning_rate": 0.1584307311900886, + "loss": 0.2735, + "num_input_tokens_seen": 14707504, + "step": 19285 + }, + { + "epoch": 40.103950103950105, + "grad_norm": 0.0003282174584455788, + "learning_rate": 0.1583719187927517, + "loss": 0.2706, + "num_input_tokens_seen": 14711312, + "step": 19290 + }, + { + "epoch": 40.11434511434511, + "grad_norm": 0.0004849833494517952, + "learning_rate": 0.15831310510435967, + "loss": 0.2742, + "num_input_tokens_seen": 14715056, + "step": 19295 + }, + { + "epoch": 40.12474012474012, + "grad_norm": 0.00028795923572033644, + "learning_rate": 0.15825429013398243, + "loss": 0.2684, + "num_input_tokens_seen": 14718992, + "step": 19300 + }, + { + "epoch": 40.13513513513514, + "grad_norm": 0.0003188580449204892, + "learning_rate": 0.15819547389068986, + "loss": 0.274, + "num_input_tokens_seen": 14722672, + "step": 19305 + }, + { + "epoch": 40.145530145530145, + "grad_norm": 0.00022718128457199782, + "learning_rate": 0.1581366563835522, + "loss": 0.2725, + "num_input_tokens_seen": 14726416, + "step": 19310 + }, + { + "epoch": 40.15592515592515, + "grad_norm": 0.00019647617591544986, + "learning_rate": 0.15807783762163993, + "loss": 0.282, + "num_input_tokens_seen": 14730256, + "step": 19315 + }, + { + "epoch": 40.16632016632017, + "grad_norm": 0.0002832565223798156, + "learning_rate": 0.15801901761402365, + "loss": 0.2766, + "num_input_tokens_seen": 14734096, + "step": 19320 + }, + { + "epoch": 40.17671517671518, + "grad_norm": 0.00012157320452388376, + "learning_rate": 0.157960196369774, + "loss": 0.2716, + "num_input_tokens_seen": 14737872, + "step": 19325 + }, + { + "epoch": 40.187110187110186, + "grad_norm": 0.00032345092040486634, + "learning_rate": 0.157901373897962, + "loss": 0.268, + "num_input_tokens_seen": 14741616, + "step": 19330 + }, + { + "epoch": 40.197505197505194, + "grad_norm": 0.00012614870502147824, + "learning_rate": 0.15784255020765892, + "loss": 0.2539, + "num_input_tokens_seen": 14745456, + "step": 19335 + }, + { + "epoch": 40.20790020790021, + "grad_norm": 4.1194787627318874e-05, + "learning_rate": 0.157783725307936, + "loss": 0.2515, + "num_input_tokens_seen": 14749200, + "step": 19340 + }, + { + "epoch": 40.21829521829522, + "grad_norm": 0.00013640409451909363, + "learning_rate": 0.15772489920786484, + "loss": 0.2936, + "num_input_tokens_seen": 14753040, + "step": 19345 + }, + { + "epoch": 40.228690228690226, + "grad_norm": 0.0002855884376913309, + "learning_rate": 0.15766607191651713, + "loss": 0.2599, + "num_input_tokens_seen": 14757072, + "step": 19350 + }, + { + "epoch": 40.23908523908524, + "grad_norm": 0.0003141595225315541, + "learning_rate": 0.1576072434429648, + "loss": 0.258, + "num_input_tokens_seen": 14760720, + "step": 19355 + }, + { + "epoch": 40.24948024948025, + "grad_norm": 6.51771406410262e-05, + "learning_rate": 0.15754841379627998, + "loss": 0.2781, + "num_input_tokens_seen": 14764656, + "step": 19360 + }, + { + "epoch": 40.25987525987526, + "grad_norm": 0.00017045206914190203, + "learning_rate": 0.15748958298553484, + "loss": 0.2669, + "num_input_tokens_seen": 14768464, + "step": 19365 + }, + { + "epoch": 40.270270270270274, + "grad_norm": 4.464116500457749e-05, + "learning_rate": 0.1574307510198019, + "loss": 0.2639, + "num_input_tokens_seen": 14772272, + "step": 19370 + }, + { + "epoch": 40.28066528066528, + "grad_norm": 0.00023470954329241067, + "learning_rate": 0.15737191790815375, + "loss": 0.2554, + "num_input_tokens_seen": 14776080, + "step": 19375 + }, + { + "epoch": 40.29106029106029, + "grad_norm": 6.293384649325162e-05, + "learning_rate": 0.15731308365966323, + "loss": 0.2256, + "num_input_tokens_seen": 14779760, + "step": 19380 + }, + { + "epoch": 40.3014553014553, + "grad_norm": 0.00042729778215289116, + "learning_rate": 0.15725424828340331, + "loss": 0.2532, + "num_input_tokens_seen": 14783760, + "step": 19385 + }, + { + "epoch": 40.311850311850314, + "grad_norm": 3.17596313834656e-05, + "learning_rate": 0.15719541178844715, + "loss": 0.3026, + "num_input_tokens_seen": 14787664, + "step": 19390 + }, + { + "epoch": 40.32224532224532, + "grad_norm": 0.00015527821960859, + "learning_rate": 0.15713657418386806, + "loss": 0.286, + "num_input_tokens_seen": 14791312, + "step": 19395 + }, + { + "epoch": 40.33264033264033, + "grad_norm": 0.00028506197850219905, + "learning_rate": 0.15707773547873957, + "loss": 0.2716, + "num_input_tokens_seen": 14795184, + "step": 19400 + }, + { + "epoch": 40.33264033264033, + "eval_loss": 0.25817909836769104, + "eval_runtime": 13.4178, + "eval_samples_per_second": 63.796, + "eval_steps_per_second": 15.949, + "num_input_tokens_seen": 14795184, + "step": 19400 + }, + { + "epoch": 40.343035343035346, + "grad_norm": 0.00026690305094234645, + "learning_rate": 0.1570188956821353, + "loss": 0.2664, + "num_input_tokens_seen": 14798800, + "step": 19405 + }, + { + "epoch": 40.353430353430355, + "grad_norm": 0.00019005748617928475, + "learning_rate": 0.1569600548031291, + "loss": 0.2697, + "num_input_tokens_seen": 14802768, + "step": 19410 + }, + { + "epoch": 40.36382536382536, + "grad_norm": 0.00023546448210254312, + "learning_rate": 0.156901212850795, + "loss": 0.2922, + "num_input_tokens_seen": 14806736, + "step": 19415 + }, + { + "epoch": 40.37422037422037, + "grad_norm": 0.0002392926689935848, + "learning_rate": 0.15684236983420716, + "loss": 0.2601, + "num_input_tokens_seen": 14810512, + "step": 19420 + }, + { + "epoch": 40.38461538461539, + "grad_norm": 0.00047618846292607486, + "learning_rate": 0.1567835257624399, + "loss": 0.2573, + "num_input_tokens_seen": 14814320, + "step": 19425 + }, + { + "epoch": 40.395010395010395, + "grad_norm": 0.0002257501328131184, + "learning_rate": 0.1567246806445677, + "loss": 0.2687, + "num_input_tokens_seen": 14818224, + "step": 19430 + }, + { + "epoch": 40.4054054054054, + "grad_norm": 0.0001277759438380599, + "learning_rate": 0.15666583448966526, + "loss": 0.2598, + "num_input_tokens_seen": 14821936, + "step": 19435 + }, + { + "epoch": 40.41580041580042, + "grad_norm": 9.289153967984021e-05, + "learning_rate": 0.1566069873068074, + "loss": 0.2615, + "num_input_tokens_seen": 14825712, + "step": 19440 + }, + { + "epoch": 40.42619542619543, + "grad_norm": 0.0002395427436567843, + "learning_rate": 0.156548139105069, + "loss": 0.2575, + "num_input_tokens_seen": 14829424, + "step": 19445 + }, + { + "epoch": 40.436590436590436, + "grad_norm": 0.0002534229715820402, + "learning_rate": 0.15648928989352529, + "loss": 0.258, + "num_input_tokens_seen": 14833232, + "step": 19450 + }, + { + "epoch": 40.446985446985444, + "grad_norm": 0.00021143726189620793, + "learning_rate": 0.15643043968125156, + "loss": 0.2726, + "num_input_tokens_seen": 14837040, + "step": 19455 + }, + { + "epoch": 40.45738045738046, + "grad_norm": 0.0006262267706915736, + "learning_rate": 0.15637158847732316, + "loss": 0.2856, + "num_input_tokens_seen": 14840880, + "step": 19460 + }, + { + "epoch": 40.46777546777547, + "grad_norm": 0.000190770675544627, + "learning_rate": 0.15631273629081582, + "loss": 0.2828, + "num_input_tokens_seen": 14844688, + "step": 19465 + }, + { + "epoch": 40.478170478170476, + "grad_norm": 0.00032719149021431804, + "learning_rate": 0.15625388313080518, + "loss": 0.286, + "num_input_tokens_seen": 14848464, + "step": 19470 + }, + { + "epoch": 40.48856548856549, + "grad_norm": 0.0003644162497948855, + "learning_rate": 0.15619502900636714, + "loss": 0.2747, + "num_input_tokens_seen": 14852112, + "step": 19475 + }, + { + "epoch": 40.4989604989605, + "grad_norm": 0.00034886348294094205, + "learning_rate": 0.15613617392657783, + "loss": 0.2548, + "num_input_tokens_seen": 14855952, + "step": 19480 + }, + { + "epoch": 40.50935550935551, + "grad_norm": 0.00010568360448814929, + "learning_rate": 0.15607731790051335, + "loss": 0.2699, + "num_input_tokens_seen": 14859792, + "step": 19485 + }, + { + "epoch": 40.51975051975052, + "grad_norm": 0.0002857714134734124, + "learning_rate": 0.15601846093725008, + "loss": 0.2586, + "num_input_tokens_seen": 14863632, + "step": 19490 + }, + { + "epoch": 40.53014553014553, + "grad_norm": 0.00010057652980322018, + "learning_rate": 0.1559596030458645, + "loss": 0.2436, + "num_input_tokens_seen": 14867376, + "step": 19495 + }, + { + "epoch": 40.54054054054054, + "grad_norm": 0.00026237242855131626, + "learning_rate": 0.1559007442354333, + "loss": 0.2637, + "num_input_tokens_seen": 14871152, + "step": 19500 + }, + { + "epoch": 40.55093555093555, + "grad_norm": 0.00014281108451541513, + "learning_rate": 0.15584188451503314, + "loss": 0.259, + "num_input_tokens_seen": 14875088, + "step": 19505 + }, + { + "epoch": 40.561330561330564, + "grad_norm": 0.00023837907065171748, + "learning_rate": 0.15578302389374094, + "loss": 0.2035, + "num_input_tokens_seen": 14878864, + "step": 19510 + }, + { + "epoch": 40.57172557172557, + "grad_norm": 0.0005240822792984545, + "learning_rate": 0.1557241623806338, + "loss": 0.2166, + "num_input_tokens_seen": 14882704, + "step": 19515 + }, + { + "epoch": 40.58212058212058, + "grad_norm": 8.65408728714101e-05, + "learning_rate": 0.15566529998478887, + "loss": 0.3236, + "num_input_tokens_seen": 14886608, + "step": 19520 + }, + { + "epoch": 40.59251559251559, + "grad_norm": 0.00020667626813519746, + "learning_rate": 0.15560643671528354, + "loss": 0.2797, + "num_input_tokens_seen": 14890544, + "step": 19525 + }, + { + "epoch": 40.602910602910605, + "grad_norm": 9.813201177166775e-05, + "learning_rate": 0.15554757258119514, + "loss": 0.2688, + "num_input_tokens_seen": 14894448, + "step": 19530 + }, + { + "epoch": 40.61330561330561, + "grad_norm": 0.00023076613433659077, + "learning_rate": 0.1554887075916014, + "loss": 0.2781, + "num_input_tokens_seen": 14898128, + "step": 19535 + }, + { + "epoch": 40.62370062370062, + "grad_norm": 0.00041677054832689464, + "learning_rate": 0.15542984175558, + "loss": 0.2725, + "num_input_tokens_seen": 14901872, + "step": 19540 + }, + { + "epoch": 40.63409563409564, + "grad_norm": 7.777262362651527e-05, + "learning_rate": 0.1553709750822087, + "loss": 0.2738, + "num_input_tokens_seen": 14905520, + "step": 19545 + }, + { + "epoch": 40.644490644490645, + "grad_norm": 0.0003350620099809021, + "learning_rate": 0.15531210758056554, + "loss": 0.2908, + "num_input_tokens_seen": 14909424, + "step": 19550 + }, + { + "epoch": 40.65488565488565, + "grad_norm": 0.00021306442795321345, + "learning_rate": 0.15525323925972867, + "loss": 0.2305, + "num_input_tokens_seen": 14913232, + "step": 19555 + }, + { + "epoch": 40.66528066528066, + "grad_norm": 0.0005110530182719231, + "learning_rate": 0.15519437012877627, + "loss": 0.276, + "num_input_tokens_seen": 14917008, + "step": 19560 + }, + { + "epoch": 40.67567567567568, + "grad_norm": 0.00038366674561984837, + "learning_rate": 0.15513550019678676, + "loss": 0.2922, + "num_input_tokens_seen": 14920784, + "step": 19565 + }, + { + "epoch": 40.686070686070686, + "grad_norm": 0.00013617715740110725, + "learning_rate": 0.15507662947283854, + "loss": 0.2634, + "num_input_tokens_seen": 14924464, + "step": 19570 + }, + { + "epoch": 40.696465696465694, + "grad_norm": 8.178168354788795e-05, + "learning_rate": 0.15501775796601028, + "loss": 0.2756, + "num_input_tokens_seen": 14928176, + "step": 19575 + }, + { + "epoch": 40.70686070686071, + "grad_norm": 0.00022606848506256938, + "learning_rate": 0.15495888568538066, + "loss": 0.2737, + "num_input_tokens_seen": 14932048, + "step": 19580 + }, + { + "epoch": 40.71725571725572, + "grad_norm": 8.381425868719816e-05, + "learning_rate": 0.1549000126400286, + "loss": 0.2778, + "num_input_tokens_seen": 14935728, + "step": 19585 + }, + { + "epoch": 40.727650727650726, + "grad_norm": 0.0003244522085878998, + "learning_rate": 0.15484113883903294, + "loss": 0.2629, + "num_input_tokens_seen": 14939760, + "step": 19590 + }, + { + "epoch": 40.73804573804574, + "grad_norm": 0.00022940777125768363, + "learning_rate": 0.15478226429147288, + "loss": 0.2587, + "num_input_tokens_seen": 14943568, + "step": 19595 + }, + { + "epoch": 40.74844074844075, + "grad_norm": 0.00017733285494614393, + "learning_rate": 0.15472338900642757, + "loss": 0.2627, + "num_input_tokens_seen": 14947312, + "step": 19600 + }, + { + "epoch": 40.74844074844075, + "eval_loss": 0.24833182990550995, + "eval_runtime": 13.4689, + "eval_samples_per_second": 63.554, + "eval_steps_per_second": 15.888, + "num_input_tokens_seen": 14947312, + "step": 19600 + }, + { + "epoch": 40.75883575883576, + "grad_norm": 0.000331509712850675, + "learning_rate": 0.15466451299297632, + "loss": 0.2446, + "num_input_tokens_seen": 14951152, + "step": 19605 + }, + { + "epoch": 40.76923076923077, + "grad_norm": 0.00031276073423214257, + "learning_rate": 0.15460563626019852, + "loss": 0.2699, + "num_input_tokens_seen": 14955152, + "step": 19610 + }, + { + "epoch": 40.77962577962578, + "grad_norm": 6.963621854083613e-05, + "learning_rate": 0.15454675881717375, + "loss": 0.2641, + "num_input_tokens_seen": 14959088, + "step": 19615 + }, + { + "epoch": 40.79002079002079, + "grad_norm": 0.00032989951432682574, + "learning_rate": 0.1544878806729816, + "loss": 0.2574, + "num_input_tokens_seen": 14963056, + "step": 19620 + }, + { + "epoch": 40.8004158004158, + "grad_norm": 0.00017532636411488056, + "learning_rate": 0.1544290018367019, + "loss": 0.2553, + "num_input_tokens_seen": 14966832, + "step": 19625 + }, + { + "epoch": 40.810810810810814, + "grad_norm": 6.198434857651591e-05, + "learning_rate": 0.15437012231741445, + "loss": 0.2727, + "num_input_tokens_seen": 14970768, + "step": 19630 + }, + { + "epoch": 40.82120582120582, + "grad_norm": 0.0003781394334509969, + "learning_rate": 0.1543112421241992, + "loss": 0.2693, + "num_input_tokens_seen": 14974512, + "step": 19635 + }, + { + "epoch": 40.83160083160083, + "grad_norm": 0.00020204659085720778, + "learning_rate": 0.15425236126613626, + "loss": 0.2649, + "num_input_tokens_seen": 14978224, + "step": 19640 + }, + { + "epoch": 40.84199584199584, + "grad_norm": 0.0001514485338702798, + "learning_rate": 0.15419347975230577, + "loss": 0.2632, + "num_input_tokens_seen": 14982096, + "step": 19645 + }, + { + "epoch": 40.852390852390855, + "grad_norm": 8.541520946891978e-05, + "learning_rate": 0.154134597591788, + "loss": 0.2685, + "num_input_tokens_seen": 14985872, + "step": 19650 + }, + { + "epoch": 40.86278586278586, + "grad_norm": 0.00046932362602092326, + "learning_rate": 0.1540757147936633, + "loss": 0.2742, + "num_input_tokens_seen": 14989680, + "step": 19655 + }, + { + "epoch": 40.87318087318087, + "grad_norm": 0.00019683956634253263, + "learning_rate": 0.1540168313670122, + "loss": 0.2632, + "num_input_tokens_seen": 14993520, + "step": 19660 + }, + { + "epoch": 40.88357588357589, + "grad_norm": 0.00010083048982778564, + "learning_rate": 0.1539579473209152, + "loss": 0.2537, + "num_input_tokens_seen": 14997232, + "step": 19665 + }, + { + "epoch": 40.893970893970895, + "grad_norm": 0.00039551942609250546, + "learning_rate": 0.15389906266445294, + "loss": 0.2311, + "num_input_tokens_seen": 15000848, + "step": 19670 + }, + { + "epoch": 40.9043659043659, + "grad_norm": 0.00023763089848216623, + "learning_rate": 0.15384017740670627, + "loss": 0.2742, + "num_input_tokens_seen": 15004720, + "step": 19675 + }, + { + "epoch": 40.91476091476091, + "grad_norm": 0.00028581643709912896, + "learning_rate": 0.15378129155675602, + "loss": 0.2701, + "num_input_tokens_seen": 15008432, + "step": 19680 + }, + { + "epoch": 40.92515592515593, + "grad_norm": 0.000126958591863513, + "learning_rate": 0.15372240512368307, + "loss": 0.269, + "num_input_tokens_seen": 15012144, + "step": 19685 + }, + { + "epoch": 40.935550935550935, + "grad_norm": 0.0002582821762189269, + "learning_rate": 0.1536635181165684, + "loss": 0.2742, + "num_input_tokens_seen": 15015920, + "step": 19690 + }, + { + "epoch": 40.945945945945944, + "grad_norm": 0.0004045071837026626, + "learning_rate": 0.15360463054449328, + "loss": 0.2439, + "num_input_tokens_seen": 15019824, + "step": 19695 + }, + { + "epoch": 40.95634095634096, + "grad_norm": 0.00011655008711386472, + "learning_rate": 0.1535457424165388, + "loss": 0.258, + "num_input_tokens_seen": 15023824, + "step": 19700 + }, + { + "epoch": 40.96673596673597, + "grad_norm": 0.0004144597623962909, + "learning_rate": 0.15348685374178628, + "loss": 0.2972, + "num_input_tokens_seen": 15027632, + "step": 19705 + }, + { + "epoch": 40.977130977130976, + "grad_norm": 0.00020229780056979507, + "learning_rate": 0.1534279645293171, + "loss": 0.2745, + "num_input_tokens_seen": 15031504, + "step": 19710 + }, + { + "epoch": 40.987525987525984, + "grad_norm": 0.00015313579933717847, + "learning_rate": 0.1533690747882127, + "loss": 0.2766, + "num_input_tokens_seen": 15035312, + "step": 19715 + }, + { + "epoch": 40.997920997921, + "grad_norm": 0.00034366751788184047, + "learning_rate": 0.15331018452755465, + "loss": 0.2641, + "num_input_tokens_seen": 15039216, + "step": 19720 + }, + { + "epoch": 41.00831600831601, + "grad_norm": 0.00024178162857424468, + "learning_rate": 0.15325129375642457, + "loss": 0.2574, + "num_input_tokens_seen": 15042960, + "step": 19725 + }, + { + "epoch": 41.018711018711016, + "grad_norm": 0.0002120835124514997, + "learning_rate": 0.15319240248390406, + "loss": 0.2499, + "num_input_tokens_seen": 15046704, + "step": 19730 + }, + { + "epoch": 41.02910602910603, + "grad_norm": 0.00015668269770685583, + "learning_rate": 0.153133510719075, + "loss": 0.2894, + "num_input_tokens_seen": 15050608, + "step": 19735 + }, + { + "epoch": 41.03950103950104, + "grad_norm": 0.00043070336687378585, + "learning_rate": 0.15307461847101922, + "loss": 0.2509, + "num_input_tokens_seen": 15054320, + "step": 19740 + }, + { + "epoch": 41.04989604989605, + "grad_norm": 0.00043466646457090974, + "learning_rate": 0.15301572574881864, + "loss": 0.2546, + "num_input_tokens_seen": 15058128, + "step": 19745 + }, + { + "epoch": 41.06029106029106, + "grad_norm": 5.977210821583867e-05, + "learning_rate": 0.15295683256155523, + "loss": 0.2372, + "num_input_tokens_seen": 15061968, + "step": 19750 + }, + { + "epoch": 41.07068607068607, + "grad_norm": 0.0004895455203950405, + "learning_rate": 0.15289793891831113, + "loss": 0.2737, + "num_input_tokens_seen": 15065648, + "step": 19755 + }, + { + "epoch": 41.08108108108108, + "grad_norm": 0.00026473545585758984, + "learning_rate": 0.15283904482816837, + "loss": 0.2759, + "num_input_tokens_seen": 15069424, + "step": 19760 + }, + { + "epoch": 41.09147609147609, + "grad_norm": 0.0005848917062394321, + "learning_rate": 0.15278015030020928, + "loss": 0.221, + "num_input_tokens_seen": 15073168, + "step": 19765 + }, + { + "epoch": 41.101871101871104, + "grad_norm": 2.508679244783707e-05, + "learning_rate": 0.152721255343516, + "loss": 0.2508, + "num_input_tokens_seen": 15076976, + "step": 19770 + }, + { + "epoch": 41.11226611226611, + "grad_norm": 0.00019580198568291962, + "learning_rate": 0.15266235996717098, + "loss": 0.2633, + "num_input_tokens_seen": 15080912, + "step": 19775 + }, + { + "epoch": 41.12266112266112, + "grad_norm": 0.00043932415428571403, + "learning_rate": 0.15260346418025664, + "loss": 0.2614, + "num_input_tokens_seen": 15084944, + "step": 19780 + }, + { + "epoch": 41.13305613305613, + "grad_norm": 0.0004289295757189393, + "learning_rate": 0.15254456799185537, + "loss": 0.2495, + "num_input_tokens_seen": 15088816, + "step": 19785 + }, + { + "epoch": 41.143451143451145, + "grad_norm": 0.00013837787264492363, + "learning_rate": 0.15248567141104974, + "loss": 0.2657, + "num_input_tokens_seen": 15092688, + "step": 19790 + }, + { + "epoch": 41.15384615384615, + "grad_norm": 0.00011153219384141266, + "learning_rate": 0.15242677444692232, + "loss": 0.2959, + "num_input_tokens_seen": 15096656, + "step": 19795 + }, + { + "epoch": 41.16424116424116, + "grad_norm": 0.0001644265721552074, + "learning_rate": 0.15236787710855584, + "loss": 0.2659, + "num_input_tokens_seen": 15100336, + "step": 19800 + }, + { + "epoch": 41.16424116424116, + "eval_loss": 0.2633832097053528, + "eval_runtime": 13.3984, + "eval_samples_per_second": 63.888, + "eval_steps_per_second": 15.972, + "num_input_tokens_seen": 15100336, + "step": 19800 + }, + { + "epoch": 41.17463617463618, + "grad_norm": 0.0004443080979399383, + "learning_rate": 0.1523089794050329, + "loss": 0.2816, + "num_input_tokens_seen": 15104208, + "step": 19805 + }, + { + "epoch": 41.185031185031185, + "grad_norm": 0.00034143132506869733, + "learning_rate": 0.15225008134543633, + "loss": 0.2514, + "num_input_tokens_seen": 15107920, + "step": 19810 + }, + { + "epoch": 41.195426195426194, + "grad_norm": 0.00013911582936998457, + "learning_rate": 0.15219118293884895, + "loss": 0.274, + "num_input_tokens_seen": 15111760, + "step": 19815 + }, + { + "epoch": 41.20582120582121, + "grad_norm": 0.0008609531214460731, + "learning_rate": 0.15213228419435362, + "loss": 0.2805, + "num_input_tokens_seen": 15115440, + "step": 19820 + }, + { + "epoch": 41.21621621621622, + "grad_norm": 7.07189756212756e-05, + "learning_rate": 0.15207338512103327, + "loss": 0.2657, + "num_input_tokens_seen": 15119280, + "step": 19825 + }, + { + "epoch": 41.226611226611226, + "grad_norm": 0.00023028695432003587, + "learning_rate": 0.1520144857279709, + "loss": 0.26, + "num_input_tokens_seen": 15123088, + "step": 19830 + }, + { + "epoch": 41.237006237006234, + "grad_norm": 0.00030570439412258565, + "learning_rate": 0.1519555860242495, + "loss": 0.2575, + "num_input_tokens_seen": 15126832, + "step": 19835 + }, + { + "epoch": 41.24740124740125, + "grad_norm": 5.598764983005822e-05, + "learning_rate": 0.15189668601895218, + "loss": 0.257, + "num_input_tokens_seen": 15130704, + "step": 19840 + }, + { + "epoch": 41.25779625779626, + "grad_norm": 7.639340765308589e-05, + "learning_rate": 0.151837785721162, + "loss": 0.2696, + "num_input_tokens_seen": 15134640, + "step": 19845 + }, + { + "epoch": 41.268191268191266, + "grad_norm": 0.00020071813196409494, + "learning_rate": 0.15177888513996218, + "loss": 0.2521, + "num_input_tokens_seen": 15138384, + "step": 19850 + }, + { + "epoch": 41.27858627858628, + "grad_norm": 0.00043783418368548155, + "learning_rate": 0.15171998428443592, + "loss": 0.2777, + "num_input_tokens_seen": 15142160, + "step": 19855 + }, + { + "epoch": 41.28898128898129, + "grad_norm": 0.0006280025700107217, + "learning_rate": 0.1516610831636665, + "loss": 0.2676, + "num_input_tokens_seen": 15145968, + "step": 19860 + }, + { + "epoch": 41.2993762993763, + "grad_norm": 0.00031761726131662726, + "learning_rate": 0.15160218178673715, + "loss": 0.2731, + "num_input_tokens_seen": 15149648, + "step": 19865 + }, + { + "epoch": 41.30977130977131, + "grad_norm": 0.00023938054800964892, + "learning_rate": 0.15154328016273122, + "loss": 0.247, + "num_input_tokens_seen": 15153360, + "step": 19870 + }, + { + "epoch": 41.32016632016632, + "grad_norm": 0.0005181441665627062, + "learning_rate": 0.1514843783007321, + "loss": 0.2572, + "num_input_tokens_seen": 15157136, + "step": 19875 + }, + { + "epoch": 41.33056133056133, + "grad_norm": 0.00014944201393518597, + "learning_rate": 0.15142547620982322, + "loss": 0.2472, + "num_input_tokens_seen": 15160912, + "step": 19880 + }, + { + "epoch": 41.34095634095634, + "grad_norm": 0.0005177579587325454, + "learning_rate": 0.15136657389908797, + "loss": 0.2575, + "num_input_tokens_seen": 15164784, + "step": 19885 + }, + { + "epoch": 41.351351351351354, + "grad_norm": 0.00048187319771386683, + "learning_rate": 0.15130767137760986, + "loss": 0.2663, + "num_input_tokens_seen": 15168656, + "step": 19890 + }, + { + "epoch": 41.36174636174636, + "grad_norm": 0.0007022770587354898, + "learning_rate": 0.15124876865447243, + "loss": 0.2551, + "num_input_tokens_seen": 15172528, + "step": 19895 + }, + { + "epoch": 41.37214137214137, + "grad_norm": 0.00039229870890267193, + "learning_rate": 0.15118986573875912, + "loss": 0.2728, + "num_input_tokens_seen": 15176432, + "step": 19900 + }, + { + "epoch": 41.38253638253638, + "grad_norm": 0.0005361128132790327, + "learning_rate": 0.15113096263955358, + "loss": 0.2422, + "num_input_tokens_seen": 15180240, + "step": 19905 + }, + { + "epoch": 41.392931392931395, + "grad_norm": 0.0004896206082776189, + "learning_rate": 0.1510720593659394, + "loss": 0.2695, + "num_input_tokens_seen": 15183792, + "step": 19910 + }, + { + "epoch": 41.4033264033264, + "grad_norm": 0.00017406239931005985, + "learning_rate": 0.15101315592700015, + "loss": 0.2689, + "num_input_tokens_seen": 15187600, + "step": 19915 + }, + { + "epoch": 41.41372141372141, + "grad_norm": 0.0004011830023955554, + "learning_rate": 0.15095425233181956, + "loss": 0.2704, + "num_input_tokens_seen": 15191440, + "step": 19920 + }, + { + "epoch": 41.42411642411643, + "grad_norm": 0.00024775919155217707, + "learning_rate": 0.15089534858948128, + "loss": 0.2817, + "num_input_tokens_seen": 15195312, + "step": 19925 + }, + { + "epoch": 41.434511434511435, + "grad_norm": 0.0001525456755189225, + "learning_rate": 0.15083644470906898, + "loss": 0.2696, + "num_input_tokens_seen": 15199216, + "step": 19930 + }, + { + "epoch": 41.444906444906444, + "grad_norm": 0.0005426400457508862, + "learning_rate": 0.1507775406996664, + "loss": 0.2578, + "num_input_tokens_seen": 15202992, + "step": 19935 + }, + { + "epoch": 41.45530145530145, + "grad_norm": 9.243025124305859e-05, + "learning_rate": 0.15071863657035725, + "loss": 0.266, + "num_input_tokens_seen": 15206832, + "step": 19940 + }, + { + "epoch": 41.46569646569647, + "grad_norm": 9.435565152671188e-05, + "learning_rate": 0.15065973233022534, + "loss": 0.2689, + "num_input_tokens_seen": 15210640, + "step": 19945 + }, + { + "epoch": 41.476091476091476, + "grad_norm": 0.0004644471046049148, + "learning_rate": 0.15060082798835442, + "loss": 0.2718, + "num_input_tokens_seen": 15214448, + "step": 19950 + }, + { + "epoch": 41.486486486486484, + "grad_norm": 0.0002996270777657628, + "learning_rate": 0.15054192355382823, + "loss": 0.2732, + "num_input_tokens_seen": 15218288, + "step": 19955 + }, + { + "epoch": 41.4968814968815, + "grad_norm": 0.0004439016920514405, + "learning_rate": 0.15048301903573066, + "loss": 0.2485, + "num_input_tokens_seen": 15222064, + "step": 19960 + }, + { + "epoch": 41.50727650727651, + "grad_norm": 0.00010482939978828654, + "learning_rate": 0.15042411444314546, + "loss": 0.2551, + "num_input_tokens_seen": 15225936, + "step": 19965 + }, + { + "epoch": 41.517671517671516, + "grad_norm": 0.00015237348270602524, + "learning_rate": 0.1503652097851565, + "loss": 0.2537, + "num_input_tokens_seen": 15229648, + "step": 19970 + }, + { + "epoch": 41.528066528066525, + "grad_norm": 0.00012742483522742987, + "learning_rate": 0.15030630507084758, + "loss": 0.2862, + "num_input_tokens_seen": 15233424, + "step": 19975 + }, + { + "epoch": 41.53846153846154, + "grad_norm": 0.0005171297816559672, + "learning_rate": 0.1502474003093026, + "loss": 0.2683, + "num_input_tokens_seen": 15237200, + "step": 19980 + }, + { + "epoch": 41.54885654885655, + "grad_norm": 7.528687274316326e-05, + "learning_rate": 0.15018849550960536, + "loss": 0.2732, + "num_input_tokens_seen": 15240944, + "step": 19985 + }, + { + "epoch": 41.55925155925156, + "grad_norm": 6.710041634505615e-05, + "learning_rate": 0.15012959068083975, + "loss": 0.2649, + "num_input_tokens_seen": 15244624, + "step": 19990 + }, + { + "epoch": 41.56964656964657, + "grad_norm": 0.00038850161945447326, + "learning_rate": 0.1500706858320896, + "loss": 0.2466, + "num_input_tokens_seen": 15248656, + "step": 19995 + }, + { + "epoch": 41.58004158004158, + "grad_norm": 0.0003976233711000532, + "learning_rate": 0.15001178097243886, + "loss": 0.306, + "num_input_tokens_seen": 15252464, + "step": 20000 + }, + { + "epoch": 41.58004158004158, + "eval_loss": 0.24729791283607483, + "eval_runtime": 13.4069, + "eval_samples_per_second": 63.848, + "eval_steps_per_second": 15.962, + "num_input_tokens_seen": 15252464, + "step": 20000 + }, + { + "epoch": 41.59043659043659, + "grad_norm": 0.00014859740622341633, + "learning_rate": 0.1499528761109713, + "loss": 0.2745, + "num_input_tokens_seen": 15256144, + "step": 20005 + }, + { + "epoch": 41.6008316008316, + "grad_norm": 0.00039404688868671656, + "learning_rate": 0.14989397125677087, + "loss": 0.262, + "num_input_tokens_seen": 15260144, + "step": 20010 + }, + { + "epoch": 41.61122661122661, + "grad_norm": 0.00026431374135427177, + "learning_rate": 0.14983506641892141, + "loss": 0.278, + "num_input_tokens_seen": 15264016, + "step": 20015 + }, + { + "epoch": 41.62162162162162, + "grad_norm": 9.346404112875462e-05, + "learning_rate": 0.14977616160650672, + "loss": 0.2474, + "num_input_tokens_seen": 15267824, + "step": 20020 + }, + { + "epoch": 41.63201663201663, + "grad_norm": 0.00016083351511042565, + "learning_rate": 0.14971725682861076, + "loss": 0.2468, + "num_input_tokens_seen": 15271632, + "step": 20025 + }, + { + "epoch": 41.642411642411645, + "grad_norm": 0.00033178890589624643, + "learning_rate": 0.14965835209431738, + "loss": 0.2893, + "num_input_tokens_seen": 15275440, + "step": 20030 + }, + { + "epoch": 41.65280665280665, + "grad_norm": 7.432598795276135e-05, + "learning_rate": 0.14959944741271036, + "loss": 0.2665, + "num_input_tokens_seen": 15279408, + "step": 20035 + }, + { + "epoch": 41.66320166320166, + "grad_norm": 0.00012657100160140544, + "learning_rate": 0.14954054279287363, + "loss": 0.2685, + "num_input_tokens_seen": 15283184, + "step": 20040 + }, + { + "epoch": 41.67359667359668, + "grad_norm": 0.00029358078609220684, + "learning_rate": 0.14948163824389094, + "loss": 0.2742, + "num_input_tokens_seen": 15286736, + "step": 20045 + }, + { + "epoch": 41.683991683991685, + "grad_norm": 0.00027662538923323154, + "learning_rate": 0.14942273377484613, + "loss": 0.2704, + "num_input_tokens_seen": 15290416, + "step": 20050 + }, + { + "epoch": 41.694386694386694, + "grad_norm": 0.00013197209045756608, + "learning_rate": 0.1493638293948231, + "loss": 0.2767, + "num_input_tokens_seen": 15294192, + "step": 20055 + }, + { + "epoch": 41.7047817047817, + "grad_norm": 0.00020125800801906735, + "learning_rate": 0.14930492511290547, + "loss": 0.2698, + "num_input_tokens_seen": 15298192, + "step": 20060 + }, + { + "epoch": 41.71517671517672, + "grad_norm": 0.00020611828949768096, + "learning_rate": 0.14924602093817715, + "loss": 0.2619, + "num_input_tokens_seen": 15301968, + "step": 20065 + }, + { + "epoch": 41.725571725571726, + "grad_norm": 0.00025062181521207094, + "learning_rate": 0.14918711687972194, + "loss": 0.2719, + "num_input_tokens_seen": 15305648, + "step": 20070 + }, + { + "epoch": 41.735966735966734, + "grad_norm": 0.0004917155019938946, + "learning_rate": 0.14912821294662346, + "loss": 0.2632, + "num_input_tokens_seen": 15309648, + "step": 20075 + }, + { + "epoch": 41.74636174636175, + "grad_norm": 0.00011816350161097944, + "learning_rate": 0.14906930914796554, + "loss": 0.2685, + "num_input_tokens_seen": 15313552, + "step": 20080 + }, + { + "epoch": 41.75675675675676, + "grad_norm": 0.00045738540939055383, + "learning_rate": 0.14901040549283182, + "loss": 0.2552, + "num_input_tokens_seen": 15317296, + "step": 20085 + }, + { + "epoch": 41.767151767151766, + "grad_norm": 0.00012525790953077376, + "learning_rate": 0.148951501990306, + "loss": 0.2594, + "num_input_tokens_seen": 15321072, + "step": 20090 + }, + { + "epoch": 41.777546777546775, + "grad_norm": 0.00028359986026771367, + "learning_rate": 0.14889259864947177, + "loss": 0.2771, + "num_input_tokens_seen": 15324880, + "step": 20095 + }, + { + "epoch": 41.78794178794179, + "grad_norm": 0.0008753768051974475, + "learning_rate": 0.14883369547941272, + "loss": 0.2706, + "num_input_tokens_seen": 15328720, + "step": 20100 + }, + { + "epoch": 41.7983367983368, + "grad_norm": 7.559965160908177e-05, + "learning_rate": 0.14877479248921247, + "loss": 0.2477, + "num_input_tokens_seen": 15332400, + "step": 20105 + }, + { + "epoch": 41.80873180873181, + "grad_norm": 0.00014708726666867733, + "learning_rate": 0.14871588968795468, + "loss": 0.2625, + "num_input_tokens_seen": 15336272, + "step": 20110 + }, + { + "epoch": 41.81912681912682, + "grad_norm": 0.00046606335672549903, + "learning_rate": 0.1486569870847228, + "loss": 0.2536, + "num_input_tokens_seen": 15340016, + "step": 20115 + }, + { + "epoch": 41.82952182952183, + "grad_norm": 0.0002363435341976583, + "learning_rate": 0.1485980846886004, + "loss": 0.2489, + "num_input_tokens_seen": 15343856, + "step": 20120 + }, + { + "epoch": 41.83991683991684, + "grad_norm": 0.0005542057915590703, + "learning_rate": 0.14853918250867096, + "loss": 0.2674, + "num_input_tokens_seen": 15347664, + "step": 20125 + }, + { + "epoch": 41.85031185031185, + "grad_norm": 0.0008636984275653958, + "learning_rate": 0.1484802805540179, + "loss": 0.2848, + "num_input_tokens_seen": 15351440, + "step": 20130 + }, + { + "epoch": 41.86070686070686, + "grad_norm": 0.0004019705520477146, + "learning_rate": 0.14842137883372472, + "loss": 0.2838, + "num_input_tokens_seen": 15355440, + "step": 20135 + }, + { + "epoch": 41.87110187110187, + "grad_norm": 0.0004886981914751232, + "learning_rate": 0.14836247735687474, + "loss": 0.2834, + "num_input_tokens_seen": 15359280, + "step": 20140 + }, + { + "epoch": 41.88149688149688, + "grad_norm": 0.0001917227345984429, + "learning_rate": 0.14830357613255132, + "loss": 0.2815, + "num_input_tokens_seen": 15362992, + "step": 20145 + }, + { + "epoch": 41.891891891891895, + "grad_norm": 0.0002934641670435667, + "learning_rate": 0.1482446751698378, + "loss": 0.2748, + "num_input_tokens_seen": 15366864, + "step": 20150 + }, + { + "epoch": 41.9022869022869, + "grad_norm": 5.305219747242518e-05, + "learning_rate": 0.14818577447781744, + "loss": 0.2646, + "num_input_tokens_seen": 15370736, + "step": 20155 + }, + { + "epoch": 41.91268191268191, + "grad_norm": 0.00040848529897630215, + "learning_rate": 0.14812687406557346, + "loss": 0.2949, + "num_input_tokens_seen": 15374448, + "step": 20160 + }, + { + "epoch": 41.92307692307692, + "grad_norm": 0.0002495793451089412, + "learning_rate": 0.14806797394218899, + "loss": 0.2682, + "num_input_tokens_seen": 15378160, + "step": 20165 + }, + { + "epoch": 41.933471933471935, + "grad_norm": 0.00020230650261510164, + "learning_rate": 0.1480090741167472, + "loss": 0.2822, + "num_input_tokens_seen": 15381936, + "step": 20170 + }, + { + "epoch": 41.943866943866944, + "grad_norm": 0.00026740634348243475, + "learning_rate": 0.1479501745983313, + "loss": 0.2721, + "num_input_tokens_seen": 15385680, + "step": 20175 + }, + { + "epoch": 41.95426195426195, + "grad_norm": 0.0004154109919909388, + "learning_rate": 0.14789127539602415, + "loss": 0.2604, + "num_input_tokens_seen": 15389456, + "step": 20180 + }, + { + "epoch": 41.96465696465697, + "grad_norm": 0.00041083982796408236, + "learning_rate": 0.14783237651890885, + "loss": 0.2635, + "num_input_tokens_seen": 15393456, + "step": 20185 + }, + { + "epoch": 41.975051975051976, + "grad_norm": 0.00016418518498539925, + "learning_rate": 0.14777347797606838, + "loss": 0.2865, + "num_input_tokens_seen": 15397328, + "step": 20190 + }, + { + "epoch": 41.985446985446984, + "grad_norm": 0.0005811268929392099, + "learning_rate": 0.14771457977658553, + "loss": 0.2779, + "num_input_tokens_seen": 15401168, + "step": 20195 + }, + { + "epoch": 41.99584199584199, + "grad_norm": 0.00013299698184709996, + "learning_rate": 0.14765568192954326, + "loss": 0.2733, + "num_input_tokens_seen": 15404912, + "step": 20200 + }, + { + "epoch": 41.99584199584199, + "eval_loss": 0.26801010966300964, + "eval_runtime": 13.4426, + "eval_samples_per_second": 63.678, + "eval_steps_per_second": 15.92, + "num_input_tokens_seen": 15404912, + "step": 20200 + }, + { + "epoch": 42.00623700623701, + "grad_norm": 0.00031006947392597795, + "learning_rate": 0.14759678444402421, + "loss": 0.278, + "num_input_tokens_seen": 15408760, + "step": 20205 + }, + { + "epoch": 42.016632016632016, + "grad_norm": 9.294260962633416e-05, + "learning_rate": 0.14753788732911122, + "loss": 0.2859, + "num_input_tokens_seen": 15412568, + "step": 20210 + }, + { + "epoch": 42.027027027027025, + "grad_norm": 4.357831130619161e-05, + "learning_rate": 0.147478990593887, + "loss": 0.257, + "num_input_tokens_seen": 15416312, + "step": 20215 + }, + { + "epoch": 42.03742203742204, + "grad_norm": 0.000271494995104149, + "learning_rate": 0.14742009424743405, + "loss": 0.2626, + "num_input_tokens_seen": 15420056, + "step": 20220 + }, + { + "epoch": 42.04781704781705, + "grad_norm": 0.0005096501554362476, + "learning_rate": 0.14736119829883504, + "loss": 0.2538, + "num_input_tokens_seen": 15423992, + "step": 20225 + }, + { + "epoch": 42.05821205821206, + "grad_norm": 0.0002849528391379863, + "learning_rate": 0.14730230275717243, + "loss": 0.2684, + "num_input_tokens_seen": 15427992, + "step": 20230 + }, + { + "epoch": 42.06860706860707, + "grad_norm": 5.8325760619482026e-05, + "learning_rate": 0.14724340763152854, + "loss": 0.2771, + "num_input_tokens_seen": 15431704, + "step": 20235 + }, + { + "epoch": 42.07900207900208, + "grad_norm": 0.0004823118506465107, + "learning_rate": 0.14718451293098594, + "loss": 0.2547, + "num_input_tokens_seen": 15435672, + "step": 20240 + }, + { + "epoch": 42.08939708939709, + "grad_norm": 0.0007917492766864598, + "learning_rate": 0.14712561866462676, + "loss": 0.2757, + "num_input_tokens_seen": 15439512, + "step": 20245 + }, + { + "epoch": 42.0997920997921, + "grad_norm": 0.00016996277554426342, + "learning_rate": 0.1470667248415333, + "loss": 0.2724, + "num_input_tokens_seen": 15443256, + "step": 20250 + }, + { + "epoch": 42.11018711018711, + "grad_norm": 0.00012539848103187978, + "learning_rate": 0.1470078314707878, + "loss": 0.269, + "num_input_tokens_seen": 15447032, + "step": 20255 + }, + { + "epoch": 42.12058212058212, + "grad_norm": 0.00018367639859206975, + "learning_rate": 0.14694893856147223, + "loss": 0.2693, + "num_input_tokens_seen": 15450808, + "step": 20260 + }, + { + "epoch": 42.13097713097713, + "grad_norm": 0.0003650474245660007, + "learning_rate": 0.14689004612266868, + "loss": 0.2738, + "num_input_tokens_seen": 15454712, + "step": 20265 + }, + { + "epoch": 42.141372141372145, + "grad_norm": 0.00014819738862570375, + "learning_rate": 0.14683115416345913, + "loss": 0.2643, + "num_input_tokens_seen": 15458488, + "step": 20270 + }, + { + "epoch": 42.15176715176715, + "grad_norm": 0.00036469101905822754, + "learning_rate": 0.1467722626929254, + "loss": 0.2737, + "num_input_tokens_seen": 15462360, + "step": 20275 + }, + { + "epoch": 42.16216216216216, + "grad_norm": 4.69807127956301e-05, + "learning_rate": 0.14671337172014937, + "loss": 0.2487, + "num_input_tokens_seen": 15466232, + "step": 20280 + }, + { + "epoch": 42.17255717255717, + "grad_norm": 0.0005535532836802304, + "learning_rate": 0.14665448125421265, + "loss": 0.2879, + "num_input_tokens_seen": 15470200, + "step": 20285 + }, + { + "epoch": 42.182952182952185, + "grad_norm": 0.00025059442850761116, + "learning_rate": 0.146595591304197, + "loss": 0.2658, + "num_input_tokens_seen": 15474104, + "step": 20290 + }, + { + "epoch": 42.19334719334719, + "grad_norm": 0.00022680446272715926, + "learning_rate": 0.14653670187918397, + "loss": 0.2669, + "num_input_tokens_seen": 15477816, + "step": 20295 + }, + { + "epoch": 42.2037422037422, + "grad_norm": 0.00017650314839556813, + "learning_rate": 0.14647781298825502, + "loss": 0.2684, + "num_input_tokens_seen": 15481592, + "step": 20300 + }, + { + "epoch": 42.21413721413722, + "grad_norm": 9.735923958942294e-05, + "learning_rate": 0.14641892464049153, + "loss": 0.2544, + "num_input_tokens_seen": 15485368, + "step": 20305 + }, + { + "epoch": 42.224532224532226, + "grad_norm": 6.39576610410586e-05, + "learning_rate": 0.14636003684497495, + "loss": 0.2842, + "num_input_tokens_seen": 15489240, + "step": 20310 + }, + { + "epoch": 42.234927234927234, + "grad_norm": 0.0002896163787227124, + "learning_rate": 0.14630114961078636, + "loss": 0.2651, + "num_input_tokens_seen": 15493112, + "step": 20315 + }, + { + "epoch": 42.24532224532224, + "grad_norm": 0.0006976997246965766, + "learning_rate": 0.14624226294700704, + "loss": 0.2497, + "num_input_tokens_seen": 15496984, + "step": 20320 + }, + { + "epoch": 42.25571725571726, + "grad_norm": 0.0006990879774093628, + "learning_rate": 0.14618337686271793, + "loss": 0.3222, + "num_input_tokens_seen": 15500696, + "step": 20325 + }, + { + "epoch": 42.266112266112266, + "grad_norm": 0.000442896707681939, + "learning_rate": 0.1461244913670001, + "loss": 0.275, + "num_input_tokens_seen": 15504536, + "step": 20330 + }, + { + "epoch": 42.276507276507274, + "grad_norm": 0.00045019612298347056, + "learning_rate": 0.1460656064689344, + "loss": 0.2586, + "num_input_tokens_seen": 15508152, + "step": 20335 + }, + { + "epoch": 42.28690228690229, + "grad_norm": 0.00014745046792086214, + "learning_rate": 0.14600672217760163, + "loss": 0.2578, + "num_input_tokens_seen": 15512024, + "step": 20340 + }, + { + "epoch": 42.2972972972973, + "grad_norm": 0.00022589118452742696, + "learning_rate": 0.14594783850208248, + "loss": 0.2269, + "num_input_tokens_seen": 15515608, + "step": 20345 + }, + { + "epoch": 42.30769230769231, + "grad_norm": 1.6374231563531794e-05, + "learning_rate": 0.14588895545145758, + "loss": 0.2678, + "num_input_tokens_seen": 15519448, + "step": 20350 + }, + { + "epoch": 42.318087318087315, + "grad_norm": 0.0001519451179774478, + "learning_rate": 0.14583007303480738, + "loss": 0.2742, + "num_input_tokens_seen": 15523064, + "step": 20355 + }, + { + "epoch": 42.32848232848233, + "grad_norm": 3.054229455301538e-05, + "learning_rate": 0.14577119126121235, + "loss": 0.2743, + "num_input_tokens_seen": 15526840, + "step": 20360 + }, + { + "epoch": 42.33887733887734, + "grad_norm": 0.00022006033395882696, + "learning_rate": 0.14571231013975272, + "loss": 0.2557, + "num_input_tokens_seen": 15530616, + "step": 20365 + }, + { + "epoch": 42.34927234927235, + "grad_norm": 0.0002961000718642026, + "learning_rate": 0.1456534296795088, + "loss": 0.2625, + "num_input_tokens_seen": 15534520, + "step": 20370 + }, + { + "epoch": 42.35966735966736, + "grad_norm": 9.562297782395035e-05, + "learning_rate": 0.14559454988956066, + "loss": 0.2679, + "num_input_tokens_seen": 15538360, + "step": 20375 + }, + { + "epoch": 42.37006237006237, + "grad_norm": 0.0006060078158043325, + "learning_rate": 0.1455356707789882, + "loss": 0.2767, + "num_input_tokens_seen": 15541976, + "step": 20380 + }, + { + "epoch": 42.38045738045738, + "grad_norm": 0.0003607240505516529, + "learning_rate": 0.14547679235687147, + "loss": 0.2382, + "num_input_tokens_seen": 15545784, + "step": 20385 + }, + { + "epoch": 42.39085239085239, + "grad_norm": 0.00011270622053416446, + "learning_rate": 0.14541791463229023, + "loss": 0.2718, + "num_input_tokens_seen": 15549592, + "step": 20390 + }, + { + "epoch": 42.4012474012474, + "grad_norm": 0.00017741376359481364, + "learning_rate": 0.14535903761432406, + "loss": 0.2586, + "num_input_tokens_seen": 15553368, + "step": 20395 + }, + { + "epoch": 42.41164241164241, + "grad_norm": 8.503738354193047e-05, + "learning_rate": 0.1453001613120527, + "loss": 0.2541, + "num_input_tokens_seen": 15557176, + "step": 20400 + }, + { + "epoch": 42.41164241164241, + "eval_loss": 0.2488873153924942, + "eval_runtime": 13.399, + "eval_samples_per_second": 63.886, + "eval_steps_per_second": 15.971, + "num_input_tokens_seen": 15557176, + "step": 20400 + }, + { + "epoch": 42.42203742203742, + "grad_norm": 0.00014495462528429925, + "learning_rate": 0.14524128573455547, + "loss": 0.2615, + "num_input_tokens_seen": 15561112, + "step": 20405 + }, + { + "epoch": 42.432432432432435, + "grad_norm": 0.0005751390708610415, + "learning_rate": 0.14518241089091177, + "loss": 0.2665, + "num_input_tokens_seen": 15564984, + "step": 20410 + }, + { + "epoch": 42.44282744282744, + "grad_norm": 0.00036329164868220687, + "learning_rate": 0.1451235367902009, + "loss": 0.278, + "num_input_tokens_seen": 15568952, + "step": 20415 + }, + { + "epoch": 42.45322245322245, + "grad_norm": 0.0006299822707660496, + "learning_rate": 0.1450646634415019, + "loss": 0.2688, + "num_input_tokens_seen": 15572696, + "step": 20420 + }, + { + "epoch": 42.46361746361746, + "grad_norm": 0.0002691232948563993, + "learning_rate": 0.1450057908538938, + "loss": 0.2781, + "num_input_tokens_seen": 15576568, + "step": 20425 + }, + { + "epoch": 42.474012474012476, + "grad_norm": 0.0002322325890418142, + "learning_rate": 0.14494691903645557, + "loss": 0.2768, + "num_input_tokens_seen": 15580536, + "step": 20430 + }, + { + "epoch": 42.484407484407484, + "grad_norm": 0.00011574888048926368, + "learning_rate": 0.14488804799826588, + "loss": 0.27, + "num_input_tokens_seen": 15584312, + "step": 20435 + }, + { + "epoch": 42.49480249480249, + "grad_norm": 0.00023743572819512337, + "learning_rate": 0.14482917774840348, + "loss": 0.2532, + "num_input_tokens_seen": 15588024, + "step": 20440 + }, + { + "epoch": 42.50519750519751, + "grad_norm": 0.00011074316716985777, + "learning_rate": 0.14477030829594684, + "loss": 0.2726, + "num_input_tokens_seen": 15591928, + "step": 20445 + }, + { + "epoch": 42.515592515592516, + "grad_norm": 9.598352335160598e-05, + "learning_rate": 0.14471143964997432, + "loss": 0.2704, + "num_input_tokens_seen": 15595800, + "step": 20450 + }, + { + "epoch": 42.525987525987524, + "grad_norm": 0.00012516751303337514, + "learning_rate": 0.14465257181956434, + "loss": 0.2755, + "num_input_tokens_seen": 15599704, + "step": 20455 + }, + { + "epoch": 42.53638253638254, + "grad_norm": 0.0007414927240461111, + "learning_rate": 0.1445937048137949, + "loss": 0.266, + "num_input_tokens_seen": 15603544, + "step": 20460 + }, + { + "epoch": 42.54677754677755, + "grad_norm": 3.4345830499660224e-05, + "learning_rate": 0.14453483864174416, + "loss": 0.2767, + "num_input_tokens_seen": 15607224, + "step": 20465 + }, + { + "epoch": 42.55717255717256, + "grad_norm": 0.0004697073600254953, + "learning_rate": 0.14447597331249, + "loss": 0.2602, + "num_input_tokens_seen": 15611064, + "step": 20470 + }, + { + "epoch": 42.567567567567565, + "grad_norm": 0.00012706419511232525, + "learning_rate": 0.1444171088351102, + "loss": 0.2526, + "num_input_tokens_seen": 15614968, + "step": 20475 + }, + { + "epoch": 42.57796257796258, + "grad_norm": 0.000415144138969481, + "learning_rate": 0.14435824521868235, + "loss": 0.2632, + "num_input_tokens_seen": 15618712, + "step": 20480 + }, + { + "epoch": 42.58835758835759, + "grad_norm": 0.00030475438688881695, + "learning_rate": 0.14429938247228397, + "loss": 0.2924, + "num_input_tokens_seen": 15622552, + "step": 20485 + }, + { + "epoch": 42.5987525987526, + "grad_norm": 0.00022947760589886457, + "learning_rate": 0.14424052060499243, + "loss": 0.2628, + "num_input_tokens_seen": 15626392, + "step": 20490 + }, + { + "epoch": 42.60914760914761, + "grad_norm": 0.0003667368437163532, + "learning_rate": 0.14418165962588506, + "loss": 0.2653, + "num_input_tokens_seen": 15630072, + "step": 20495 + }, + { + "epoch": 42.61954261954262, + "grad_norm": 0.0002552943187765777, + "learning_rate": 0.1441227995440388, + "loss": 0.2567, + "num_input_tokens_seen": 15633944, + "step": 20500 + }, + { + "epoch": 42.62993762993763, + "grad_norm": 0.0002152569213649258, + "learning_rate": 0.14406394036853082, + "loss": 0.252, + "num_input_tokens_seen": 15637848, + "step": 20505 + }, + { + "epoch": 42.64033264033264, + "grad_norm": 0.0001386357325827703, + "learning_rate": 0.14400508210843774, + "loss": 0.2963, + "num_input_tokens_seen": 15641720, + "step": 20510 + }, + { + "epoch": 42.65072765072765, + "grad_norm": 0.00015914726827759296, + "learning_rate": 0.1439462247728364, + "loss": 0.271, + "num_input_tokens_seen": 15645496, + "step": 20515 + }, + { + "epoch": 42.66112266112266, + "grad_norm": 0.00020199711434543133, + "learning_rate": 0.14388736837080326, + "loss": 0.2591, + "num_input_tokens_seen": 15649208, + "step": 20520 + }, + { + "epoch": 42.67151767151767, + "grad_norm": 7.850349356886e-05, + "learning_rate": 0.14382851291141469, + "loss": 0.2786, + "num_input_tokens_seen": 15653048, + "step": 20525 + }, + { + "epoch": 42.681912681912685, + "grad_norm": 5.353394226403907e-05, + "learning_rate": 0.14376965840374697, + "loss": 0.2544, + "num_input_tokens_seen": 15656856, + "step": 20530 + }, + { + "epoch": 42.69230769230769, + "grad_norm": 0.00013941607903689146, + "learning_rate": 0.14371080485687632, + "loss": 0.2533, + "num_input_tokens_seen": 15660664, + "step": 20535 + }, + { + "epoch": 42.7027027027027, + "grad_norm": 6.965376087464392e-05, + "learning_rate": 0.1436519522798785, + "loss": 0.2421, + "num_input_tokens_seen": 15664376, + "step": 20540 + }, + { + "epoch": 42.71309771309771, + "grad_norm": 0.00013098256022203714, + "learning_rate": 0.14359310068182948, + "loss": 0.2805, + "num_input_tokens_seen": 15668184, + "step": 20545 + }, + { + "epoch": 42.723492723492726, + "grad_norm": 0.00017030752496793866, + "learning_rate": 0.14353425007180484, + "loss": 0.2301, + "num_input_tokens_seen": 15671896, + "step": 20550 + }, + { + "epoch": 42.733887733887734, + "grad_norm": 0.0005642463220283389, + "learning_rate": 0.14347540045888005, + "loss": 0.2759, + "num_input_tokens_seen": 15675672, + "step": 20555 + }, + { + "epoch": 42.74428274428274, + "grad_norm": 0.00025218207156285644, + "learning_rate": 0.14341655185213056, + "loss": 0.2658, + "num_input_tokens_seen": 15679384, + "step": 20560 + }, + { + "epoch": 42.75467775467776, + "grad_norm": 0.00020625402976293117, + "learning_rate": 0.14335770426063144, + "loss": 0.2579, + "num_input_tokens_seen": 15683448, + "step": 20565 + }, + { + "epoch": 42.765072765072766, + "grad_norm": 0.00019057997269555926, + "learning_rate": 0.1432988576934578, + "loss": 0.2433, + "num_input_tokens_seen": 15687224, + "step": 20570 + }, + { + "epoch": 42.775467775467774, + "grad_norm": 0.00020051821775268763, + "learning_rate": 0.14324001215968457, + "loss": 0.2732, + "num_input_tokens_seen": 15690968, + "step": 20575 + }, + { + "epoch": 42.78586278586278, + "grad_norm": 6.650371506111696e-05, + "learning_rate": 0.14318116766838637, + "loss": 0.2739, + "num_input_tokens_seen": 15694712, + "step": 20580 + }, + { + "epoch": 42.7962577962578, + "grad_norm": 6.460359873017296e-05, + "learning_rate": 0.14312232422863788, + "loss": 0.2549, + "num_input_tokens_seen": 15698584, + "step": 20585 + }, + { + "epoch": 42.80665280665281, + "grad_norm": 5.0154740165453404e-05, + "learning_rate": 0.14306348184951334, + "loss": 0.2665, + "num_input_tokens_seen": 15702360, + "step": 20590 + }, + { + "epoch": 42.817047817047815, + "grad_norm": 0.000191645827726461, + "learning_rate": 0.1430046405400871, + "loss": 0.2626, + "num_input_tokens_seen": 15706168, + "step": 20595 + }, + { + "epoch": 42.82744282744283, + "grad_norm": 0.0001799809542717412, + "learning_rate": 0.14294580030943324, + "loss": 0.2715, + "num_input_tokens_seen": 15709912, + "step": 20600 + }, + { + "epoch": 42.82744282744283, + "eval_loss": 0.2510470747947693, + "eval_runtime": 13.4238, + "eval_samples_per_second": 63.767, + "eval_steps_per_second": 15.942, + "num_input_tokens_seen": 15709912, + "step": 20600 + }, + { + "epoch": 42.83783783783784, + "grad_norm": 0.0003508008667267859, + "learning_rate": 0.14288696116662553, + "loss": 0.2921, + "num_input_tokens_seen": 15713752, + "step": 20605 + }, + { + "epoch": 42.84823284823285, + "grad_norm": 0.0002212283870903775, + "learning_rate": 0.1428281231207378, + "loss": 0.2492, + "num_input_tokens_seen": 15717656, + "step": 20610 + }, + { + "epoch": 42.858627858627855, + "grad_norm": 0.00030565151246264577, + "learning_rate": 0.1427692861808437, + "loss": 0.2732, + "num_input_tokens_seen": 15721528, + "step": 20615 + }, + { + "epoch": 42.86902286902287, + "grad_norm": 0.0005401631933636963, + "learning_rate": 0.1427104503560165, + "loss": 0.2434, + "num_input_tokens_seen": 15725432, + "step": 20620 + }, + { + "epoch": 42.87941787941788, + "grad_norm": 0.0001575544592924416, + "learning_rate": 0.14265161565532947, + "loss": 0.2421, + "num_input_tokens_seen": 15729272, + "step": 20625 + }, + { + "epoch": 42.88981288981289, + "grad_norm": 0.00032919671502895653, + "learning_rate": 0.14259278208785564, + "loss": 0.2902, + "num_input_tokens_seen": 15733112, + "step": 20630 + }, + { + "epoch": 42.9002079002079, + "grad_norm": 0.00030925741884857416, + "learning_rate": 0.14253394966266789, + "loss": 0.2714, + "num_input_tokens_seen": 15737016, + "step": 20635 + }, + { + "epoch": 42.91060291060291, + "grad_norm": 0.00011206260387552902, + "learning_rate": 0.14247511838883894, + "loss": 0.2575, + "num_input_tokens_seen": 15740792, + "step": 20640 + }, + { + "epoch": 42.92099792099792, + "grad_norm": 0.00021110885427333415, + "learning_rate": 0.14241628827544126, + "loss": 0.2626, + "num_input_tokens_seen": 15744600, + "step": 20645 + }, + { + "epoch": 42.931392931392935, + "grad_norm": 6.634168676100671e-05, + "learning_rate": 0.14235745933154723, + "loss": 0.2642, + "num_input_tokens_seen": 15748280, + "step": 20650 + }, + { + "epoch": 42.94178794178794, + "grad_norm": 0.0002306604728801176, + "learning_rate": 0.14229863156622907, + "loss": 0.2787, + "num_input_tokens_seen": 15752120, + "step": 20655 + }, + { + "epoch": 42.95218295218295, + "grad_norm": 0.00016673390928190202, + "learning_rate": 0.14223980498855868, + "loss": 0.2656, + "num_input_tokens_seen": 15755864, + "step": 20660 + }, + { + "epoch": 42.96257796257796, + "grad_norm": 0.0002595261612441391, + "learning_rate": 0.14218097960760792, + "loss": 0.2578, + "num_input_tokens_seen": 15759704, + "step": 20665 + }, + { + "epoch": 42.972972972972975, + "grad_norm": 0.0002815193438436836, + "learning_rate": 0.1421221554324483, + "loss": 0.2898, + "num_input_tokens_seen": 15763480, + "step": 20670 + }, + { + "epoch": 42.983367983367984, + "grad_norm": 0.00032388733234256506, + "learning_rate": 0.1420633324721513, + "loss": 0.29, + "num_input_tokens_seen": 15767224, + "step": 20675 + }, + { + "epoch": 42.99376299376299, + "grad_norm": 0.00013253749057184905, + "learning_rate": 0.14200451073578824, + "loss": 0.2703, + "num_input_tokens_seen": 15771000, + "step": 20680 + }, + { + "epoch": 43.00415800415801, + "grad_norm": 0.00013155497435946018, + "learning_rate": 0.14194569023243003, + "loss": 0.2628, + "num_input_tokens_seen": 15774816, + "step": 20685 + }, + { + "epoch": 43.014553014553016, + "grad_norm": 7.800347520969808e-05, + "learning_rate": 0.14188687097114766, + "loss": 0.2649, + "num_input_tokens_seen": 15778592, + "step": 20690 + }, + { + "epoch": 43.024948024948024, + "grad_norm": 7.313511741813272e-05, + "learning_rate": 0.14182805296101172, + "loss": 0.2566, + "num_input_tokens_seen": 15782496, + "step": 20695 + }, + { + "epoch": 43.03534303534303, + "grad_norm": 0.0006190743879415095, + "learning_rate": 0.14176923621109272, + "loss": 0.264, + "num_input_tokens_seen": 15786368, + "step": 20700 + }, + { + "epoch": 43.04573804573805, + "grad_norm": 0.00037532905116677284, + "learning_rate": 0.14171042073046097, + "loss": 0.2925, + "num_input_tokens_seen": 15790144, + "step": 20705 + }, + { + "epoch": 43.056133056133056, + "grad_norm": 0.00034636480268090963, + "learning_rate": 0.14165160652818642, + "loss": 0.2753, + "num_input_tokens_seen": 15793984, + "step": 20710 + }, + { + "epoch": 43.066528066528065, + "grad_norm": 0.0009093452244997025, + "learning_rate": 0.14159279361333907, + "loss": 0.2687, + "num_input_tokens_seen": 15797760, + "step": 20715 + }, + { + "epoch": 43.07692307692308, + "grad_norm": 0.00019883294589817524, + "learning_rate": 0.14153398199498868, + "loss": 0.2828, + "num_input_tokens_seen": 15801600, + "step": 20720 + }, + { + "epoch": 43.08731808731809, + "grad_norm": 0.0004070269060321152, + "learning_rate": 0.14147517168220458, + "loss": 0.2499, + "num_input_tokens_seen": 15805216, + "step": 20725 + }, + { + "epoch": 43.0977130977131, + "grad_norm": 0.00012920625158585608, + "learning_rate": 0.14141636268405616, + "loss": 0.262, + "num_input_tokens_seen": 15808928, + "step": 20730 + }, + { + "epoch": 43.108108108108105, + "grad_norm": 5.8282905229134485e-05, + "learning_rate": 0.14135755500961253, + "loss": 0.2624, + "num_input_tokens_seen": 15812800, + "step": 20735 + }, + { + "epoch": 43.11850311850312, + "grad_norm": 0.00021145261416677386, + "learning_rate": 0.14129874866794245, + "loss": 0.2564, + "num_input_tokens_seen": 15816576, + "step": 20740 + }, + { + "epoch": 43.12889812889813, + "grad_norm": 0.00025781645672395825, + "learning_rate": 0.14123994366811476, + "loss": 0.2702, + "num_input_tokens_seen": 15820320, + "step": 20745 + }, + { + "epoch": 43.13929313929314, + "grad_norm": 0.0002806819975376129, + "learning_rate": 0.14118114001919774, + "loss": 0.2827, + "num_input_tokens_seen": 15824096, + "step": 20750 + }, + { + "epoch": 43.14968814968815, + "grad_norm": 0.00036128226201981306, + "learning_rate": 0.14112233773025978, + "loss": 0.2599, + "num_input_tokens_seen": 15827840, + "step": 20755 + }, + { + "epoch": 43.16008316008316, + "grad_norm": 0.00016363158647436649, + "learning_rate": 0.14106353681036896, + "loss": 0.2678, + "num_input_tokens_seen": 15831584, + "step": 20760 + }, + { + "epoch": 43.17047817047817, + "grad_norm": 0.00034647106076590717, + "learning_rate": 0.14100473726859303, + "loss": 0.2771, + "num_input_tokens_seen": 15835200, + "step": 20765 + }, + { + "epoch": 43.18087318087318, + "grad_norm": 0.00036992289824411273, + "learning_rate": 0.14094593911399964, + "loss": 0.2824, + "num_input_tokens_seen": 15839104, + "step": 20770 + }, + { + "epoch": 43.19126819126819, + "grad_norm": 0.00013782289170194417, + "learning_rate": 0.14088714235565625, + "loss": 0.2724, + "num_input_tokens_seen": 15843008, + "step": 20775 + }, + { + "epoch": 43.2016632016632, + "grad_norm": 0.00027012662030756474, + "learning_rate": 0.14082834700263, + "loss": 0.2624, + "num_input_tokens_seen": 15846752, + "step": 20780 + }, + { + "epoch": 43.21205821205821, + "grad_norm": 0.0004066957044415176, + "learning_rate": 0.14076955306398795, + "loss": 0.278, + "num_input_tokens_seen": 15850784, + "step": 20785 + }, + { + "epoch": 43.222453222453225, + "grad_norm": 0.0003724557172972709, + "learning_rate": 0.14071076054879675, + "loss": 0.2888, + "num_input_tokens_seen": 15854592, + "step": 20790 + }, + { + "epoch": 43.232848232848234, + "grad_norm": 0.00025916879530996084, + "learning_rate": 0.14065196946612302, + "loss": 0.2562, + "num_input_tokens_seen": 15858400, + "step": 20795 + }, + { + "epoch": 43.24324324324324, + "grad_norm": 9.486638009548187e-05, + "learning_rate": 0.1405931798250331, + "loss": 0.2508, + "num_input_tokens_seen": 15862336, + "step": 20800 + }, + { + "epoch": 43.24324324324324, + "eval_loss": 0.24934041500091553, + "eval_runtime": 13.4151, + "eval_samples_per_second": 63.809, + "eval_steps_per_second": 15.952, + "num_input_tokens_seen": 15862336, + "step": 20800 + }, + { + "epoch": 43.25363825363825, + "grad_norm": 0.0001949878642335534, + "learning_rate": 0.14053439163459308, + "loss": 0.2596, + "num_input_tokens_seen": 15866080, + "step": 20805 + }, + { + "epoch": 43.264033264033266, + "grad_norm": 6.169541302369907e-05, + "learning_rate": 0.14047560490386876, + "loss": 0.2953, + "num_input_tokens_seen": 15869952, + "step": 20810 + }, + { + "epoch": 43.274428274428274, + "grad_norm": 0.00015231332508847117, + "learning_rate": 0.14041681964192593, + "loss": 0.2465, + "num_input_tokens_seen": 15873792, + "step": 20815 + }, + { + "epoch": 43.28482328482328, + "grad_norm": 0.00011879668454639614, + "learning_rate": 0.14035803585782988, + "loss": 0.2698, + "num_input_tokens_seen": 15877728, + "step": 20820 + }, + { + "epoch": 43.2952182952183, + "grad_norm": 0.0001493665622547269, + "learning_rate": 0.14029925356064593, + "loss": 0.2731, + "num_input_tokens_seen": 15881504, + "step": 20825 + }, + { + "epoch": 43.305613305613306, + "grad_norm": 0.0001392518461216241, + "learning_rate": 0.1402404727594389, + "loss": 0.2603, + "num_input_tokens_seen": 15885280, + "step": 20830 + }, + { + "epoch": 43.316008316008315, + "grad_norm": 6.962760380702093e-05, + "learning_rate": 0.1401816934632737, + "loss": 0.2392, + "num_input_tokens_seen": 15889216, + "step": 20835 + }, + { + "epoch": 43.32640332640332, + "grad_norm": 0.00012231871369294822, + "learning_rate": 0.1401229156812147, + "loss": 0.2309, + "num_input_tokens_seen": 15892896, + "step": 20840 + }, + { + "epoch": 43.33679833679834, + "grad_norm": 0.00012839803821407259, + "learning_rate": 0.14006413942232626, + "loss": 0.2805, + "num_input_tokens_seen": 15896672, + "step": 20845 + }, + { + "epoch": 43.34719334719335, + "grad_norm": 0.00024317018687725067, + "learning_rate": 0.14000536469567235, + "loss": 0.2954, + "num_input_tokens_seen": 15900480, + "step": 20850 + }, + { + "epoch": 43.357588357588355, + "grad_norm": 8.986486500361934e-05, + "learning_rate": 0.13994659151031685, + "loss": 0.2819, + "num_input_tokens_seen": 15904288, + "step": 20855 + }, + { + "epoch": 43.36798336798337, + "grad_norm": 0.000595803139731288, + "learning_rate": 0.13988781987532323, + "loss": 0.2806, + "num_input_tokens_seen": 15908224, + "step": 20860 + }, + { + "epoch": 43.37837837837838, + "grad_norm": 7.580459350720048e-05, + "learning_rate": 0.1398290497997549, + "loss": 0.2667, + "num_input_tokens_seen": 15912064, + "step": 20865 + }, + { + "epoch": 43.38877338877339, + "grad_norm": 0.00041864762897603214, + "learning_rate": 0.13977028129267488, + "loss": 0.2654, + "num_input_tokens_seen": 15915712, + "step": 20870 + }, + { + "epoch": 43.3991683991684, + "grad_norm": 7.840335456421599e-05, + "learning_rate": 0.13971151436314605, + "loss": 0.2692, + "num_input_tokens_seen": 15919424, + "step": 20875 + }, + { + "epoch": 43.40956340956341, + "grad_norm": 0.00011510030890349299, + "learning_rate": 0.13965274902023103, + "loss": 0.2738, + "num_input_tokens_seen": 15923168, + "step": 20880 + }, + { + "epoch": 43.41995841995842, + "grad_norm": 7.706279575359076e-05, + "learning_rate": 0.13959398527299208, + "loss": 0.2366, + "num_input_tokens_seen": 15926976, + "step": 20885 + }, + { + "epoch": 43.43035343035343, + "grad_norm": 0.00017795112216845155, + "learning_rate": 0.13953522313049138, + "loss": 0.2525, + "num_input_tokens_seen": 15930656, + "step": 20890 + }, + { + "epoch": 43.44074844074844, + "grad_norm": 0.00012256923946551979, + "learning_rate": 0.13947646260179083, + "loss": 0.2791, + "num_input_tokens_seen": 15934368, + "step": 20895 + }, + { + "epoch": 43.45114345114345, + "grad_norm": 0.00025555898901075125, + "learning_rate": 0.13941770369595194, + "loss": 0.2681, + "num_input_tokens_seen": 15938144, + "step": 20900 + }, + { + "epoch": 43.46153846153846, + "grad_norm": 0.0002503453870303929, + "learning_rate": 0.1393589464220362, + "loss": 0.2565, + "num_input_tokens_seen": 15941824, + "step": 20905 + }, + { + "epoch": 43.471933471933475, + "grad_norm": 5.845542182214558e-05, + "learning_rate": 0.13930019078910455, + "loss": 0.2749, + "num_input_tokens_seen": 15945600, + "step": 20910 + }, + { + "epoch": 43.482328482328484, + "grad_norm": 0.0002480710099916905, + "learning_rate": 0.139241436806218, + "loss": 0.2758, + "num_input_tokens_seen": 15949440, + "step": 20915 + }, + { + "epoch": 43.49272349272349, + "grad_norm": 0.000213769351830706, + "learning_rate": 0.13918268448243712, + "loss": 0.2725, + "num_input_tokens_seen": 15953280, + "step": 20920 + }, + { + "epoch": 43.5031185031185, + "grad_norm": 0.00019795297703240067, + "learning_rate": 0.13912393382682217, + "loss": 0.249, + "num_input_tokens_seen": 15957152, + "step": 20925 + }, + { + "epoch": 43.513513513513516, + "grad_norm": 0.0003817606484517455, + "learning_rate": 0.1390651848484333, + "loss": 0.2943, + "num_input_tokens_seen": 15960896, + "step": 20930 + }, + { + "epoch": 43.523908523908524, + "grad_norm": 0.00038954117917455733, + "learning_rate": 0.1390064375563304, + "loss": 0.2826, + "num_input_tokens_seen": 15964832, + "step": 20935 + }, + { + "epoch": 43.53430353430353, + "grad_norm": 0.00027874193619936705, + "learning_rate": 0.13894769195957293, + "loss": 0.2716, + "num_input_tokens_seen": 15968608, + "step": 20940 + }, + { + "epoch": 43.54469854469855, + "grad_norm": 9.038222196977586e-05, + "learning_rate": 0.13888894806722032, + "loss": 0.2722, + "num_input_tokens_seen": 15972576, + "step": 20945 + }, + { + "epoch": 43.555093555093556, + "grad_norm": 0.00016321065777447075, + "learning_rate": 0.1388302058883315, + "loss": 0.2678, + "num_input_tokens_seen": 15976352, + "step": 20950 + }, + { + "epoch": 43.565488565488565, + "grad_norm": 6.552860577357933e-05, + "learning_rate": 0.13877146543196528, + "loss": 0.2513, + "num_input_tokens_seen": 15980256, + "step": 20955 + }, + { + "epoch": 43.57588357588357, + "grad_norm": 0.00022836010612081736, + "learning_rate": 0.13871272670718027, + "loss": 0.2828, + "num_input_tokens_seen": 15984096, + "step": 20960 + }, + { + "epoch": 43.58627858627859, + "grad_norm": 0.00011581848957575858, + "learning_rate": 0.13865398972303455, + "loss": 0.2695, + "num_input_tokens_seen": 15987840, + "step": 20965 + }, + { + "epoch": 43.5966735966736, + "grad_norm": 6.513256812468171e-05, + "learning_rate": 0.13859525448858623, + "loss": 0.2775, + "num_input_tokens_seen": 15991744, + "step": 20970 + }, + { + "epoch": 43.607068607068605, + "grad_norm": 0.00017235387349501252, + "learning_rate": 0.13853652101289304, + "loss": 0.2578, + "num_input_tokens_seen": 15995392, + "step": 20975 + }, + { + "epoch": 43.61746361746362, + "grad_norm": 0.00042952195508405566, + "learning_rate": 0.13847778930501234, + "loss": 0.2856, + "num_input_tokens_seen": 15999168, + "step": 20980 + }, + { + "epoch": 43.62785862785863, + "grad_norm": 3.493340773275122e-05, + "learning_rate": 0.1384190593740013, + "loss": 0.2557, + "num_input_tokens_seen": 16003008, + "step": 20985 + }, + { + "epoch": 43.63825363825364, + "grad_norm": 0.00024644704535603523, + "learning_rate": 0.13836033122891686, + "loss": 0.2892, + "num_input_tokens_seen": 16006784, + "step": 20990 + }, + { + "epoch": 43.648648648648646, + "grad_norm": 0.00036609030212275684, + "learning_rate": 0.1383016048788156, + "loss": 0.2563, + "num_input_tokens_seen": 16010496, + "step": 20995 + }, + { + "epoch": 43.65904365904366, + "grad_norm": 7.281095895450562e-05, + "learning_rate": 0.13824288033275392, + "loss": 0.2563, + "num_input_tokens_seen": 16014304, + "step": 21000 + }, + { + "epoch": 43.65904365904366, + "eval_loss": 0.24751318991184235, + "eval_runtime": 13.3865, + "eval_samples_per_second": 63.945, + "eval_steps_per_second": 15.986, + "num_input_tokens_seen": 16014304, + "step": 21000 + }, + { + "epoch": 43.66943866943867, + "grad_norm": 5.122081347508356e-05, + "learning_rate": 0.1381841575997878, + "loss": 0.2434, + "num_input_tokens_seen": 16017952, + "step": 21005 + }, + { + "epoch": 43.67983367983368, + "grad_norm": 6.415200914489105e-05, + "learning_rate": 0.13812543668897306, + "loss": 0.2609, + "num_input_tokens_seen": 16021760, + "step": 21010 + }, + { + "epoch": 43.69022869022869, + "grad_norm": 4.1278857679571956e-05, + "learning_rate": 0.13806671760936526, + "loss": 0.2855, + "num_input_tokens_seen": 16025696, + "step": 21015 + }, + { + "epoch": 43.7006237006237, + "grad_norm": 0.0004114708863198757, + "learning_rate": 0.13800800037001956, + "loss": 0.253, + "num_input_tokens_seen": 16029504, + "step": 21020 + }, + { + "epoch": 43.71101871101871, + "grad_norm": 0.0001820306497393176, + "learning_rate": 0.13794928497999087, + "loss": 0.27, + "num_input_tokens_seen": 16033376, + "step": 21025 + }, + { + "epoch": 43.72141372141372, + "grad_norm": 0.00010494939488125965, + "learning_rate": 0.1378905714483339, + "loss": 0.2495, + "num_input_tokens_seen": 16037120, + "step": 21030 + }, + { + "epoch": 43.731808731808734, + "grad_norm": 0.00015724134573247284, + "learning_rate": 0.13783185978410295, + "loss": 0.2367, + "num_input_tokens_seen": 16040928, + "step": 21035 + }, + { + "epoch": 43.74220374220374, + "grad_norm": 5.533166040549986e-05, + "learning_rate": 0.13777314999635218, + "loss": 0.269, + "num_input_tokens_seen": 16044512, + "step": 21040 + }, + { + "epoch": 43.75259875259875, + "grad_norm": 0.0002520230773370713, + "learning_rate": 0.1377144420941353, + "loss": 0.2852, + "num_input_tokens_seen": 16048320, + "step": 21045 + }, + { + "epoch": 43.762993762993766, + "grad_norm": 5.8905341575155035e-05, + "learning_rate": 0.13765573608650586, + "loss": 0.2635, + "num_input_tokens_seen": 16051904, + "step": 21050 + }, + { + "epoch": 43.773388773388774, + "grad_norm": 0.00038820350891910493, + "learning_rate": 0.13759703198251702, + "loss": 0.2594, + "num_input_tokens_seen": 16055744, + "step": 21055 + }, + { + "epoch": 43.78378378378378, + "grad_norm": 0.00010310626385034993, + "learning_rate": 0.13753832979122174, + "loss": 0.2633, + "num_input_tokens_seen": 16059584, + "step": 21060 + }, + { + "epoch": 43.79417879417879, + "grad_norm": 0.00019183687982149422, + "learning_rate": 0.13747962952167264, + "loss": 0.262, + "num_input_tokens_seen": 16063296, + "step": 21065 + }, + { + "epoch": 43.804573804573806, + "grad_norm": 0.0001501396909588948, + "learning_rate": 0.13742093118292192, + "loss": 0.2561, + "num_input_tokens_seen": 16067040, + "step": 21070 + }, + { + "epoch": 43.814968814968815, + "grad_norm": 0.00017984324949793518, + "learning_rate": 0.13736223478402174, + "loss": 0.2413, + "num_input_tokens_seen": 16070816, + "step": 21075 + }, + { + "epoch": 43.82536382536382, + "grad_norm": 8.610917575424537e-05, + "learning_rate": 0.1373035403340238, + "loss": 0.2857, + "num_input_tokens_seen": 16074688, + "step": 21080 + }, + { + "epoch": 43.83575883575884, + "grad_norm": 0.0003861830336973071, + "learning_rate": 0.13724484784197943, + "loss": 0.2629, + "num_input_tokens_seen": 16078624, + "step": 21085 + }, + { + "epoch": 43.84615384615385, + "grad_norm": 0.00041653879452496767, + "learning_rate": 0.13718615731693987, + "loss": 0.2679, + "num_input_tokens_seen": 16082464, + "step": 21090 + }, + { + "epoch": 43.856548856548855, + "grad_norm": 0.0006006438052281737, + "learning_rate": 0.13712746876795587, + "loss": 0.249, + "num_input_tokens_seen": 16086272, + "step": 21095 + }, + { + "epoch": 43.86694386694387, + "grad_norm": 0.0001967371063074097, + "learning_rate": 0.13706878220407792, + "loss": 0.2672, + "num_input_tokens_seen": 16090112, + "step": 21100 + }, + { + "epoch": 43.87733887733888, + "grad_norm": 0.00026333611458539963, + "learning_rate": 0.13701009763435631, + "loss": 0.2522, + "num_input_tokens_seen": 16094048, + "step": 21105 + }, + { + "epoch": 43.88773388773389, + "grad_norm": 8.01010537543334e-05, + "learning_rate": 0.13695141506784084, + "loss": 0.2892, + "num_input_tokens_seen": 16097952, + "step": 21110 + }, + { + "epoch": 43.898128898128896, + "grad_norm": 0.000379104953026399, + "learning_rate": 0.13689273451358114, + "loss": 0.2573, + "num_input_tokens_seen": 16101888, + "step": 21115 + }, + { + "epoch": 43.90852390852391, + "grad_norm": 4.9818871048046276e-05, + "learning_rate": 0.13683405598062653, + "loss": 0.2506, + "num_input_tokens_seen": 16105600, + "step": 21120 + }, + { + "epoch": 43.91891891891892, + "grad_norm": 0.0002987574553117156, + "learning_rate": 0.1367753794780259, + "loss": 0.281, + "num_input_tokens_seen": 16109568, + "step": 21125 + }, + { + "epoch": 43.92931392931393, + "grad_norm": 0.0002677681914065033, + "learning_rate": 0.13671670501482802, + "loss": 0.2504, + "num_input_tokens_seen": 16113696, + "step": 21130 + }, + { + "epoch": 43.93970893970894, + "grad_norm": 0.0002804089745040983, + "learning_rate": 0.1366580326000811, + "loss": 0.2548, + "num_input_tokens_seen": 16117632, + "step": 21135 + }, + { + "epoch": 43.95010395010395, + "grad_norm": 0.0002980169956572354, + "learning_rate": 0.1365993622428332, + "loss": 0.2743, + "num_input_tokens_seen": 16121504, + "step": 21140 + }, + { + "epoch": 43.96049896049896, + "grad_norm": 0.00032101667602546513, + "learning_rate": 0.13654069395213211, + "loss": 0.2769, + "num_input_tokens_seen": 16125312, + "step": 21145 + }, + { + "epoch": 43.97089397089397, + "grad_norm": 0.0002919249818660319, + "learning_rate": 0.13648202773702509, + "loss": 0.2369, + "num_input_tokens_seen": 16129056, + "step": 21150 + }, + { + "epoch": 43.981288981288984, + "grad_norm": 0.00021261710207909346, + "learning_rate": 0.13642336360655927, + "loss": 0.2511, + "num_input_tokens_seen": 16132992, + "step": 21155 + }, + { + "epoch": 43.99168399168399, + "grad_norm": 0.00012051967496518046, + "learning_rate": 0.13636470156978145, + "loss": 0.2674, + "num_input_tokens_seen": 16136736, + "step": 21160 + }, + { + "epoch": 44.002079002079, + "grad_norm": 7.050829299259931e-05, + "learning_rate": 0.13630604163573798, + "loss": 0.2782, + "num_input_tokens_seen": 16140408, + "step": 21165 + }, + { + "epoch": 44.012474012474016, + "grad_norm": 0.0005001010722480714, + "learning_rate": 0.13624738381347495, + "loss": 0.2609, + "num_input_tokens_seen": 16144120, + "step": 21170 + }, + { + "epoch": 44.022869022869024, + "grad_norm": 0.00011979175178566948, + "learning_rate": 0.1361887281120382, + "loss": 0.2605, + "num_input_tokens_seen": 16147928, + "step": 21175 + }, + { + "epoch": 44.03326403326403, + "grad_norm": 0.00019757766858674586, + "learning_rate": 0.13613007454047307, + "loss": 0.2593, + "num_input_tokens_seen": 16151768, + "step": 21180 + }, + { + "epoch": 44.04365904365904, + "grad_norm": 0.0002386389533057809, + "learning_rate": 0.13607142310782486, + "loss": 0.2648, + "num_input_tokens_seen": 16155448, + "step": 21185 + }, + { + "epoch": 44.054054054054056, + "grad_norm": 9.231379226548597e-05, + "learning_rate": 0.13601277382313814, + "loss": 0.2665, + "num_input_tokens_seen": 16159128, + "step": 21190 + }, + { + "epoch": 44.064449064449065, + "grad_norm": 0.00022090598940849304, + "learning_rate": 0.1359541266954575, + "loss": 0.2378, + "num_input_tokens_seen": 16162872, + "step": 21195 + }, + { + "epoch": 44.07484407484407, + "grad_norm": 0.00014628010103479028, + "learning_rate": 0.13589548173382707, + "loss": 0.2613, + "num_input_tokens_seen": 16166680, + "step": 21200 + }, + { + "epoch": 44.07484407484407, + "eval_loss": 0.24760009348392487, + "eval_runtime": 13.3812, + "eval_samples_per_second": 63.97, + "eval_steps_per_second": 15.993, + "num_input_tokens_seen": 16166680, + "step": 21200 + }, + { + "epoch": 44.08523908523909, + "grad_norm": 0.00021189768449403346, + "learning_rate": 0.1358368389472906, + "loss": 0.2681, + "num_input_tokens_seen": 16170616, + "step": 21205 + }, + { + "epoch": 44.0956340956341, + "grad_norm": 0.0001461546344216913, + "learning_rate": 0.13577819834489155, + "loss": 0.2629, + "num_input_tokens_seen": 16174680, + "step": 21210 + }, + { + "epoch": 44.106029106029105, + "grad_norm": 0.0001874374720500782, + "learning_rate": 0.135719559935673, + "loss": 0.2433, + "num_input_tokens_seen": 16178488, + "step": 21215 + }, + { + "epoch": 44.11642411642411, + "grad_norm": 0.00021781085524708033, + "learning_rate": 0.13566092372867775, + "loss": 0.2846, + "num_input_tokens_seen": 16182392, + "step": 21220 + }, + { + "epoch": 44.12681912681913, + "grad_norm": 0.0004241634451318532, + "learning_rate": 0.13560228973294833, + "loss": 0.2877, + "num_input_tokens_seen": 16186264, + "step": 21225 + }, + { + "epoch": 44.13721413721414, + "grad_norm": 0.00023653804964851588, + "learning_rate": 0.13554365795752668, + "loss": 0.2609, + "num_input_tokens_seen": 16190040, + "step": 21230 + }, + { + "epoch": 44.147609147609145, + "grad_norm": 0.00027910235803574324, + "learning_rate": 0.1354850284114547, + "loss": 0.271, + "num_input_tokens_seen": 16193848, + "step": 21235 + }, + { + "epoch": 44.15800415800416, + "grad_norm": 0.0005078093381598592, + "learning_rate": 0.13542640110377374, + "loss": 0.2901, + "num_input_tokens_seen": 16197816, + "step": 21240 + }, + { + "epoch": 44.16839916839917, + "grad_norm": 0.00020517937082331628, + "learning_rate": 0.13536777604352487, + "loss": 0.2488, + "num_input_tokens_seen": 16201528, + "step": 21245 + }, + { + "epoch": 44.17879417879418, + "grad_norm": 0.00017930624017026275, + "learning_rate": 0.13530915323974887, + "loss": 0.2511, + "num_input_tokens_seen": 16205432, + "step": 21250 + }, + { + "epoch": 44.189189189189186, + "grad_norm": 0.00025428240769542754, + "learning_rate": 0.13525053270148596, + "loss": 0.2682, + "num_input_tokens_seen": 16209272, + "step": 21255 + }, + { + "epoch": 44.1995841995842, + "grad_norm": 0.0003311745822429657, + "learning_rate": 0.13519191443777628, + "loss": 0.2715, + "num_input_tokens_seen": 16213144, + "step": 21260 + }, + { + "epoch": 44.20997920997921, + "grad_norm": 0.0004858675820287317, + "learning_rate": 0.13513329845765953, + "loss": 0.2344, + "num_input_tokens_seen": 16216984, + "step": 21265 + }, + { + "epoch": 44.22037422037422, + "grad_norm": 0.000263742811512202, + "learning_rate": 0.13507468477017495, + "loss": 0.2783, + "num_input_tokens_seen": 16220728, + "step": 21270 + }, + { + "epoch": 44.23076923076923, + "grad_norm": 7.299939898075536e-05, + "learning_rate": 0.13501607338436153, + "loss": 0.2342, + "num_input_tokens_seen": 16224632, + "step": 21275 + }, + { + "epoch": 44.24116424116424, + "grad_norm": 0.0003884996403940022, + "learning_rate": 0.13495746430925798, + "loss": 0.2751, + "num_input_tokens_seen": 16228504, + "step": 21280 + }, + { + "epoch": 44.25155925155925, + "grad_norm": 0.00044079963117837906, + "learning_rate": 0.13489885755390238, + "loss": 0.2712, + "num_input_tokens_seen": 16232216, + "step": 21285 + }, + { + "epoch": 44.26195426195426, + "grad_norm": 4.1018534830072895e-05, + "learning_rate": 0.13484025312733275, + "loss": 0.2783, + "num_input_tokens_seen": 16235928, + "step": 21290 + }, + { + "epoch": 44.272349272349274, + "grad_norm": 0.00010333069803891703, + "learning_rate": 0.13478165103858658, + "loss": 0.268, + "num_input_tokens_seen": 16239832, + "step": 21295 + }, + { + "epoch": 44.28274428274428, + "grad_norm": 7.203936547739431e-05, + "learning_rate": 0.13472305129670106, + "loss": 0.2615, + "num_input_tokens_seen": 16243736, + "step": 21300 + }, + { + "epoch": 44.29313929313929, + "grad_norm": 0.00014360516797751188, + "learning_rate": 0.13466445391071305, + "loss": 0.2603, + "num_input_tokens_seen": 16247512, + "step": 21305 + }, + { + "epoch": 44.303534303534306, + "grad_norm": 0.00028150840080343187, + "learning_rate": 0.13460585888965895, + "loss": 0.2603, + "num_input_tokens_seen": 16251352, + "step": 21310 + }, + { + "epoch": 44.313929313929314, + "grad_norm": 0.0003955138963647187, + "learning_rate": 0.13454726624257482, + "loss": 0.2977, + "num_input_tokens_seen": 16255096, + "step": 21315 + }, + { + "epoch": 44.32432432432432, + "grad_norm": 4.251376958563924e-05, + "learning_rate": 0.1344886759784965, + "loss": 0.2766, + "num_input_tokens_seen": 16258936, + "step": 21320 + }, + { + "epoch": 44.33471933471934, + "grad_norm": 0.0001331439707428217, + "learning_rate": 0.13443008810645923, + "loss": 0.2738, + "num_input_tokens_seen": 16262680, + "step": 21325 + }, + { + "epoch": 44.34511434511435, + "grad_norm": 6.833983934484422e-05, + "learning_rate": 0.13437150263549807, + "loss": 0.2508, + "num_input_tokens_seen": 16266296, + "step": 21330 + }, + { + "epoch": 44.355509355509355, + "grad_norm": 0.0004538811044767499, + "learning_rate": 0.13431291957464755, + "loss": 0.266, + "num_input_tokens_seen": 16270200, + "step": 21335 + }, + { + "epoch": 44.36590436590436, + "grad_norm": 0.00030418054666370153, + "learning_rate": 0.13425433893294197, + "loss": 0.3076, + "num_input_tokens_seen": 16274104, + "step": 21340 + }, + { + "epoch": 44.37629937629938, + "grad_norm": 0.00047397197340615094, + "learning_rate": 0.13419576071941525, + "loss": 0.285, + "num_input_tokens_seen": 16278008, + "step": 21345 + }, + { + "epoch": 44.38669438669439, + "grad_norm": 2.938338184321765e-05, + "learning_rate": 0.1341371849431008, + "loss": 0.2723, + "num_input_tokens_seen": 16281912, + "step": 21350 + }, + { + "epoch": 44.397089397089395, + "grad_norm": 0.00019900860206689686, + "learning_rate": 0.13407861161303178, + "loss": 0.2602, + "num_input_tokens_seen": 16285784, + "step": 21355 + }, + { + "epoch": 44.40748440748441, + "grad_norm": 5.167853305465542e-05, + "learning_rate": 0.13402004073824098, + "loss": 0.2354, + "num_input_tokens_seen": 16289528, + "step": 21360 + }, + { + "epoch": 44.41787941787942, + "grad_norm": 0.0004037044127471745, + "learning_rate": 0.13396147232776062, + "loss": 0.2468, + "num_input_tokens_seen": 16293368, + "step": 21365 + }, + { + "epoch": 44.42827442827443, + "grad_norm": 0.00038667209446430206, + "learning_rate": 0.13390290639062288, + "loss": 0.2911, + "num_input_tokens_seen": 16297112, + "step": 21370 + }, + { + "epoch": 44.438669438669436, + "grad_norm": 3.342255877214484e-05, + "learning_rate": 0.13384434293585917, + "loss": 0.2607, + "num_input_tokens_seen": 16301048, + "step": 21375 + }, + { + "epoch": 44.44906444906445, + "grad_norm": 0.0001605077413842082, + "learning_rate": 0.13378578197250088, + "loss": 0.2571, + "num_input_tokens_seen": 16305016, + "step": 21380 + }, + { + "epoch": 44.45945945945946, + "grad_norm": 0.00023463493562303483, + "learning_rate": 0.13372722350957872, + "loss": 0.269, + "num_input_tokens_seen": 16308952, + "step": 21385 + }, + { + "epoch": 44.46985446985447, + "grad_norm": 7.709422061452642e-05, + "learning_rate": 0.13366866755612322, + "loss": 0.271, + "num_input_tokens_seen": 16312696, + "step": 21390 + }, + { + "epoch": 44.48024948024948, + "grad_norm": 9.872400551103055e-05, + "learning_rate": 0.13361011412116436, + "loss": 0.274, + "num_input_tokens_seen": 16316504, + "step": 21395 + }, + { + "epoch": 44.49064449064449, + "grad_norm": 4.162257755524479e-05, + "learning_rate": 0.13355156321373196, + "loss": 0.2594, + "num_input_tokens_seen": 16320408, + "step": 21400 + }, + { + "epoch": 44.49064449064449, + "eval_loss": 0.24985271692276, + "eval_runtime": 13.3843, + "eval_samples_per_second": 63.955, + "eval_steps_per_second": 15.989, + "num_input_tokens_seen": 16320408, + "step": 21400 + }, + { + "epoch": 44.5010395010395, + "grad_norm": 9.209149720845744e-05, + "learning_rate": 0.13349301484285514, + "loss": 0.2616, + "num_input_tokens_seen": 16324344, + "step": 21405 + }, + { + "epoch": 44.51143451143451, + "grad_norm": 0.00010198220843449235, + "learning_rate": 0.13343446901756295, + "loss": 0.2652, + "num_input_tokens_seen": 16328152, + "step": 21410 + }, + { + "epoch": 44.521829521829524, + "grad_norm": 0.0003755299549084157, + "learning_rate": 0.13337592574688376, + "loss": 0.2542, + "num_input_tokens_seen": 16332024, + "step": 21415 + }, + { + "epoch": 44.53222453222453, + "grad_norm": 0.00013641236000694335, + "learning_rate": 0.13331738503984572, + "loss": 0.2693, + "num_input_tokens_seen": 16335768, + "step": 21420 + }, + { + "epoch": 44.54261954261954, + "grad_norm": 0.00010613301128614694, + "learning_rate": 0.1332588469054766, + "loss": 0.2777, + "num_input_tokens_seen": 16339544, + "step": 21425 + }, + { + "epoch": 44.553014553014556, + "grad_norm": 5.538835466722958e-05, + "learning_rate": 0.1332003113528036, + "loss": 0.263, + "num_input_tokens_seen": 16343384, + "step": 21430 + }, + { + "epoch": 44.563409563409564, + "grad_norm": 0.00014307141827885062, + "learning_rate": 0.13314177839085373, + "loss": 0.2728, + "num_input_tokens_seen": 16347160, + "step": 21435 + }, + { + "epoch": 44.57380457380457, + "grad_norm": 4.037943290313706e-05, + "learning_rate": 0.13308324802865354, + "loss": 0.2579, + "num_input_tokens_seen": 16350872, + "step": 21440 + }, + { + "epoch": 44.58419958419958, + "grad_norm": 4.118874130654149e-05, + "learning_rate": 0.13302472027522905, + "loss": 0.2565, + "num_input_tokens_seen": 16354552, + "step": 21445 + }, + { + "epoch": 44.5945945945946, + "grad_norm": 0.00016343964671250433, + "learning_rate": 0.13296619513960606, + "loss": 0.2461, + "num_input_tokens_seen": 16358360, + "step": 21450 + }, + { + "epoch": 44.604989604989605, + "grad_norm": 0.0004469746199902147, + "learning_rate": 0.1329076726308098, + "loss": 0.2681, + "num_input_tokens_seen": 16362168, + "step": 21455 + }, + { + "epoch": 44.61538461538461, + "grad_norm": 0.0005577249685302377, + "learning_rate": 0.13284915275786519, + "loss": 0.2872, + "num_input_tokens_seen": 16365848, + "step": 21460 + }, + { + "epoch": 44.62577962577963, + "grad_norm": 0.00013743994350079447, + "learning_rate": 0.1327906355297968, + "loss": 0.2801, + "num_input_tokens_seen": 16369688, + "step": 21465 + }, + { + "epoch": 44.63617463617464, + "grad_norm": 0.00011450327292550355, + "learning_rate": 0.13273212095562867, + "loss": 0.2761, + "num_input_tokens_seen": 16373272, + "step": 21470 + }, + { + "epoch": 44.646569646569645, + "grad_norm": 0.00020977872190997005, + "learning_rate": 0.13267360904438444, + "loss": 0.2716, + "num_input_tokens_seen": 16376920, + "step": 21475 + }, + { + "epoch": 44.656964656964654, + "grad_norm": 8.59934079926461e-05, + "learning_rate": 0.1326150998050875, + "loss": 0.2549, + "num_input_tokens_seen": 16380792, + "step": 21480 + }, + { + "epoch": 44.66735966735967, + "grad_norm": 0.0004057687474414706, + "learning_rate": 0.1325565932467606, + "loss": 0.2594, + "num_input_tokens_seen": 16384472, + "step": 21485 + }, + { + "epoch": 44.67775467775468, + "grad_norm": 0.0005032027256675065, + "learning_rate": 0.13249808937842628, + "loss": 0.2729, + "num_input_tokens_seen": 16388408, + "step": 21490 + }, + { + "epoch": 44.688149688149686, + "grad_norm": 0.00038425010279752314, + "learning_rate": 0.1324395882091065, + "loss": 0.2492, + "num_input_tokens_seen": 16392152, + "step": 21495 + }, + { + "epoch": 44.6985446985447, + "grad_norm": 0.000612241739872843, + "learning_rate": 0.13238108974782284, + "loss": 0.2765, + "num_input_tokens_seen": 16395960, + "step": 21500 + }, + { + "epoch": 44.70893970893971, + "grad_norm": 0.00022954182350076735, + "learning_rate": 0.13232259400359664, + "loss": 0.2627, + "num_input_tokens_seen": 16399832, + "step": 21505 + }, + { + "epoch": 44.71933471933472, + "grad_norm": 7.659868424525484e-05, + "learning_rate": 0.13226410098544852, + "loss": 0.2694, + "num_input_tokens_seen": 16403704, + "step": 21510 + }, + { + "epoch": 44.729729729729726, + "grad_norm": 0.0003492979740258306, + "learning_rate": 0.13220561070239892, + "loss": 0.2645, + "num_input_tokens_seen": 16407736, + "step": 21515 + }, + { + "epoch": 44.74012474012474, + "grad_norm": 0.00045646459329873323, + "learning_rate": 0.13214712316346783, + "loss": 0.2707, + "num_input_tokens_seen": 16411576, + "step": 21520 + }, + { + "epoch": 44.75051975051975, + "grad_norm": 0.00022545094543602318, + "learning_rate": 0.13208863837767465, + "loss": 0.2708, + "num_input_tokens_seen": 16415384, + "step": 21525 + }, + { + "epoch": 44.76091476091476, + "grad_norm": 0.00014819574425928295, + "learning_rate": 0.13203015635403856, + "loss": 0.2784, + "num_input_tokens_seen": 16419448, + "step": 21530 + }, + { + "epoch": 44.771309771309774, + "grad_norm": 0.00037124508526176214, + "learning_rate": 0.13197167710157817, + "loss": 0.2796, + "num_input_tokens_seen": 16423416, + "step": 21535 + }, + { + "epoch": 44.78170478170478, + "grad_norm": 0.0003178407496307045, + "learning_rate": 0.13191320062931167, + "loss": 0.2702, + "num_input_tokens_seen": 16427192, + "step": 21540 + }, + { + "epoch": 44.79209979209979, + "grad_norm": 0.0001378702581860125, + "learning_rate": 0.13185472694625702, + "loss": 0.2714, + "num_input_tokens_seen": 16431000, + "step": 21545 + }, + { + "epoch": 44.802494802494806, + "grad_norm": 2.3721042452962138e-05, + "learning_rate": 0.13179625606143142, + "loss": 0.2677, + "num_input_tokens_seen": 16434904, + "step": 21550 + }, + { + "epoch": 44.812889812889814, + "grad_norm": 0.0001865524100139737, + "learning_rate": 0.13173778798385188, + "loss": 0.2581, + "num_input_tokens_seen": 16438776, + "step": 21555 + }, + { + "epoch": 44.82328482328482, + "grad_norm": 0.00039999911678023636, + "learning_rate": 0.13167932272253505, + "loss": 0.2577, + "num_input_tokens_seen": 16442584, + "step": 21560 + }, + { + "epoch": 44.83367983367983, + "grad_norm": 0.00036108295898884535, + "learning_rate": 0.1316208602864968, + "loss": 0.2737, + "num_input_tokens_seen": 16446488, + "step": 21565 + }, + { + "epoch": 44.84407484407485, + "grad_norm": 0.00026716841966845095, + "learning_rate": 0.13156240068475292, + "loss": 0.2437, + "num_input_tokens_seen": 16450200, + "step": 21570 + }, + { + "epoch": 44.854469854469855, + "grad_norm": 0.00023703456099610776, + "learning_rate": 0.1315039439263185, + "loss": 0.2527, + "num_input_tokens_seen": 16454008, + "step": 21575 + }, + { + "epoch": 44.86486486486486, + "grad_norm": 9.033113019540906e-05, + "learning_rate": 0.13144549002020833, + "loss": 0.2717, + "num_input_tokens_seen": 16457944, + "step": 21580 + }, + { + "epoch": 44.87525987525988, + "grad_norm": 0.00025927714887075126, + "learning_rate": 0.13138703897543688, + "loss": 0.2546, + "num_input_tokens_seen": 16461592, + "step": 21585 + }, + { + "epoch": 44.88565488565489, + "grad_norm": 0.0004093958414159715, + "learning_rate": 0.1313285908010178, + "loss": 0.2506, + "num_input_tokens_seen": 16465336, + "step": 21590 + }, + { + "epoch": 44.896049896049895, + "grad_norm": 2.9531140171457082e-05, + "learning_rate": 0.13127014550596475, + "loss": 0.2629, + "num_input_tokens_seen": 16468952, + "step": 21595 + }, + { + "epoch": 44.906444906444904, + "grad_norm": 0.0003926590725313872, + "learning_rate": 0.1312117030992906, + "loss": 0.2615, + "num_input_tokens_seen": 16472888, + "step": 21600 + }, + { + "epoch": 44.906444906444904, + "eval_loss": 0.2488902509212494, + "eval_runtime": 13.3818, + "eval_samples_per_second": 63.967, + "eval_steps_per_second": 15.992, + "num_input_tokens_seen": 16472888, + "step": 21600 + }, + { + "epoch": 44.91683991683992, + "grad_norm": 0.0001376639847876504, + "learning_rate": 0.13115326359000795, + "loss": 0.286, + "num_input_tokens_seen": 16476664, + "step": 21605 + }, + { + "epoch": 44.92723492723493, + "grad_norm": 7.41363619454205e-05, + "learning_rate": 0.13109482698712896, + "loss": 0.2699, + "num_input_tokens_seen": 16480344, + "step": 21610 + }, + { + "epoch": 44.937629937629936, + "grad_norm": 0.00011974335939157754, + "learning_rate": 0.1310363932996651, + "loss": 0.2781, + "num_input_tokens_seen": 16484184, + "step": 21615 + }, + { + "epoch": 44.94802494802495, + "grad_norm": 0.0003414248349145055, + "learning_rate": 0.13097796253662775, + "loss": 0.2584, + "num_input_tokens_seen": 16488088, + "step": 21620 + }, + { + "epoch": 44.95841995841996, + "grad_norm": 0.00020359441987238824, + "learning_rate": 0.1309195347070277, + "loss": 0.282, + "num_input_tokens_seen": 16491928, + "step": 21625 + }, + { + "epoch": 44.96881496881497, + "grad_norm": 0.0002986837353091687, + "learning_rate": 0.13086110981987506, + "loss": 0.2529, + "num_input_tokens_seen": 16495672, + "step": 21630 + }, + { + "epoch": 44.979209979209976, + "grad_norm": 0.0004336940182838589, + "learning_rate": 0.13080268788417987, + "loss": 0.2697, + "num_input_tokens_seen": 16499544, + "step": 21635 + }, + { + "epoch": 44.98960498960499, + "grad_norm": 0.00025523744989186525, + "learning_rate": 0.1307442689089515, + "loss": 0.2793, + "num_input_tokens_seen": 16503288, + "step": 21640 + }, + { + "epoch": 45.0, + "grad_norm": 7.988721336005256e-05, + "learning_rate": 0.13068585290319873, + "loss": 0.2803, + "num_input_tokens_seen": 16507056, + "step": 21645 + }, + { + "epoch": 45.01039501039501, + "grad_norm": 0.000308271060930565, + "learning_rate": 0.13062743987593026, + "loss": 0.2599, + "num_input_tokens_seen": 16510864, + "step": 21650 + }, + { + "epoch": 45.020790020790024, + "grad_norm": 0.0004340185842011124, + "learning_rate": 0.13056902983615395, + "loss": 0.269, + "num_input_tokens_seen": 16514640, + "step": 21655 + }, + { + "epoch": 45.03118503118503, + "grad_norm": 4.979573714081198e-05, + "learning_rate": 0.13051062279287742, + "loss": 0.2588, + "num_input_tokens_seen": 16518512, + "step": 21660 + }, + { + "epoch": 45.04158004158004, + "grad_norm": 0.00011068989988416433, + "learning_rate": 0.13045221875510782, + "loss": 0.261, + "num_input_tokens_seen": 16522320, + "step": 21665 + }, + { + "epoch": 45.05197505197505, + "grad_norm": 7.1961541834753e-05, + "learning_rate": 0.13039381773185174, + "loss": 0.2871, + "num_input_tokens_seen": 16526096, + "step": 21670 + }, + { + "epoch": 45.062370062370064, + "grad_norm": 0.0006801162380725145, + "learning_rate": 0.1303354197321153, + "loss": 0.2508, + "num_input_tokens_seen": 16529936, + "step": 21675 + }, + { + "epoch": 45.07276507276507, + "grad_norm": 0.00022337948030326515, + "learning_rate": 0.13027702476490433, + "loss": 0.2663, + "num_input_tokens_seen": 16533712, + "step": 21680 + }, + { + "epoch": 45.08316008316008, + "grad_norm": 8.445126150036231e-05, + "learning_rate": 0.1302186328392239, + "loss": 0.2756, + "num_input_tokens_seen": 16537616, + "step": 21685 + }, + { + "epoch": 45.093555093555096, + "grad_norm": 0.00040929674287326634, + "learning_rate": 0.130160243964079, + "loss": 0.2762, + "num_input_tokens_seen": 16541520, + "step": 21690 + }, + { + "epoch": 45.103950103950105, + "grad_norm": 0.00021836582163814455, + "learning_rate": 0.13010185814847372, + "loss": 0.2717, + "num_input_tokens_seen": 16545296, + "step": 21695 + }, + { + "epoch": 45.11434511434511, + "grad_norm": 0.0003535347932484001, + "learning_rate": 0.13004347540141192, + "loss": 0.2828, + "num_input_tokens_seen": 16549264, + "step": 21700 + }, + { + "epoch": 45.12474012474012, + "grad_norm": 0.00041933933971449733, + "learning_rate": 0.12998509573189712, + "loss": 0.2801, + "num_input_tokens_seen": 16553072, + "step": 21705 + }, + { + "epoch": 45.13513513513514, + "grad_norm": 0.0001975226477952674, + "learning_rate": 0.12992671914893203, + "loss": 0.2754, + "num_input_tokens_seen": 16556880, + "step": 21710 + }, + { + "epoch": 45.145530145530145, + "grad_norm": 0.0002361765509704128, + "learning_rate": 0.12986834566151909, + "loss": 0.2671, + "num_input_tokens_seen": 16560656, + "step": 21715 + }, + { + "epoch": 45.15592515592515, + "grad_norm": 7.07785293343477e-05, + "learning_rate": 0.12980997527866028, + "loss": 0.2707, + "num_input_tokens_seen": 16564528, + "step": 21720 + }, + { + "epoch": 45.16632016632017, + "grad_norm": 0.00020954417414031923, + "learning_rate": 0.12975160800935692, + "loss": 0.2843, + "num_input_tokens_seen": 16568336, + "step": 21725 + }, + { + "epoch": 45.17671517671518, + "grad_norm": 0.00036131739034317434, + "learning_rate": 0.12969324386261016, + "loss": 0.2696, + "num_input_tokens_seen": 16572240, + "step": 21730 + }, + { + "epoch": 45.187110187110186, + "grad_norm": 0.00024694547755643725, + "learning_rate": 0.12963488284742034, + "loss": 0.2597, + "num_input_tokens_seen": 16576080, + "step": 21735 + }, + { + "epoch": 45.197505197505194, + "grad_norm": 0.0002636623103171587, + "learning_rate": 0.12957652497278752, + "loss": 0.2259, + "num_input_tokens_seen": 16579792, + "step": 21740 + }, + { + "epoch": 45.20790020790021, + "grad_norm": 0.00033735408214852214, + "learning_rate": 0.12951817024771117, + "loss": 0.2546, + "num_input_tokens_seen": 16583504, + "step": 21745 + }, + { + "epoch": 45.21829521829522, + "grad_norm": 0.0004906202084384859, + "learning_rate": 0.12945981868119041, + "loss": 0.3106, + "num_input_tokens_seen": 16587376, + "step": 21750 + }, + { + "epoch": 45.228690228690226, + "grad_norm": 0.00024589296663179994, + "learning_rate": 0.12940147028222376, + "loss": 0.2801, + "num_input_tokens_seen": 16591248, + "step": 21755 + }, + { + "epoch": 45.23908523908524, + "grad_norm": 0.0003033537359442562, + "learning_rate": 0.12934312505980916, + "loss": 0.2627, + "num_input_tokens_seen": 16595088, + "step": 21760 + }, + { + "epoch": 45.24948024948025, + "grad_norm": 0.0005645820056088269, + "learning_rate": 0.1292847830229443, + "loss": 0.2442, + "num_input_tokens_seen": 16598992, + "step": 21765 + }, + { + "epoch": 45.25987525987526, + "grad_norm": 0.0002161420852644369, + "learning_rate": 0.12922644418062626, + "loss": 0.2772, + "num_input_tokens_seen": 16602640, + "step": 21770 + }, + { + "epoch": 45.270270270270274, + "grad_norm": 0.00021539830777328461, + "learning_rate": 0.1291681085418515, + "loss": 0.2748, + "num_input_tokens_seen": 16606480, + "step": 21775 + }, + { + "epoch": 45.28066528066528, + "grad_norm": 0.0001854811271186918, + "learning_rate": 0.12910977611561628, + "loss": 0.2613, + "num_input_tokens_seen": 16610448, + "step": 21780 + }, + { + "epoch": 45.29106029106029, + "grad_norm": 0.00014989111514296383, + "learning_rate": 0.1290514469109161, + "loss": 0.2737, + "num_input_tokens_seen": 16614192, + "step": 21785 + }, + { + "epoch": 45.3014553014553, + "grad_norm": 0.00013945419050287455, + "learning_rate": 0.128993120936746, + "loss": 0.2496, + "num_input_tokens_seen": 16618096, + "step": 21790 + }, + { + "epoch": 45.311850311850314, + "grad_norm": 0.00015988264931365848, + "learning_rate": 0.12893479820210071, + "loss": 0.2831, + "num_input_tokens_seen": 16621840, + "step": 21795 + }, + { + "epoch": 45.32224532224532, + "grad_norm": 0.00027094673714600503, + "learning_rate": 0.1288764787159742, + "loss": 0.2551, + "num_input_tokens_seen": 16625808, + "step": 21800 + }, + { + "epoch": 45.32224532224532, + "eval_loss": 0.25192537903785706, + "eval_runtime": 13.3766, + "eval_samples_per_second": 63.993, + "eval_steps_per_second": 15.998, + "num_input_tokens_seen": 16625808, + "step": 21800 + }, + { + "epoch": 45.33264033264033, + "grad_norm": 0.0002906887384597212, + "learning_rate": 0.1288181624873601, + "loss": 0.2629, + "num_input_tokens_seen": 16629872, + "step": 21805 + }, + { + "epoch": 45.343035343035346, + "grad_norm": 8.257607987616211e-05, + "learning_rate": 0.12875984952525163, + "loss": 0.2681, + "num_input_tokens_seen": 16633808, + "step": 21810 + }, + { + "epoch": 45.353430353430355, + "grad_norm": 6.172782741487026e-05, + "learning_rate": 0.12870153983864122, + "loss": 0.2663, + "num_input_tokens_seen": 16637424, + "step": 21815 + }, + { + "epoch": 45.36382536382536, + "grad_norm": 0.0001658169785514474, + "learning_rate": 0.12864323343652104, + "loss": 0.2751, + "num_input_tokens_seen": 16641232, + "step": 21820 + }, + { + "epoch": 45.37422037422037, + "grad_norm": 9.83078934950754e-05, + "learning_rate": 0.12858493032788268, + "loss": 0.2792, + "num_input_tokens_seen": 16644880, + "step": 21825 + }, + { + "epoch": 45.38461538461539, + "grad_norm": 0.0005819424404762685, + "learning_rate": 0.12852663052171714, + "loss": 0.2407, + "num_input_tokens_seen": 16648720, + "step": 21830 + }, + { + "epoch": 45.395010395010395, + "grad_norm": 0.00017988719628192484, + "learning_rate": 0.12846833402701507, + "loss": 0.2569, + "num_input_tokens_seen": 16652368, + "step": 21835 + }, + { + "epoch": 45.4054054054054, + "grad_norm": 0.0005045937141403556, + "learning_rate": 0.12841004085276642, + "loss": 0.261, + "num_input_tokens_seen": 16656272, + "step": 21840 + }, + { + "epoch": 45.41580041580042, + "grad_norm": 0.00023076646903064102, + "learning_rate": 0.12835175100796076, + "loss": 0.2494, + "num_input_tokens_seen": 16660176, + "step": 21845 + }, + { + "epoch": 45.42619542619543, + "grad_norm": 0.00021518296853173524, + "learning_rate": 0.12829346450158724, + "loss": 0.2894, + "num_input_tokens_seen": 16664016, + "step": 21850 + }, + { + "epoch": 45.436590436590436, + "grad_norm": 0.00023761062766425312, + "learning_rate": 0.12823518134263423, + "loss": 0.2632, + "num_input_tokens_seen": 16667888, + "step": 21855 + }, + { + "epoch": 45.446985446985444, + "grad_norm": 0.0002611563540995121, + "learning_rate": 0.12817690154008973, + "loss": 0.2613, + "num_input_tokens_seen": 16671728, + "step": 21860 + }, + { + "epoch": 45.45738045738046, + "grad_norm": 0.0006271555321291089, + "learning_rate": 0.12811862510294134, + "loss": 0.2664, + "num_input_tokens_seen": 16675472, + "step": 21865 + }, + { + "epoch": 45.46777546777547, + "grad_norm": 7.920320058474317e-05, + "learning_rate": 0.12806035204017585, + "loss": 0.2603, + "num_input_tokens_seen": 16679280, + "step": 21870 + }, + { + "epoch": 45.478170478170476, + "grad_norm": 0.00011259713937761262, + "learning_rate": 0.12800208236077987, + "loss": 0.2733, + "num_input_tokens_seen": 16683088, + "step": 21875 + }, + { + "epoch": 45.48856548856549, + "grad_norm": 0.00017516405205242336, + "learning_rate": 0.12794381607373917, + "loss": 0.2943, + "num_input_tokens_seen": 16687088, + "step": 21880 + }, + { + "epoch": 45.4989604989605, + "grad_norm": 7.293133239727467e-05, + "learning_rate": 0.12788555318803924, + "loss": 0.2586, + "num_input_tokens_seen": 16690864, + "step": 21885 + }, + { + "epoch": 45.50935550935551, + "grad_norm": 0.00026224294560961425, + "learning_rate": 0.1278272937126649, + "loss": 0.2594, + "num_input_tokens_seen": 16694704, + "step": 21890 + }, + { + "epoch": 45.51975051975052, + "grad_norm": 0.00033755958429537714, + "learning_rate": 0.1277690376566005, + "loss": 0.2683, + "num_input_tokens_seen": 16698576, + "step": 21895 + }, + { + "epoch": 45.53014553014553, + "grad_norm": 7.646583981113508e-05, + "learning_rate": 0.12771078502882985, + "loss": 0.2612, + "num_input_tokens_seen": 16702288, + "step": 21900 + }, + { + "epoch": 45.54054054054054, + "grad_norm": 7.660726987523958e-05, + "learning_rate": 0.12765253583833633, + "loss": 0.2728, + "num_input_tokens_seen": 16706160, + "step": 21905 + }, + { + "epoch": 45.55093555093555, + "grad_norm": 0.00036447244929149747, + "learning_rate": 0.12759429009410256, + "loss": 0.2802, + "num_input_tokens_seen": 16710064, + "step": 21910 + }, + { + "epoch": 45.561330561330564, + "grad_norm": 0.000583614397328347, + "learning_rate": 0.12753604780511085, + "loss": 0.2658, + "num_input_tokens_seen": 16713808, + "step": 21915 + }, + { + "epoch": 45.57172557172557, + "grad_norm": 0.00017305099754594266, + "learning_rate": 0.12747780898034283, + "loss": 0.2614, + "num_input_tokens_seen": 16717680, + "step": 21920 + }, + { + "epoch": 45.58212058212058, + "grad_norm": 0.00039512160583399236, + "learning_rate": 0.12741957362877973, + "loss": 0.2824, + "num_input_tokens_seen": 16721328, + "step": 21925 + }, + { + "epoch": 45.59251559251559, + "grad_norm": 0.00017481579561717808, + "learning_rate": 0.12736134175940214, + "loss": 0.2762, + "num_input_tokens_seen": 16725040, + "step": 21930 + }, + { + "epoch": 45.602910602910605, + "grad_norm": 5.549371780944057e-05, + "learning_rate": 0.12730311338119016, + "loss": 0.2401, + "num_input_tokens_seen": 16728816, + "step": 21935 + }, + { + "epoch": 45.61330561330561, + "grad_norm": 0.0002514217921998352, + "learning_rate": 0.12724488850312327, + "loss": 0.2678, + "num_input_tokens_seen": 16732688, + "step": 21940 + }, + { + "epoch": 45.62370062370062, + "grad_norm": 0.0001222207793034613, + "learning_rate": 0.1271866671341806, + "loss": 0.2481, + "num_input_tokens_seen": 16736368, + "step": 21945 + }, + { + "epoch": 45.63409563409564, + "grad_norm": 0.0004071629955433309, + "learning_rate": 0.12712844928334047, + "loss": 0.2726, + "num_input_tokens_seen": 16740240, + "step": 21950 + }, + { + "epoch": 45.644490644490645, + "grad_norm": 0.0003476296260487288, + "learning_rate": 0.12707023495958095, + "loss": 0.2314, + "num_input_tokens_seen": 16743920, + "step": 21955 + }, + { + "epoch": 45.65488565488565, + "grad_norm": 6.0887155996169895e-05, + "learning_rate": 0.12701202417187932, + "loss": 0.2261, + "num_input_tokens_seen": 16747824, + "step": 21960 + }, + { + "epoch": 45.66528066528066, + "grad_norm": 0.0005414694896899164, + "learning_rate": 0.12695381692921243, + "loss": 0.3144, + "num_input_tokens_seen": 16751824, + "step": 21965 + }, + { + "epoch": 45.67567567567568, + "grad_norm": 0.00018057672423310578, + "learning_rate": 0.12689561324055665, + "loss": 0.2664, + "num_input_tokens_seen": 16755568, + "step": 21970 + }, + { + "epoch": 45.686070686070686, + "grad_norm": 0.00013447966193780303, + "learning_rate": 0.12683741311488758, + "loss": 0.268, + "num_input_tokens_seen": 16759184, + "step": 21975 + }, + { + "epoch": 45.696465696465694, + "grad_norm": 0.00041670765494927764, + "learning_rate": 0.1267792165611805, + "loss": 0.2732, + "num_input_tokens_seen": 16762928, + "step": 21980 + }, + { + "epoch": 45.70686070686071, + "grad_norm": 0.00018882739823311567, + "learning_rate": 0.1267210235884101, + "loss": 0.2674, + "num_input_tokens_seen": 16766704, + "step": 21985 + }, + { + "epoch": 45.71725571725572, + "grad_norm": 0.00014354908489622176, + "learning_rate": 0.12666283420555033, + "loss": 0.2761, + "num_input_tokens_seen": 16770544, + "step": 21990 + }, + { + "epoch": 45.727650727650726, + "grad_norm": 0.00031965802190825343, + "learning_rate": 0.12660464842157487, + "loss": 0.2598, + "num_input_tokens_seen": 16774384, + "step": 21995 + }, + { + "epoch": 45.73804573804574, + "grad_norm": 0.00010237919195787981, + "learning_rate": 0.1265464662454566, + "loss": 0.261, + "num_input_tokens_seen": 16778288, + "step": 22000 + }, + { + "epoch": 45.73804573804574, + "eval_loss": 0.2485373318195343, + "eval_runtime": 13.3685, + "eval_samples_per_second": 64.031, + "eval_steps_per_second": 16.008, + "num_input_tokens_seen": 16778288, + "step": 22000 + }, + { + "epoch": 45.74844074844075, + "grad_norm": 0.00037750197225250304, + "learning_rate": 0.12648828768616793, + "loss": 0.2745, + "num_input_tokens_seen": 16782032, + "step": 22005 + }, + { + "epoch": 45.75883575883576, + "grad_norm": 0.0003712645557243377, + "learning_rate": 0.12643011275268085, + "loss": 0.2424, + "num_input_tokens_seen": 16786064, + "step": 22010 + }, + { + "epoch": 45.76923076923077, + "grad_norm": 5.295631126500666e-05, + "learning_rate": 0.1263719414539665, + "loss": 0.2647, + "num_input_tokens_seen": 16789872, + "step": 22015 + }, + { + "epoch": 45.77962577962578, + "grad_norm": 0.00017828235286287963, + "learning_rate": 0.1263137737989957, + "loss": 0.2614, + "num_input_tokens_seen": 16793776, + "step": 22020 + }, + { + "epoch": 45.79002079002079, + "grad_norm": 0.00031687956652604043, + "learning_rate": 0.1262556097967387, + "loss": 0.2726, + "num_input_tokens_seen": 16797680, + "step": 22025 + }, + { + "epoch": 45.8004158004158, + "grad_norm": 8.944228466134518e-05, + "learning_rate": 0.126197449456165, + "loss": 0.2565, + "num_input_tokens_seen": 16801616, + "step": 22030 + }, + { + "epoch": 45.810810810810814, + "grad_norm": 8.557483670301735e-05, + "learning_rate": 0.12613929278624378, + "loss": 0.2689, + "num_input_tokens_seen": 16805456, + "step": 22035 + }, + { + "epoch": 45.82120582120582, + "grad_norm": 0.00010479663615114987, + "learning_rate": 0.12608113979594343, + "loss": 0.2486, + "num_input_tokens_seen": 16809200, + "step": 22040 + }, + { + "epoch": 45.83160083160083, + "grad_norm": 7.175657810876146e-05, + "learning_rate": 0.1260229904942319, + "loss": 0.2912, + "num_input_tokens_seen": 16813072, + "step": 22045 + }, + { + "epoch": 45.84199584199584, + "grad_norm": 5.344607779989019e-05, + "learning_rate": 0.12596484489007662, + "loss": 0.2695, + "num_input_tokens_seen": 16816944, + "step": 22050 + }, + { + "epoch": 45.852390852390855, + "grad_norm": 0.00020244310144335032, + "learning_rate": 0.1259067029924442, + "loss": 0.2673, + "num_input_tokens_seen": 16820688, + "step": 22055 + }, + { + "epoch": 45.86278586278586, + "grad_norm": 0.00012761795369442552, + "learning_rate": 0.12584856481030096, + "loss": 0.253, + "num_input_tokens_seen": 16824464, + "step": 22060 + }, + { + "epoch": 45.87318087318087, + "grad_norm": 3.073545667575672e-05, + "learning_rate": 0.12579043035261261, + "loss": 0.2452, + "num_input_tokens_seen": 16828240, + "step": 22065 + }, + { + "epoch": 45.88357588357589, + "grad_norm": 0.000197981673409231, + "learning_rate": 0.1257322996283441, + "loss": 0.2538, + "num_input_tokens_seen": 16832112, + "step": 22070 + }, + { + "epoch": 45.893970893970895, + "grad_norm": 0.00019772742234636098, + "learning_rate": 0.12567417264645994, + "loss": 0.2208, + "num_input_tokens_seen": 16835984, + "step": 22075 + }, + { + "epoch": 45.9043659043659, + "grad_norm": 5.160536238690838e-05, + "learning_rate": 0.12561604941592408, + "loss": 0.3101, + "num_input_tokens_seen": 16839888, + "step": 22080 + }, + { + "epoch": 45.91476091476091, + "grad_norm": 6.732078327331692e-05, + "learning_rate": 0.12555792994569978, + "loss": 0.2735, + "num_input_tokens_seen": 16843600, + "step": 22085 + }, + { + "epoch": 45.92515592515593, + "grad_norm": 0.0004544588446151465, + "learning_rate": 0.1254998142447499, + "loss": 0.2676, + "num_input_tokens_seen": 16847568, + "step": 22090 + }, + { + "epoch": 45.935550935550935, + "grad_norm": 4.943975727655925e-05, + "learning_rate": 0.1254417023220365, + "loss": 0.263, + "num_input_tokens_seen": 16851376, + "step": 22095 + }, + { + "epoch": 45.945945945945944, + "grad_norm": 0.00025042553897947073, + "learning_rate": 0.12538359418652126, + "loss": 0.2797, + "num_input_tokens_seen": 16855088, + "step": 22100 + }, + { + "epoch": 45.95634095634096, + "grad_norm": 8.274863648694009e-05, + "learning_rate": 0.12532548984716513, + "loss": 0.2728, + "num_input_tokens_seen": 16858864, + "step": 22105 + }, + { + "epoch": 45.96673596673597, + "grad_norm": 0.00010645118891261518, + "learning_rate": 0.12526738931292855, + "loss": 0.265, + "num_input_tokens_seen": 16862608, + "step": 22110 + }, + { + "epoch": 45.977130977130976, + "grad_norm": 0.00015119793533813208, + "learning_rate": 0.1252092925927714, + "loss": 0.2646, + "num_input_tokens_seen": 16866416, + "step": 22115 + }, + { + "epoch": 45.987525987525984, + "grad_norm": 0.00017371904687024653, + "learning_rate": 0.12515119969565278, + "loss": 0.2775, + "num_input_tokens_seen": 16870160, + "step": 22120 + }, + { + "epoch": 45.997920997921, + "grad_norm": 0.00016638576926197857, + "learning_rate": 0.12509311063053144, + "loss": 0.2666, + "num_input_tokens_seen": 16873872, + "step": 22125 + }, + { + "epoch": 46.00831600831601, + "grad_norm": 0.0004470042185857892, + "learning_rate": 0.1250350254063655, + "loss": 0.2609, + "num_input_tokens_seen": 16877672, + "step": 22130 + }, + { + "epoch": 46.018711018711016, + "grad_norm": 0.0001230023626703769, + "learning_rate": 0.1249769440321123, + "loss": 0.2712, + "num_input_tokens_seen": 16881512, + "step": 22135 + }, + { + "epoch": 46.02910602910603, + "grad_norm": 7.218630344141275e-05, + "learning_rate": 0.12491886651672884, + "loss": 0.2638, + "num_input_tokens_seen": 16885320, + "step": 22140 + }, + { + "epoch": 46.03950103950104, + "grad_norm": 0.0003252634487580508, + "learning_rate": 0.12486079286917139, + "loss": 0.2761, + "num_input_tokens_seen": 16889224, + "step": 22145 + }, + { + "epoch": 46.04989604989605, + "grad_norm": 0.00032207617186941206, + "learning_rate": 0.12480272309839553, + "loss": 0.2598, + "num_input_tokens_seen": 16893032, + "step": 22150 + }, + { + "epoch": 46.06029106029106, + "grad_norm": 0.0003125071234535426, + "learning_rate": 0.12474465721335648, + "loss": 0.2944, + "num_input_tokens_seen": 16896808, + "step": 22155 + }, + { + "epoch": 46.07068607068607, + "grad_norm": 0.0003102972113993019, + "learning_rate": 0.12468659522300861, + "loss": 0.2682, + "num_input_tokens_seen": 16900680, + "step": 22160 + }, + { + "epoch": 46.08108108108108, + "grad_norm": 0.00018275667389389127, + "learning_rate": 0.12462853713630584, + "loss": 0.2626, + "num_input_tokens_seen": 16904680, + "step": 22165 + }, + { + "epoch": 46.09147609147609, + "grad_norm": 0.00015696596528869122, + "learning_rate": 0.12457048296220156, + "loss": 0.2624, + "num_input_tokens_seen": 16908392, + "step": 22170 + }, + { + "epoch": 46.101871101871104, + "grad_norm": 7.242074934765697e-05, + "learning_rate": 0.12451243270964832, + "loss": 0.2855, + "num_input_tokens_seen": 16912200, + "step": 22175 + }, + { + "epoch": 46.11226611226611, + "grad_norm": 0.00020489588496275246, + "learning_rate": 0.12445438638759827, + "loss": 0.2547, + "num_input_tokens_seen": 16916072, + "step": 22180 + }, + { + "epoch": 46.12266112266112, + "grad_norm": 0.0005089446785859764, + "learning_rate": 0.1243963440050029, + "loss": 0.2841, + "num_input_tokens_seen": 16919976, + "step": 22185 + }, + { + "epoch": 46.13305613305613, + "grad_norm": 0.0003243469982407987, + "learning_rate": 0.12433830557081298, + "loss": 0.2576, + "num_input_tokens_seen": 16923720, + "step": 22190 + }, + { + "epoch": 46.143451143451145, + "grad_norm": 0.00043446035124361515, + "learning_rate": 0.12428027109397889, + "loss": 0.2558, + "num_input_tokens_seen": 16927688, + "step": 22195 + }, + { + "epoch": 46.15384615384615, + "grad_norm": 0.0001216045202454552, + "learning_rate": 0.12422224058345015, + "loss": 0.2597, + "num_input_tokens_seen": 16931560, + "step": 22200 + }, + { + "epoch": 46.15384615384615, + "eval_loss": 0.24865058064460754, + "eval_runtime": 13.3945, + "eval_samples_per_second": 63.907, + "eval_steps_per_second": 15.977, + "num_input_tokens_seen": 16931560, + "step": 22200 + }, + { + "epoch": 46.16424116424116, + "grad_norm": 0.0001540331431897357, + "learning_rate": 0.12416421404817583, + "loss": 0.2954, + "num_input_tokens_seen": 16935400, + "step": 22205 + }, + { + "epoch": 46.17463617463618, + "grad_norm": 0.00020360035705380142, + "learning_rate": 0.12410619149710447, + "loss": 0.2773, + "num_input_tokens_seen": 16939208, + "step": 22210 + }, + { + "epoch": 46.185031185031185, + "grad_norm": 0.00012955625425092876, + "learning_rate": 0.12404817293918374, + "loss": 0.2601, + "num_input_tokens_seen": 16943112, + "step": 22215 + }, + { + "epoch": 46.195426195426194, + "grad_norm": 0.00030278105987235904, + "learning_rate": 0.12399015838336086, + "loss": 0.2698, + "num_input_tokens_seen": 16946920, + "step": 22220 + }, + { + "epoch": 46.20582120582121, + "grad_norm": 0.000351201044395566, + "learning_rate": 0.12393214783858246, + "loss": 0.2542, + "num_input_tokens_seen": 16950600, + "step": 22225 + }, + { + "epoch": 46.21621621621622, + "grad_norm": 0.0005220185848884284, + "learning_rate": 0.1238741413137944, + "loss": 0.2409, + "num_input_tokens_seen": 16954472, + "step": 22230 + }, + { + "epoch": 46.226611226611226, + "grad_norm": 0.00029369135154411197, + "learning_rate": 0.12381613881794212, + "loss": 0.2901, + "num_input_tokens_seen": 16958440, + "step": 22235 + }, + { + "epoch": 46.237006237006234, + "grad_norm": 0.00020896198111586273, + "learning_rate": 0.12375814035997022, + "loss": 0.2628, + "num_input_tokens_seen": 16962312, + "step": 22240 + }, + { + "epoch": 46.24740124740125, + "grad_norm": 0.0006270611193031073, + "learning_rate": 0.12370014594882285, + "loss": 0.2468, + "num_input_tokens_seen": 16966184, + "step": 22245 + }, + { + "epoch": 46.25779625779626, + "grad_norm": 0.00010485291568329558, + "learning_rate": 0.12364215559344356, + "loss": 0.2621, + "num_input_tokens_seen": 16970216, + "step": 22250 + }, + { + "epoch": 46.268191268191266, + "grad_norm": 7.513363379985094e-05, + "learning_rate": 0.12358416930277506, + "loss": 0.2634, + "num_input_tokens_seen": 16974056, + "step": 22255 + }, + { + "epoch": 46.27858627858628, + "grad_norm": 0.00019662165141198784, + "learning_rate": 0.1235261870857596, + "loss": 0.2647, + "num_input_tokens_seen": 16977864, + "step": 22260 + }, + { + "epoch": 46.28898128898129, + "grad_norm": 0.0004189387254882604, + "learning_rate": 0.12346820895133884, + "loss": 0.2746, + "num_input_tokens_seen": 16981576, + "step": 22265 + }, + { + "epoch": 46.2993762993763, + "grad_norm": 0.0003355878870934248, + "learning_rate": 0.12341023490845361, + "loss": 0.2627, + "num_input_tokens_seen": 16985224, + "step": 22270 + }, + { + "epoch": 46.30977130977131, + "grad_norm": 0.000288547424133867, + "learning_rate": 0.12335226496604437, + "loss": 0.2521, + "num_input_tokens_seen": 16988936, + "step": 22275 + }, + { + "epoch": 46.32016632016632, + "grad_norm": 7.834876305423677e-05, + "learning_rate": 0.12329429913305069, + "loss": 0.2528, + "num_input_tokens_seen": 16992808, + "step": 22280 + }, + { + "epoch": 46.33056133056133, + "grad_norm": 0.0001973739272216335, + "learning_rate": 0.12323633741841171, + "loss": 0.2833, + "num_input_tokens_seen": 16996680, + "step": 22285 + }, + { + "epoch": 46.34095634095634, + "grad_norm": 0.00031665078131482005, + "learning_rate": 0.12317837983106583, + "loss": 0.2475, + "num_input_tokens_seen": 17000456, + "step": 22290 + }, + { + "epoch": 46.351351351351354, + "grad_norm": 0.0001569976011523977, + "learning_rate": 0.12312042637995087, + "loss": 0.2595, + "num_input_tokens_seen": 17004168, + "step": 22295 + }, + { + "epoch": 46.36174636174636, + "grad_norm": 5.593411697191186e-05, + "learning_rate": 0.12306247707400389, + "loss": 0.275, + "num_input_tokens_seen": 17007976, + "step": 22300 + }, + { + "epoch": 46.37214137214137, + "grad_norm": 0.00031737436074763536, + "learning_rate": 0.12300453192216154, + "loss": 0.2541, + "num_input_tokens_seen": 17011944, + "step": 22305 + }, + { + "epoch": 46.38253638253638, + "grad_norm": 0.00016067716933321208, + "learning_rate": 0.12294659093335956, + "loss": 0.2831, + "num_input_tokens_seen": 17015688, + "step": 22310 + }, + { + "epoch": 46.392931392931395, + "grad_norm": 0.0001555733324494213, + "learning_rate": 0.12288865411653327, + "loss": 0.2697, + "num_input_tokens_seen": 17019464, + "step": 22315 + }, + { + "epoch": 46.4033264033264, + "grad_norm": 0.0001924388052430004, + "learning_rate": 0.12283072148061717, + "loss": 0.2856, + "num_input_tokens_seen": 17023176, + "step": 22320 + }, + { + "epoch": 46.41372141372141, + "grad_norm": 0.00019046859233640134, + "learning_rate": 0.12277279303454529, + "loss": 0.2779, + "num_input_tokens_seen": 17026952, + "step": 22325 + }, + { + "epoch": 46.42411642411643, + "grad_norm": 0.0005711017292924225, + "learning_rate": 0.12271486878725091, + "loss": 0.2753, + "num_input_tokens_seen": 17030728, + "step": 22330 + }, + { + "epoch": 46.434511434511435, + "grad_norm": 0.0002052215568255633, + "learning_rate": 0.12265694874766658, + "loss": 0.2866, + "num_input_tokens_seen": 17034536, + "step": 22335 + }, + { + "epoch": 46.444906444906444, + "grad_norm": 0.00029944657580927014, + "learning_rate": 0.12259903292472435, + "loss": 0.2645, + "num_input_tokens_seen": 17038472, + "step": 22340 + }, + { + "epoch": 46.45530145530145, + "grad_norm": 0.00022672700288239866, + "learning_rate": 0.12254112132735567, + "loss": 0.2449, + "num_input_tokens_seen": 17042248, + "step": 22345 + }, + { + "epoch": 46.46569646569647, + "grad_norm": 0.0005383013049140573, + "learning_rate": 0.12248321396449108, + "loss": 0.2702, + "num_input_tokens_seen": 17046184, + "step": 22350 + }, + { + "epoch": 46.476091476091476, + "grad_norm": 0.0002797792258206755, + "learning_rate": 0.12242531084506075, + "loss": 0.2762, + "num_input_tokens_seen": 17049864, + "step": 22355 + }, + { + "epoch": 46.486486486486484, + "grad_norm": 0.0002097926044370979, + "learning_rate": 0.122367411977994, + "loss": 0.2496, + "num_input_tokens_seen": 17053576, + "step": 22360 + }, + { + "epoch": 46.4968814968815, + "grad_norm": 0.00028601373196579516, + "learning_rate": 0.12230951737221954, + "loss": 0.2597, + "num_input_tokens_seen": 17057352, + "step": 22365 + }, + { + "epoch": 46.50727650727651, + "grad_norm": 0.0004895106540061533, + "learning_rate": 0.12225162703666555, + "loss": 0.2658, + "num_input_tokens_seen": 17061224, + "step": 22370 + }, + { + "epoch": 46.517671517671516, + "grad_norm": 0.0003517533768899739, + "learning_rate": 0.1221937409802593, + "loss": 0.2405, + "num_input_tokens_seen": 17064840, + "step": 22375 + }, + { + "epoch": 46.528066528066525, + "grad_norm": 3.7381454603746533e-05, + "learning_rate": 0.12213585921192768, + "loss": 0.2638, + "num_input_tokens_seen": 17068616, + "step": 22380 + }, + { + "epoch": 46.53846153846154, + "grad_norm": 0.0004954173928126693, + "learning_rate": 0.1220779817405967, + "loss": 0.2754, + "num_input_tokens_seen": 17072520, + "step": 22385 + }, + { + "epoch": 46.54885654885655, + "grad_norm": 0.00013085729733575135, + "learning_rate": 0.12202010857519181, + "loss": 0.2824, + "num_input_tokens_seen": 17076264, + "step": 22390 + }, + { + "epoch": 46.55925155925156, + "grad_norm": 0.00023271025565918535, + "learning_rate": 0.12196223972463785, + "loss": 0.2693, + "num_input_tokens_seen": 17080136, + "step": 22395 + }, + { + "epoch": 46.56964656964657, + "grad_norm": 0.00025900444597937167, + "learning_rate": 0.12190437519785885, + "loss": 0.2781, + "num_input_tokens_seen": 17083880, + "step": 22400 + }, + { + "epoch": 46.56964656964657, + "eval_loss": 0.25912460684776306, + "eval_runtime": 13.3799, + "eval_samples_per_second": 63.976, + "eval_steps_per_second": 15.994, + "num_input_tokens_seen": 17083880, + "step": 22400 + }, + { + "epoch": 46.58004158004158, + "grad_norm": 6.228961865417659e-05, + "learning_rate": 0.12184651500377823, + "loss": 0.2744, + "num_input_tokens_seen": 17087880, + "step": 22405 + }, + { + "epoch": 46.59043659043659, + "grad_norm": 7.690744678257033e-05, + "learning_rate": 0.12178865915131885, + "loss": 0.2647, + "num_input_tokens_seen": 17091688, + "step": 22410 + }, + { + "epoch": 46.6008316008316, + "grad_norm": 0.00030996662098914385, + "learning_rate": 0.1217308076494027, + "loss": 0.2421, + "num_input_tokens_seen": 17095400, + "step": 22415 + }, + { + "epoch": 46.61122661122661, + "grad_norm": 0.0003689787699840963, + "learning_rate": 0.12167296050695134, + "loss": 0.2943, + "num_input_tokens_seen": 17099272, + "step": 22420 + }, + { + "epoch": 46.62162162162162, + "grad_norm": 0.00016878554015420377, + "learning_rate": 0.12161511773288536, + "loss": 0.2675, + "num_input_tokens_seen": 17103016, + "step": 22425 + }, + { + "epoch": 46.63201663201663, + "grad_norm": 0.00015906919725239277, + "learning_rate": 0.121557279336125, + "loss": 0.2513, + "num_input_tokens_seen": 17106728, + "step": 22430 + }, + { + "epoch": 46.642411642411645, + "grad_norm": 0.0004891976714134216, + "learning_rate": 0.12149944532558957, + "loss": 0.2527, + "num_input_tokens_seen": 17110568, + "step": 22435 + }, + { + "epoch": 46.65280665280665, + "grad_norm": 0.00012872193474322557, + "learning_rate": 0.12144161571019785, + "loss": 0.261, + "num_input_tokens_seen": 17114408, + "step": 22440 + }, + { + "epoch": 46.66320166320166, + "grad_norm": 0.000386368774343282, + "learning_rate": 0.12138379049886781, + "loss": 0.2676, + "num_input_tokens_seen": 17118088, + "step": 22445 + }, + { + "epoch": 46.67359667359668, + "grad_norm": 0.00017513229977339506, + "learning_rate": 0.12132596970051697, + "loss": 0.2884, + "num_input_tokens_seen": 17121864, + "step": 22450 + }, + { + "epoch": 46.683991683991685, + "grad_norm": 0.0004549797740764916, + "learning_rate": 0.12126815332406189, + "loss": 0.2779, + "num_input_tokens_seen": 17125544, + "step": 22455 + }, + { + "epoch": 46.694386694386694, + "grad_norm": 0.0001555889321025461, + "learning_rate": 0.12121034137841868, + "loss": 0.2678, + "num_input_tokens_seen": 17129256, + "step": 22460 + }, + { + "epoch": 46.7047817047817, + "grad_norm": 0.000434901041444391, + "learning_rate": 0.12115253387250258, + "loss": 0.2624, + "num_input_tokens_seen": 17133096, + "step": 22465 + }, + { + "epoch": 46.71517671517672, + "grad_norm": 0.00012585431977640837, + "learning_rate": 0.12109473081522831, + "loss": 0.2499, + "num_input_tokens_seen": 17136712, + "step": 22470 + }, + { + "epoch": 46.725571725571726, + "grad_norm": 0.00016275845700874925, + "learning_rate": 0.12103693221550982, + "loss": 0.2534, + "num_input_tokens_seen": 17140488, + "step": 22475 + }, + { + "epoch": 46.735966735966734, + "grad_norm": 0.00021853778162039816, + "learning_rate": 0.12097913808226027, + "loss": 0.2505, + "num_input_tokens_seen": 17144520, + "step": 22480 + }, + { + "epoch": 46.74636174636175, + "grad_norm": 8.459365199087188e-05, + "learning_rate": 0.12092134842439234, + "loss": 0.2746, + "num_input_tokens_seen": 17148296, + "step": 22485 + }, + { + "epoch": 46.75675675675676, + "grad_norm": 9.782022971194237e-05, + "learning_rate": 0.12086356325081798, + "loss": 0.2927, + "num_input_tokens_seen": 17152136, + "step": 22490 + }, + { + "epoch": 46.767151767151766, + "grad_norm": 5.016908471588977e-05, + "learning_rate": 0.12080578257044824, + "loss": 0.2796, + "num_input_tokens_seen": 17155848, + "step": 22495 + }, + { + "epoch": 46.777546777546775, + "grad_norm": 0.00017641666636336595, + "learning_rate": 0.12074800639219378, + "loss": 0.2688, + "num_input_tokens_seen": 17159720, + "step": 22500 + }, + { + "epoch": 46.78794178794179, + "grad_norm": 0.0013393232366070151, + "learning_rate": 0.12069023472496428, + "loss": 0.2668, + "num_input_tokens_seen": 17163560, + "step": 22505 + }, + { + "epoch": 46.7983367983368, + "grad_norm": 0.00012373989738989621, + "learning_rate": 0.12063246757766893, + "loss": 0.2719, + "num_input_tokens_seen": 17167304, + "step": 22510 + }, + { + "epoch": 46.80873180873181, + "grad_norm": 9.288824367104098e-05, + "learning_rate": 0.12057470495921618, + "loss": 0.2603, + "num_input_tokens_seen": 17171144, + "step": 22515 + }, + { + "epoch": 46.81912681912682, + "grad_norm": 0.0002501018170733005, + "learning_rate": 0.12051694687851364, + "loss": 0.2506, + "num_input_tokens_seen": 17175016, + "step": 22520 + }, + { + "epoch": 46.82952182952183, + "grad_norm": 0.0002631780516821891, + "learning_rate": 0.12045919334446839, + "loss": 0.2585, + "num_input_tokens_seen": 17178984, + "step": 22525 + }, + { + "epoch": 46.83991683991684, + "grad_norm": 0.0005281229969114065, + "learning_rate": 0.12040144436598683, + "loss": 0.2568, + "num_input_tokens_seen": 17182696, + "step": 22530 + }, + { + "epoch": 46.85031185031185, + "grad_norm": 8.637745486339554e-05, + "learning_rate": 0.12034369995197444, + "loss": 0.2928, + "num_input_tokens_seen": 17186376, + "step": 22535 + }, + { + "epoch": 46.86070686070686, + "grad_norm": 4.935479228151962e-05, + "learning_rate": 0.12028596011133627, + "loss": 0.2581, + "num_input_tokens_seen": 17190056, + "step": 22540 + }, + { + "epoch": 46.87110187110187, + "grad_norm": 0.0001000378470052965, + "learning_rate": 0.12022822485297643, + "loss": 0.2628, + "num_input_tokens_seen": 17193992, + "step": 22545 + }, + { + "epoch": 46.88149688149688, + "grad_norm": 0.0003920421586371958, + "learning_rate": 0.12017049418579843, + "loss": 0.2618, + "num_input_tokens_seen": 17197800, + "step": 22550 + }, + { + "epoch": 46.891891891891895, + "grad_norm": 0.00031206224230118096, + "learning_rate": 0.12011276811870514, + "loss": 0.2529, + "num_input_tokens_seen": 17201576, + "step": 22555 + }, + { + "epoch": 46.9022869022869, + "grad_norm": 0.00015635065210517496, + "learning_rate": 0.12005504666059852, + "loss": 0.2822, + "num_input_tokens_seen": 17205448, + "step": 22560 + }, + { + "epoch": 46.91268191268191, + "grad_norm": 0.00044488132698461413, + "learning_rate": 0.11999732982038003, + "loss": 0.2573, + "num_input_tokens_seen": 17209160, + "step": 22565 + }, + { + "epoch": 46.92307692307692, + "grad_norm": 0.0001751144736772403, + "learning_rate": 0.11993961760695038, + "loss": 0.2656, + "num_input_tokens_seen": 17212936, + "step": 22570 + }, + { + "epoch": 46.933471933471935, + "grad_norm": 0.00012267709826119244, + "learning_rate": 0.11988191002920942, + "loss": 0.2515, + "num_input_tokens_seen": 17216936, + "step": 22575 + }, + { + "epoch": 46.943866943866944, + "grad_norm": 0.000148308725329116, + "learning_rate": 0.11982420709605641, + "loss": 0.2445, + "num_input_tokens_seen": 17220680, + "step": 22580 + }, + { + "epoch": 46.95426195426195, + "grad_norm": 6.409818888641894e-05, + "learning_rate": 0.11976650881638991, + "loss": 0.2813, + "num_input_tokens_seen": 17224360, + "step": 22585 + }, + { + "epoch": 46.96465696465697, + "grad_norm": 0.00019950537534896284, + "learning_rate": 0.11970881519910764, + "loss": 0.2372, + "num_input_tokens_seen": 17228296, + "step": 22590 + }, + { + "epoch": 46.975051975051976, + "grad_norm": 0.00038636833778582513, + "learning_rate": 0.1196511262531068, + "loss": 0.2836, + "num_input_tokens_seen": 17232200, + "step": 22595 + }, + { + "epoch": 46.985446985446984, + "grad_norm": 0.0002492482017260045, + "learning_rate": 0.11959344198728361, + "loss": 0.2659, + "num_input_tokens_seen": 17235976, + "step": 22600 + }, + { + "epoch": 46.985446985446984, + "eval_loss": 0.25825560092926025, + "eval_runtime": 13.3845, + "eval_samples_per_second": 63.954, + "eval_steps_per_second": 15.989, + "num_input_tokens_seen": 17235976, + "step": 22600 + }, + { + "epoch": 46.99584199584199, + "grad_norm": 0.0002010871103266254, + "learning_rate": 0.11953576241053378, + "loss": 0.2625, + "num_input_tokens_seen": 17239784, + "step": 22605 + }, + { + "epoch": 47.00623700623701, + "grad_norm": 5.701424743165262e-05, + "learning_rate": 0.11947808753175228, + "loss": 0.2599, + "num_input_tokens_seen": 17243352, + "step": 22610 + }, + { + "epoch": 47.016632016632016, + "grad_norm": 0.00010315680265193805, + "learning_rate": 0.1194204173598332, + "loss": 0.2889, + "num_input_tokens_seen": 17247224, + "step": 22615 + }, + { + "epoch": 47.027027027027025, + "grad_norm": 0.00019773351959884167, + "learning_rate": 0.11936275190367007, + "loss": 0.2765, + "num_input_tokens_seen": 17251128, + "step": 22620 + }, + { + "epoch": 47.03742203742204, + "grad_norm": 0.00014053906488697976, + "learning_rate": 0.11930509117215563, + "loss": 0.249, + "num_input_tokens_seen": 17254968, + "step": 22625 + }, + { + "epoch": 47.04781704781705, + "grad_norm": 0.00018515456758905202, + "learning_rate": 0.11924743517418179, + "loss": 0.2724, + "num_input_tokens_seen": 17258712, + "step": 22630 + }, + { + "epoch": 47.05821205821206, + "grad_norm": 0.0003924377670045942, + "learning_rate": 0.11918978391864, + "loss": 0.2446, + "num_input_tokens_seen": 17262616, + "step": 22635 + }, + { + "epoch": 47.06860706860707, + "grad_norm": 0.00011191734665771946, + "learning_rate": 0.11913213741442065, + "loss": 0.2719, + "num_input_tokens_seen": 17266520, + "step": 22640 + }, + { + "epoch": 47.07900207900208, + "grad_norm": 8.632544631836936e-05, + "learning_rate": 0.11907449567041364, + "loss": 0.2707, + "num_input_tokens_seen": 17270296, + "step": 22645 + }, + { + "epoch": 47.08939708939709, + "grad_norm": 0.0003190569404978305, + "learning_rate": 0.11901685869550803, + "loss": 0.2564, + "num_input_tokens_seen": 17274104, + "step": 22650 + }, + { + "epoch": 47.0997920997921, + "grad_norm": 5.2312854677438736e-05, + "learning_rate": 0.1189592264985922, + "loss": 0.2432, + "num_input_tokens_seen": 17277848, + "step": 22655 + }, + { + "epoch": 47.11018711018711, + "grad_norm": 0.00035345632932148874, + "learning_rate": 0.11890159908855373, + "loss": 0.2804, + "num_input_tokens_seen": 17281688, + "step": 22660 + }, + { + "epoch": 47.12058212058212, + "grad_norm": 5.5318269005510956e-05, + "learning_rate": 0.11884397647427941, + "loss": 0.2714, + "num_input_tokens_seen": 17285432, + "step": 22665 + }, + { + "epoch": 47.13097713097713, + "grad_norm": 4.3209976865909994e-05, + "learning_rate": 0.11878635866465546, + "loss": 0.2578, + "num_input_tokens_seen": 17289112, + "step": 22670 + }, + { + "epoch": 47.141372141372145, + "grad_norm": 0.00018643047951627523, + "learning_rate": 0.11872874566856734, + "loss": 0.2663, + "num_input_tokens_seen": 17292856, + "step": 22675 + }, + { + "epoch": 47.15176715176715, + "grad_norm": 8.557530964026228e-05, + "learning_rate": 0.11867113749489955, + "loss": 0.2748, + "num_input_tokens_seen": 17296888, + "step": 22680 + }, + { + "epoch": 47.16216216216216, + "grad_norm": 0.00011160263966303319, + "learning_rate": 0.11861353415253607, + "loss": 0.2805, + "num_input_tokens_seen": 17300664, + "step": 22685 + }, + { + "epoch": 47.17255717255717, + "grad_norm": 0.00020285444043111056, + "learning_rate": 0.11855593565036011, + "loss": 0.2655, + "num_input_tokens_seen": 17304472, + "step": 22690 + }, + { + "epoch": 47.182952182952185, + "grad_norm": 0.00017236974963452667, + "learning_rate": 0.11849834199725394, + "loss": 0.2619, + "num_input_tokens_seen": 17308472, + "step": 22695 + }, + { + "epoch": 47.19334719334719, + "grad_norm": 0.0002156696718884632, + "learning_rate": 0.1184407532020994, + "loss": 0.2704, + "num_input_tokens_seen": 17312312, + "step": 22700 + }, + { + "epoch": 47.2037422037422, + "grad_norm": 7.599627133458853e-05, + "learning_rate": 0.11838316927377723, + "loss": 0.2385, + "num_input_tokens_seen": 17316216, + "step": 22705 + }, + { + "epoch": 47.21413721413722, + "grad_norm": 0.00021895587269682437, + "learning_rate": 0.11832559022116766, + "loss": 0.2873, + "num_input_tokens_seen": 17319960, + "step": 22710 + }, + { + "epoch": 47.224532224532226, + "grad_norm": 0.00022558463388122618, + "learning_rate": 0.11826801605315022, + "loss": 0.2727, + "num_input_tokens_seen": 17323768, + "step": 22715 + }, + { + "epoch": 47.234927234927234, + "grad_norm": 7.658988033654168e-05, + "learning_rate": 0.1182104467786034, + "loss": 0.2789, + "num_input_tokens_seen": 17327704, + "step": 22720 + }, + { + "epoch": 47.24532224532224, + "grad_norm": 0.000185832119314, + "learning_rate": 0.1181528824064052, + "loss": 0.2583, + "num_input_tokens_seen": 17331544, + "step": 22725 + }, + { + "epoch": 47.25571725571726, + "grad_norm": 6.409339403035119e-05, + "learning_rate": 0.11809532294543279, + "loss": 0.2793, + "num_input_tokens_seen": 17335480, + "step": 22730 + }, + { + "epoch": 47.266112266112266, + "grad_norm": 0.00033320419606752694, + "learning_rate": 0.11803776840456245, + "loss": 0.2652, + "num_input_tokens_seen": 17339416, + "step": 22735 + }, + { + "epoch": 47.276507276507274, + "grad_norm": 5.4971897043287754e-05, + "learning_rate": 0.11798021879266997, + "loss": 0.2806, + "num_input_tokens_seen": 17343192, + "step": 22740 + }, + { + "epoch": 47.28690228690229, + "grad_norm": 0.00010137514618691057, + "learning_rate": 0.11792267411863006, + "loss": 0.2747, + "num_input_tokens_seen": 17347000, + "step": 22745 + }, + { + "epoch": 47.2972972972973, + "grad_norm": 0.000202342082047835, + "learning_rate": 0.1178651343913169, + "loss": 0.2805, + "num_input_tokens_seen": 17350776, + "step": 22750 + }, + { + "epoch": 47.30769230769231, + "grad_norm": 0.00036567350616678596, + "learning_rate": 0.11780759961960392, + "loss": 0.2749, + "num_input_tokens_seen": 17354488, + "step": 22755 + }, + { + "epoch": 47.318087318087315, + "grad_norm": 0.0003085223725065589, + "learning_rate": 0.1177500698123636, + "loss": 0.2762, + "num_input_tokens_seen": 17358200, + "step": 22760 + }, + { + "epoch": 47.32848232848233, + "grad_norm": 0.00015022081788629293, + "learning_rate": 0.11769254497846778, + "loss": 0.2671, + "num_input_tokens_seen": 17361848, + "step": 22765 + }, + { + "epoch": 47.33887733887734, + "grad_norm": 0.0002670944668352604, + "learning_rate": 0.11763502512678758, + "loss": 0.2849, + "num_input_tokens_seen": 17365528, + "step": 22770 + }, + { + "epoch": 47.34927234927235, + "grad_norm": 0.00016507117834407836, + "learning_rate": 0.11757751026619315, + "loss": 0.2678, + "num_input_tokens_seen": 17369464, + "step": 22775 + }, + { + "epoch": 47.35966735966736, + "grad_norm": 0.0004100149089936167, + "learning_rate": 0.11752000040555416, + "loss": 0.2755, + "num_input_tokens_seen": 17373176, + "step": 22780 + }, + { + "epoch": 47.37006237006237, + "grad_norm": 0.0003952615079469979, + "learning_rate": 0.11746249555373921, + "loss": 0.266, + "num_input_tokens_seen": 17376952, + "step": 22785 + }, + { + "epoch": 47.38045738045738, + "grad_norm": 3.822296275757253e-05, + "learning_rate": 0.11740499571961638, + "loss": 0.2614, + "num_input_tokens_seen": 17380600, + "step": 22790 + }, + { + "epoch": 47.39085239085239, + "grad_norm": 0.0002612933167256415, + "learning_rate": 0.11734750091205279, + "loss": 0.2776, + "num_input_tokens_seen": 17384248, + "step": 22795 + }, + { + "epoch": 47.4012474012474, + "grad_norm": 0.00025843922048807144, + "learning_rate": 0.11729001113991493, + "loss": 0.2489, + "num_input_tokens_seen": 17388152, + "step": 22800 + }, + { + "epoch": 47.4012474012474, + "eval_loss": 0.24783852696418762, + "eval_runtime": 13.386, + "eval_samples_per_second": 63.947, + "eval_steps_per_second": 15.987, + "num_input_tokens_seen": 17388152, + "step": 22800 + }, + { + "epoch": 47.41164241164241, + "grad_norm": 0.00036767584970220923, + "learning_rate": 0.11723252641206837, + "loss": 0.2573, + "num_input_tokens_seen": 17391928, + "step": 22805 + }, + { + "epoch": 47.42203742203742, + "grad_norm": 0.0002765416575130075, + "learning_rate": 0.11717504673737808, + "loss": 0.2895, + "num_input_tokens_seen": 17395704, + "step": 22810 + }, + { + "epoch": 47.432432432432435, + "grad_norm": 9.059199510375038e-05, + "learning_rate": 0.11711757212470802, + "loss": 0.2707, + "num_input_tokens_seen": 17399416, + "step": 22815 + }, + { + "epoch": 47.44282744282744, + "grad_norm": 0.0003309416933916509, + "learning_rate": 0.11706010258292165, + "loss": 0.2722, + "num_input_tokens_seen": 17403192, + "step": 22820 + }, + { + "epoch": 47.45322245322245, + "grad_norm": 0.0001459595951018855, + "learning_rate": 0.11700263812088131, + "loss": 0.2677, + "num_input_tokens_seen": 17406904, + "step": 22825 + }, + { + "epoch": 47.46361746361746, + "grad_norm": 0.00016632139158900827, + "learning_rate": 0.11694517874744892, + "loss": 0.2555, + "num_input_tokens_seen": 17410744, + "step": 22830 + }, + { + "epoch": 47.474012474012476, + "grad_norm": 6.510930688818917e-05, + "learning_rate": 0.11688772447148532, + "loss": 0.2443, + "num_input_tokens_seen": 17414648, + "step": 22835 + }, + { + "epoch": 47.484407484407484, + "grad_norm": 0.00019824094488285482, + "learning_rate": 0.11683027530185074, + "loss": 0.2732, + "num_input_tokens_seen": 17418520, + "step": 22840 + }, + { + "epoch": 47.49480249480249, + "grad_norm": 0.0002203934855060652, + "learning_rate": 0.11677283124740451, + "loss": 0.2723, + "num_input_tokens_seen": 17422552, + "step": 22845 + }, + { + "epoch": 47.50519750519751, + "grad_norm": 4.549244113150053e-05, + "learning_rate": 0.11671539231700531, + "loss": 0.2519, + "num_input_tokens_seen": 17426296, + "step": 22850 + }, + { + "epoch": 47.515592515592516, + "grad_norm": 0.00019838643493130803, + "learning_rate": 0.11665795851951084, + "loss": 0.2469, + "num_input_tokens_seen": 17430104, + "step": 22855 + }, + { + "epoch": 47.525987525987524, + "grad_norm": 0.0006660513463430107, + "learning_rate": 0.11660052986377825, + "loss": 0.3004, + "num_input_tokens_seen": 17433912, + "step": 22860 + }, + { + "epoch": 47.53638253638254, + "grad_norm": 8.565449388697743e-05, + "learning_rate": 0.1165431063586636, + "loss": 0.2468, + "num_input_tokens_seen": 17437816, + "step": 22865 + }, + { + "epoch": 47.54677754677755, + "grad_norm": 8.209392399294302e-05, + "learning_rate": 0.11648568801302245, + "loss": 0.2544, + "num_input_tokens_seen": 17441656, + "step": 22870 + }, + { + "epoch": 47.55717255717256, + "grad_norm": 0.0003781815175898373, + "learning_rate": 0.11642827483570937, + "loss": 0.2558, + "num_input_tokens_seen": 17445400, + "step": 22875 + }, + { + "epoch": 47.567567567567565, + "grad_norm": 0.00015365683066193014, + "learning_rate": 0.11637086683557815, + "loss": 0.2635, + "num_input_tokens_seen": 17449272, + "step": 22880 + }, + { + "epoch": 47.57796257796258, + "grad_norm": 0.00021864441805519164, + "learning_rate": 0.11631346402148188, + "loss": 0.2781, + "num_input_tokens_seen": 17453016, + "step": 22885 + }, + { + "epoch": 47.58835758835759, + "grad_norm": 0.00017906819994095713, + "learning_rate": 0.11625606640227285, + "loss": 0.2688, + "num_input_tokens_seen": 17456984, + "step": 22890 + }, + { + "epoch": 47.5987525987526, + "grad_norm": 0.0004667652247007936, + "learning_rate": 0.11619867398680238, + "loss": 0.2536, + "num_input_tokens_seen": 17460856, + "step": 22895 + }, + { + "epoch": 47.60914760914761, + "grad_norm": 0.0004989434382878244, + "learning_rate": 0.11614128678392119, + "loss": 0.2686, + "num_input_tokens_seen": 17464536, + "step": 22900 + }, + { + "epoch": 47.61954261954262, + "grad_norm": 0.0002659035089891404, + "learning_rate": 0.11608390480247906, + "loss": 0.2552, + "num_input_tokens_seen": 17468376, + "step": 22905 + }, + { + "epoch": 47.62993762993763, + "grad_norm": 0.000250072218477726, + "learning_rate": 0.11602652805132499, + "loss": 0.2761, + "num_input_tokens_seen": 17472248, + "step": 22910 + }, + { + "epoch": 47.64033264033264, + "grad_norm": 0.00027879804838448763, + "learning_rate": 0.11596915653930731, + "loss": 0.2389, + "num_input_tokens_seen": 17476152, + "step": 22915 + }, + { + "epoch": 47.65072765072765, + "grad_norm": 8.028848242247477e-05, + "learning_rate": 0.11591179027527328, + "loss": 0.2543, + "num_input_tokens_seen": 17480056, + "step": 22920 + }, + { + "epoch": 47.66112266112266, + "grad_norm": 0.00028588916757144034, + "learning_rate": 0.11585442926806956, + "loss": 0.2311, + "num_input_tokens_seen": 17483768, + "step": 22925 + }, + { + "epoch": 47.67151767151767, + "grad_norm": 8.197624265449122e-05, + "learning_rate": 0.11579707352654202, + "loss": 0.2659, + "num_input_tokens_seen": 17487512, + "step": 22930 + }, + { + "epoch": 47.681912681912685, + "grad_norm": 0.00015521433670073748, + "learning_rate": 0.11573972305953548, + "loss": 0.2504, + "num_input_tokens_seen": 17491256, + "step": 22935 + }, + { + "epoch": 47.69230769230769, + "grad_norm": 0.00012910990335512906, + "learning_rate": 0.11568237787589426, + "loss": 0.2515, + "num_input_tokens_seen": 17495032, + "step": 22940 + }, + { + "epoch": 47.7027027027027, + "grad_norm": 0.0001905964018078521, + "learning_rate": 0.11562503798446161, + "loss": 0.2442, + "num_input_tokens_seen": 17498808, + "step": 22945 + }, + { + "epoch": 47.71309771309771, + "grad_norm": 0.00010095864126924425, + "learning_rate": 0.11556770339408005, + "loss": 0.2892, + "num_input_tokens_seen": 17502648, + "step": 22950 + }, + { + "epoch": 47.723492723492726, + "grad_norm": 0.00013444237993098795, + "learning_rate": 0.1155103741135914, + "loss": 0.2563, + "num_input_tokens_seen": 17506520, + "step": 22955 + }, + { + "epoch": 47.733887733887734, + "grad_norm": 4.4178672396810725e-05, + "learning_rate": 0.1154530501518364, + "loss": 0.2645, + "num_input_tokens_seen": 17510264, + "step": 22960 + }, + { + "epoch": 47.74428274428274, + "grad_norm": 5.697973392670974e-05, + "learning_rate": 0.11539573151765523, + "loss": 0.2674, + "num_input_tokens_seen": 17514232, + "step": 22965 + }, + { + "epoch": 47.75467775467776, + "grad_norm": 0.00017630786169320345, + "learning_rate": 0.11533841821988719, + "loss": 0.2449, + "num_input_tokens_seen": 17518104, + "step": 22970 + }, + { + "epoch": 47.765072765072766, + "grad_norm": 0.00025050199474208057, + "learning_rate": 0.11528111026737059, + "loss": 0.2833, + "num_input_tokens_seen": 17521848, + "step": 22975 + }, + { + "epoch": 47.775467775467774, + "grad_norm": 0.0002874883357435465, + "learning_rate": 0.11522380766894312, + "loss": 0.2657, + "num_input_tokens_seen": 17525624, + "step": 22980 + }, + { + "epoch": 47.78586278586278, + "grad_norm": 0.00010081563959829509, + "learning_rate": 0.11516651043344152, + "loss": 0.2707, + "num_input_tokens_seen": 17529496, + "step": 22985 + }, + { + "epoch": 47.7962577962578, + "grad_norm": 0.00041786671499721706, + "learning_rate": 0.11510921856970172, + "loss": 0.2834, + "num_input_tokens_seen": 17533240, + "step": 22990 + }, + { + "epoch": 47.80665280665281, + "grad_norm": 0.00025546379038132727, + "learning_rate": 0.11505193208655895, + "loss": 0.2668, + "num_input_tokens_seen": 17536984, + "step": 22995 + }, + { + "epoch": 47.817047817047815, + "grad_norm": 0.0004786074860021472, + "learning_rate": 0.11499465099284738, + "loss": 0.2575, + "num_input_tokens_seen": 17540824, + "step": 23000 + }, + { + "epoch": 47.817047817047815, + "eval_loss": 0.251009076833725, + "eval_runtime": 13.3981, + "eval_samples_per_second": 63.89, + "eval_steps_per_second": 15.972, + "num_input_tokens_seen": 17540824, + "step": 23000 + }, + { + "epoch": 47.82744282744283, + "grad_norm": 0.0003180828061886132, + "learning_rate": 0.1149373752974006, + "loss": 0.2523, + "num_input_tokens_seen": 17544696, + "step": 23005 + }, + { + "epoch": 47.83783783783784, + "grad_norm": 0.0008182342280633748, + "learning_rate": 0.11488010500905109, + "loss": 0.2746, + "num_input_tokens_seen": 17548472, + "step": 23010 + }, + { + "epoch": 47.84823284823285, + "grad_norm": 0.0004049630369991064, + "learning_rate": 0.11482284013663077, + "loss": 0.2793, + "num_input_tokens_seen": 17552152, + "step": 23015 + }, + { + "epoch": 47.858627858627855, + "grad_norm": 0.00036549242213368416, + "learning_rate": 0.11476558068897061, + "loss": 0.2777, + "num_input_tokens_seen": 17556056, + "step": 23020 + }, + { + "epoch": 47.86902286902287, + "grad_norm": 0.00013456857413984835, + "learning_rate": 0.11470832667490061, + "loss": 0.265, + "num_input_tokens_seen": 17559800, + "step": 23025 + }, + { + "epoch": 47.87941787941788, + "grad_norm": 0.0003066455537918955, + "learning_rate": 0.11465107810325013, + "loss": 0.2765, + "num_input_tokens_seen": 17563640, + "step": 23030 + }, + { + "epoch": 47.88981288981289, + "grad_norm": 0.000221941081690602, + "learning_rate": 0.11459383498284771, + "loss": 0.2782, + "num_input_tokens_seen": 17567384, + "step": 23035 + }, + { + "epoch": 47.9002079002079, + "grad_norm": 0.00026090172468684614, + "learning_rate": 0.11453659732252082, + "loss": 0.2637, + "num_input_tokens_seen": 17571224, + "step": 23040 + }, + { + "epoch": 47.91060291060291, + "grad_norm": 0.0005144834867678583, + "learning_rate": 0.11447936513109633, + "loss": 0.2416, + "num_input_tokens_seen": 17575000, + "step": 23045 + }, + { + "epoch": 47.92099792099792, + "grad_norm": 0.00011775966413551942, + "learning_rate": 0.11442213841740011, + "loss": 0.2657, + "num_input_tokens_seen": 17578776, + "step": 23050 + }, + { + "epoch": 47.931392931392935, + "grad_norm": 0.00058674969477579, + "learning_rate": 0.1143649171902572, + "loss": 0.287, + "num_input_tokens_seen": 17582776, + "step": 23055 + }, + { + "epoch": 47.94178794178794, + "grad_norm": 0.00017124043370131403, + "learning_rate": 0.11430770145849194, + "loss": 0.2658, + "num_input_tokens_seen": 17586680, + "step": 23060 + }, + { + "epoch": 47.95218295218295, + "grad_norm": 0.00039672275306656957, + "learning_rate": 0.11425049123092756, + "loss": 0.2675, + "num_input_tokens_seen": 17590488, + "step": 23065 + }, + { + "epoch": 47.96257796257796, + "grad_norm": 0.0003632518055383116, + "learning_rate": 0.11419328651638674, + "loss": 0.2606, + "num_input_tokens_seen": 17594264, + "step": 23070 + }, + { + "epoch": 47.972972972972975, + "grad_norm": 0.00014396029291674495, + "learning_rate": 0.11413608732369115, + "loss": 0.264, + "num_input_tokens_seen": 17598040, + "step": 23075 + }, + { + "epoch": 47.983367983367984, + "grad_norm": 0.00018414489750284702, + "learning_rate": 0.11407889366166153, + "loss": 0.2557, + "num_input_tokens_seen": 17602040, + "step": 23080 + }, + { + "epoch": 47.99376299376299, + "grad_norm": 0.00019989210704807192, + "learning_rate": 0.11402170553911797, + "loss": 0.2577, + "num_input_tokens_seen": 17605976, + "step": 23085 + }, + { + "epoch": 48.00415800415801, + "grad_norm": 0.00023254321422427893, + "learning_rate": 0.11396452296487955, + "loss": 0.2626, + "num_input_tokens_seen": 17609752, + "step": 23090 + }, + { + "epoch": 48.014553014553016, + "grad_norm": 0.00030729518039152026, + "learning_rate": 0.11390734594776449, + "loss": 0.2802, + "num_input_tokens_seen": 17613656, + "step": 23095 + }, + { + "epoch": 48.024948024948024, + "grad_norm": 5.417638749349862e-05, + "learning_rate": 0.11385017449659031, + "loss": 0.2817, + "num_input_tokens_seen": 17617592, + "step": 23100 + }, + { + "epoch": 48.03534303534303, + "grad_norm": 8.244667696999386e-05, + "learning_rate": 0.11379300862017344, + "loss": 0.2788, + "num_input_tokens_seen": 17621464, + "step": 23105 + }, + { + "epoch": 48.04573804573805, + "grad_norm": 0.0002353794261580333, + "learning_rate": 0.11373584832732966, + "loss": 0.2764, + "num_input_tokens_seen": 17625240, + "step": 23110 + }, + { + "epoch": 48.056133056133056, + "grad_norm": 0.00035346869844943285, + "learning_rate": 0.11367869362687386, + "loss": 0.2576, + "num_input_tokens_seen": 17629048, + "step": 23115 + }, + { + "epoch": 48.066528066528065, + "grad_norm": 0.00018778469529934227, + "learning_rate": 0.11362154452761988, + "loss": 0.2644, + "num_input_tokens_seen": 17632920, + "step": 23120 + }, + { + "epoch": 48.07692307692308, + "grad_norm": 0.00020919187227264047, + "learning_rate": 0.11356440103838095, + "loss": 0.2629, + "num_input_tokens_seen": 17636664, + "step": 23125 + }, + { + "epoch": 48.08731808731809, + "grad_norm": 0.00013233014033176005, + "learning_rate": 0.11350726316796922, + "loss": 0.318, + "num_input_tokens_seen": 17640632, + "step": 23130 + }, + { + "epoch": 48.0977130977131, + "grad_norm": 0.0001491636357968673, + "learning_rate": 0.11345013092519607, + "loss": 0.2836, + "num_input_tokens_seen": 17644440, + "step": 23135 + }, + { + "epoch": 48.108108108108105, + "grad_norm": 0.0004155202186666429, + "learning_rate": 0.11339300431887213, + "loss": 0.2812, + "num_input_tokens_seen": 17648376, + "step": 23140 + }, + { + "epoch": 48.11850311850312, + "grad_norm": 0.0001782684266800061, + "learning_rate": 0.11333588335780687, + "loss": 0.2826, + "num_input_tokens_seen": 17652312, + "step": 23145 + }, + { + "epoch": 48.12889812889813, + "grad_norm": 0.00014385893882717937, + "learning_rate": 0.11327876805080916, + "loss": 0.2722, + "num_input_tokens_seen": 17655960, + "step": 23150 + }, + { + "epoch": 48.13929313929314, + "grad_norm": 0.0001330423983745277, + "learning_rate": 0.11322165840668696, + "loss": 0.2744, + "num_input_tokens_seen": 17659672, + "step": 23155 + }, + { + "epoch": 48.14968814968815, + "grad_norm": 0.0007657487876713276, + "learning_rate": 0.11316455443424717, + "loss": 0.2575, + "num_input_tokens_seen": 17663384, + "step": 23160 + }, + { + "epoch": 48.16008316008316, + "grad_norm": 7.932808512123302e-05, + "learning_rate": 0.11310745614229603, + "loss": 0.2643, + "num_input_tokens_seen": 17667320, + "step": 23165 + }, + { + "epoch": 48.17047817047817, + "grad_norm": 0.00019462747150100768, + "learning_rate": 0.1130503635396387, + "loss": 0.2686, + "num_input_tokens_seen": 17671128, + "step": 23170 + }, + { + "epoch": 48.18087318087318, + "grad_norm": 0.00011662027100101113, + "learning_rate": 0.11299327663507966, + "loss": 0.263, + "num_input_tokens_seen": 17674872, + "step": 23175 + }, + { + "epoch": 48.19126819126819, + "grad_norm": 0.00014338426990434527, + "learning_rate": 0.11293619543742246, + "loss": 0.2695, + "num_input_tokens_seen": 17678808, + "step": 23180 + }, + { + "epoch": 48.2016632016632, + "grad_norm": 9.592835704097524e-05, + "learning_rate": 0.11287911995546965, + "loss": 0.2699, + "num_input_tokens_seen": 17682616, + "step": 23185 + }, + { + "epoch": 48.21205821205821, + "grad_norm": 0.0005872222245670855, + "learning_rate": 0.11282205019802308, + "loss": 0.2527, + "num_input_tokens_seen": 17686488, + "step": 23190 + }, + { + "epoch": 48.222453222453225, + "grad_norm": 0.0003532466071192175, + "learning_rate": 0.11276498617388354, + "loss": 0.2434, + "num_input_tokens_seen": 17690264, + "step": 23195 + }, + { + "epoch": 48.232848232848234, + "grad_norm": 0.00022323931625578552, + "learning_rate": 0.11270792789185109, + "loss": 0.2172, + "num_input_tokens_seen": 17693912, + "step": 23200 + }, + { + "epoch": 48.232848232848234, + "eval_loss": 0.25092241168022156, + "eval_runtime": 13.3938, + "eval_samples_per_second": 63.91, + "eval_steps_per_second": 15.977, + "num_input_tokens_seen": 17693912, + "step": 23200 + }, + { + "epoch": 48.24324324324324, + "grad_norm": 0.0005901767290197313, + "learning_rate": 0.11265087536072482, + "loss": 0.2852, + "num_input_tokens_seen": 17697752, + "step": 23205 + }, + { + "epoch": 48.25363825363825, + "grad_norm": 7.44577992008999e-05, + "learning_rate": 0.11259382858930288, + "loss": 0.2625, + "num_input_tokens_seen": 17701496, + "step": 23210 + }, + { + "epoch": 48.264033264033266, + "grad_norm": 8.658089791424572e-05, + "learning_rate": 0.11253678758638262, + "loss": 0.2566, + "num_input_tokens_seen": 17705432, + "step": 23215 + }, + { + "epoch": 48.274428274428274, + "grad_norm": 0.00031267423764802516, + "learning_rate": 0.11247975236076059, + "loss": 0.266, + "num_input_tokens_seen": 17709208, + "step": 23220 + }, + { + "epoch": 48.28482328482328, + "grad_norm": 0.00010747231135610491, + "learning_rate": 0.11242272292123218, + "loss": 0.2787, + "num_input_tokens_seen": 17713176, + "step": 23225 + }, + { + "epoch": 48.2952182952183, + "grad_norm": 0.0002441554388497025, + "learning_rate": 0.11236569927659217, + "loss": 0.2448, + "num_input_tokens_seen": 17716952, + "step": 23230 + }, + { + "epoch": 48.305613305613306, + "grad_norm": 9.575235890224576e-05, + "learning_rate": 0.11230868143563429, + "loss": 0.2276, + "num_input_tokens_seen": 17720760, + "step": 23235 + }, + { + "epoch": 48.316008316008315, + "grad_norm": 0.00047440320486202836, + "learning_rate": 0.11225166940715131, + "loss": 0.2973, + "num_input_tokens_seen": 17724568, + "step": 23240 + }, + { + "epoch": 48.32640332640332, + "grad_norm": 0.00022464468202088028, + "learning_rate": 0.11219466319993537, + "loss": 0.2286, + "num_input_tokens_seen": 17728344, + "step": 23245 + }, + { + "epoch": 48.33679833679834, + "grad_norm": 0.0004839448956772685, + "learning_rate": 0.11213766282277739, + "loss": 0.2855, + "num_input_tokens_seen": 17732248, + "step": 23250 + }, + { + "epoch": 48.34719334719335, + "grad_norm": 5.8999121392844245e-05, + "learning_rate": 0.11208066828446761, + "loss": 0.2407, + "num_input_tokens_seen": 17736248, + "step": 23255 + }, + { + "epoch": 48.357588357588355, + "grad_norm": 0.00030349905136972666, + "learning_rate": 0.11202367959379537, + "loss": 0.26, + "num_input_tokens_seen": 17740056, + "step": 23260 + }, + { + "epoch": 48.36798336798337, + "grad_norm": 0.0002658960875123739, + "learning_rate": 0.11196669675954894, + "loss": 0.2766, + "num_input_tokens_seen": 17743832, + "step": 23265 + }, + { + "epoch": 48.37837837837838, + "grad_norm": 0.0005035450449213386, + "learning_rate": 0.1119097197905158, + "loss": 0.2682, + "num_input_tokens_seen": 17747608, + "step": 23270 + }, + { + "epoch": 48.38877338877339, + "grad_norm": 0.0001554350310470909, + "learning_rate": 0.11185274869548259, + "loss": 0.2652, + "num_input_tokens_seen": 17751352, + "step": 23275 + }, + { + "epoch": 48.3991683991684, + "grad_norm": 0.00022462457127403468, + "learning_rate": 0.11179578348323486, + "loss": 0.2539, + "num_input_tokens_seen": 17755096, + "step": 23280 + }, + { + "epoch": 48.40956340956341, + "grad_norm": 0.00021113056573085487, + "learning_rate": 0.1117388241625575, + "loss": 0.2685, + "num_input_tokens_seen": 17758872, + "step": 23285 + }, + { + "epoch": 48.41995841995842, + "grad_norm": 0.00027132206014357507, + "learning_rate": 0.11168187074223421, + "loss": 0.2534, + "num_input_tokens_seen": 17762744, + "step": 23290 + }, + { + "epoch": 48.43035343035343, + "grad_norm": 0.00011608129716478288, + "learning_rate": 0.11162492323104796, + "loss": 0.2681, + "num_input_tokens_seen": 17766648, + "step": 23295 + }, + { + "epoch": 48.44074844074844, + "grad_norm": 8.047799929045141e-05, + "learning_rate": 0.11156798163778091, + "loss": 0.2618, + "num_input_tokens_seen": 17770392, + "step": 23300 + }, + { + "epoch": 48.45114345114345, + "grad_norm": 0.00026496706414036453, + "learning_rate": 0.11151104597121399, + "loss": 0.26, + "num_input_tokens_seen": 17774264, + "step": 23305 + }, + { + "epoch": 48.46153846153846, + "grad_norm": 0.00016956462059170008, + "learning_rate": 0.11145411624012742, + "loss": 0.2732, + "num_input_tokens_seen": 17778200, + "step": 23310 + }, + { + "epoch": 48.471933471933475, + "grad_norm": 9.155752923106775e-05, + "learning_rate": 0.11139719245330063, + "loss": 0.2793, + "num_input_tokens_seen": 17782104, + "step": 23315 + }, + { + "epoch": 48.482328482328484, + "grad_norm": 0.0002794752363115549, + "learning_rate": 0.11134027461951179, + "loss": 0.2746, + "num_input_tokens_seen": 17785944, + "step": 23320 + }, + { + "epoch": 48.49272349272349, + "grad_norm": 0.0001991379540413618, + "learning_rate": 0.11128336274753849, + "loss": 0.2519, + "num_input_tokens_seen": 17789720, + "step": 23325 + }, + { + "epoch": 48.5031185031185, + "grad_norm": 0.0002663952764123678, + "learning_rate": 0.11122645684615715, + "loss": 0.2371, + "num_input_tokens_seen": 17793528, + "step": 23330 + }, + { + "epoch": 48.513513513513516, + "grad_norm": 0.00020975255756638944, + "learning_rate": 0.11116955692414345, + "loss": 0.261, + "num_input_tokens_seen": 17797336, + "step": 23335 + }, + { + "epoch": 48.523908523908524, + "grad_norm": 0.00032962270779535174, + "learning_rate": 0.11111266299027203, + "loss": 0.2886, + "num_input_tokens_seen": 17801208, + "step": 23340 + }, + { + "epoch": 48.53430353430353, + "grad_norm": 9.004686580738053e-05, + "learning_rate": 0.11105577505331668, + "loss": 0.2651, + "num_input_tokens_seen": 17805016, + "step": 23345 + }, + { + "epoch": 48.54469854469855, + "grad_norm": 8.469255408272147e-05, + "learning_rate": 0.11099889312205018, + "loss": 0.2646, + "num_input_tokens_seen": 17808760, + "step": 23350 + }, + { + "epoch": 48.555093555093556, + "grad_norm": 0.00032976423972286284, + "learning_rate": 0.11094201720524455, + "loss": 0.2594, + "num_input_tokens_seen": 17812472, + "step": 23355 + }, + { + "epoch": 48.565488565488565, + "grad_norm": 0.0001228516921401024, + "learning_rate": 0.11088514731167064, + "loss": 0.2511, + "num_input_tokens_seen": 17816152, + "step": 23360 + }, + { + "epoch": 48.57588357588357, + "grad_norm": 0.00020506713190115988, + "learning_rate": 0.11082828345009862, + "loss": 0.2847, + "num_input_tokens_seen": 17819992, + "step": 23365 + }, + { + "epoch": 48.58627858627859, + "grad_norm": 0.00025999650824815035, + "learning_rate": 0.11077142562929748, + "loss": 0.2572, + "num_input_tokens_seen": 17823800, + "step": 23370 + }, + { + "epoch": 48.5966735966736, + "grad_norm": 0.00010288695193594322, + "learning_rate": 0.11071457385803554, + "loss": 0.2734, + "num_input_tokens_seen": 17827576, + "step": 23375 + }, + { + "epoch": 48.607068607068605, + "grad_norm": 0.00025981085491366684, + "learning_rate": 0.11065772814508001, + "loss": 0.2503, + "num_input_tokens_seen": 17831384, + "step": 23380 + }, + { + "epoch": 48.61746361746362, + "grad_norm": 0.00020400645735207945, + "learning_rate": 0.11060088849919715, + "loss": 0.2829, + "num_input_tokens_seen": 17835160, + "step": 23385 + }, + { + "epoch": 48.62785862785863, + "grad_norm": 0.0005605733022093773, + "learning_rate": 0.11054405492915244, + "loss": 0.27, + "num_input_tokens_seen": 17838840, + "step": 23390 + }, + { + "epoch": 48.63825363825364, + "grad_norm": 0.00032936997013166547, + "learning_rate": 0.11048722744371031, + "loss": 0.2396, + "num_input_tokens_seen": 17842584, + "step": 23395 + }, + { + "epoch": 48.648648648648646, + "grad_norm": 9.78625103016384e-05, + "learning_rate": 0.1104304060516342, + "loss": 0.2799, + "num_input_tokens_seen": 17846296, + "step": 23400 + }, + { + "epoch": 48.648648648648646, + "eval_loss": 0.2471221536397934, + "eval_runtime": 13.3803, + "eval_samples_per_second": 63.975, + "eval_steps_per_second": 15.994, + "num_input_tokens_seen": 17846296, + "step": 23400 + }, + { + "epoch": 48.65904365904366, + "grad_norm": 0.00019502625218592584, + "learning_rate": 0.11037359076168682, + "loss": 0.2664, + "num_input_tokens_seen": 17850424, + "step": 23405 + }, + { + "epoch": 48.66943866943867, + "grad_norm": 0.00023033835168462247, + "learning_rate": 0.11031678158262966, + "loss": 0.2524, + "num_input_tokens_seen": 17854296, + "step": 23410 + }, + { + "epoch": 48.67983367983368, + "grad_norm": 0.00046526658115908504, + "learning_rate": 0.11025997852322349, + "loss": 0.2632, + "num_input_tokens_seen": 17858008, + "step": 23415 + }, + { + "epoch": 48.69022869022869, + "grad_norm": 0.0001972712343558669, + "learning_rate": 0.11020318159222807, + "loss": 0.2887, + "num_input_tokens_seen": 17861848, + "step": 23420 + }, + { + "epoch": 48.7006237006237, + "grad_norm": 0.0003987428790424019, + "learning_rate": 0.1101463907984021, + "loss": 0.2798, + "num_input_tokens_seen": 17865720, + "step": 23425 + }, + { + "epoch": 48.71101871101871, + "grad_norm": 0.0002829173463396728, + "learning_rate": 0.11008960615050352, + "loss": 0.2639, + "num_input_tokens_seen": 17869592, + "step": 23430 + }, + { + "epoch": 48.72141372141372, + "grad_norm": 0.00013211939949542284, + "learning_rate": 0.11003282765728925, + "loss": 0.2692, + "num_input_tokens_seen": 17873368, + "step": 23435 + }, + { + "epoch": 48.731808731808734, + "grad_norm": 0.000244815309997648, + "learning_rate": 0.10997605532751518, + "loss": 0.2435, + "num_input_tokens_seen": 17877176, + "step": 23440 + }, + { + "epoch": 48.74220374220374, + "grad_norm": 0.00041402026545256376, + "learning_rate": 0.1099192891699364, + "loss": 0.2901, + "num_input_tokens_seen": 17880824, + "step": 23445 + }, + { + "epoch": 48.75259875259875, + "grad_norm": 0.0005632441025227308, + "learning_rate": 0.10986252919330687, + "loss": 0.3007, + "num_input_tokens_seen": 17884632, + "step": 23450 + }, + { + "epoch": 48.762993762993766, + "grad_norm": 0.00038054314791224897, + "learning_rate": 0.10980577540637973, + "loss": 0.2627, + "num_input_tokens_seen": 17888568, + "step": 23455 + }, + { + "epoch": 48.773388773388774, + "grad_norm": 0.00019710694323293865, + "learning_rate": 0.10974902781790719, + "loss": 0.256, + "num_input_tokens_seen": 17892280, + "step": 23460 + }, + { + "epoch": 48.78378378378378, + "grad_norm": 0.00022718824038747698, + "learning_rate": 0.10969228643664032, + "loss": 0.2644, + "num_input_tokens_seen": 17895992, + "step": 23465 + }, + { + "epoch": 48.79417879417879, + "grad_norm": 0.00016650119505356997, + "learning_rate": 0.10963555127132942, + "loss": 0.2628, + "num_input_tokens_seen": 17899928, + "step": 23470 + }, + { + "epoch": 48.804573804573806, + "grad_norm": 0.00012927359784953296, + "learning_rate": 0.10957882233072382, + "loss": 0.2915, + "num_input_tokens_seen": 17903736, + "step": 23475 + }, + { + "epoch": 48.814968814968815, + "grad_norm": 0.0006165364757180214, + "learning_rate": 0.10952209962357176, + "loss": 0.2735, + "num_input_tokens_seen": 17907448, + "step": 23480 + }, + { + "epoch": 48.82536382536382, + "grad_norm": 0.0004893646691925824, + "learning_rate": 0.10946538315862062, + "loss": 0.2459, + "num_input_tokens_seen": 17911288, + "step": 23485 + }, + { + "epoch": 48.83575883575884, + "grad_norm": 0.00021871081844437867, + "learning_rate": 0.10940867294461679, + "loss": 0.2382, + "num_input_tokens_seen": 17915032, + "step": 23490 + }, + { + "epoch": 48.84615384615385, + "grad_norm": 8.697227895027027e-05, + "learning_rate": 0.10935196899030565, + "loss": 0.2648, + "num_input_tokens_seen": 17918904, + "step": 23495 + }, + { + "epoch": 48.856548856548855, + "grad_norm": 0.0001984906557481736, + "learning_rate": 0.10929527130443177, + "loss": 0.2817, + "num_input_tokens_seen": 17922680, + "step": 23500 + }, + { + "epoch": 48.86694386694387, + "grad_norm": 0.00023849379795137793, + "learning_rate": 0.1092385798957385, + "loss": 0.2829, + "num_input_tokens_seen": 17926488, + "step": 23505 + }, + { + "epoch": 48.87733887733888, + "grad_norm": 0.00038961332757025957, + "learning_rate": 0.10918189477296848, + "loss": 0.2705, + "num_input_tokens_seen": 17930296, + "step": 23510 + }, + { + "epoch": 48.88773388773389, + "grad_norm": 7.555889169452712e-05, + "learning_rate": 0.1091252159448633, + "loss": 0.2771, + "num_input_tokens_seen": 17933944, + "step": 23515 + }, + { + "epoch": 48.898128898128896, + "grad_norm": 0.00027400473481975496, + "learning_rate": 0.10906854342016345, + "loss": 0.2724, + "num_input_tokens_seen": 17937720, + "step": 23520 + }, + { + "epoch": 48.90852390852391, + "grad_norm": 0.0003160927153658122, + "learning_rate": 0.10901187720760858, + "loss": 0.2696, + "num_input_tokens_seen": 17941496, + "step": 23525 + }, + { + "epoch": 48.91891891891892, + "grad_norm": 5.8038014685735106e-05, + "learning_rate": 0.10895521731593734, + "loss": 0.2712, + "num_input_tokens_seen": 17945368, + "step": 23530 + }, + { + "epoch": 48.92931392931393, + "grad_norm": 0.0006155946757644415, + "learning_rate": 0.10889856375388733, + "loss": 0.2341, + "num_input_tokens_seen": 17949112, + "step": 23535 + }, + { + "epoch": 48.93970893970894, + "grad_norm": 0.00010973793541779742, + "learning_rate": 0.1088419165301954, + "loss": 0.2541, + "num_input_tokens_seen": 17952984, + "step": 23540 + }, + { + "epoch": 48.95010395010395, + "grad_norm": 0.00017801918147597462, + "learning_rate": 0.1087852756535971, + "loss": 0.268, + "num_input_tokens_seen": 17956824, + "step": 23545 + }, + { + "epoch": 48.96049896049896, + "grad_norm": 9.488354407949373e-05, + "learning_rate": 0.10872864113282725, + "loss": 0.3091, + "num_input_tokens_seen": 17960728, + "step": 23550 + }, + { + "epoch": 48.97089397089397, + "grad_norm": 8.429075387539342e-05, + "learning_rate": 0.10867201297661958, + "loss": 0.2603, + "num_input_tokens_seen": 17964536, + "step": 23555 + }, + { + "epoch": 48.981288981288984, + "grad_norm": 0.0001953737810254097, + "learning_rate": 0.10861539119370689, + "loss": 0.2564, + "num_input_tokens_seen": 17968440, + "step": 23560 + }, + { + "epoch": 48.99168399168399, + "grad_norm": 0.00023910404706839472, + "learning_rate": 0.10855877579282096, + "loss": 0.2613, + "num_input_tokens_seen": 17972152, + "step": 23565 + }, + { + "epoch": 49.002079002079, + "grad_norm": 0.0002669554087333381, + "learning_rate": 0.10850216678269252, + "loss": 0.266, + "num_input_tokens_seen": 17975944, + "step": 23570 + }, + { + "epoch": 49.012474012474016, + "grad_norm": 0.0006303474074229598, + "learning_rate": 0.10844556417205146, + "loss": 0.3078, + "num_input_tokens_seen": 17979752, + "step": 23575 + }, + { + "epoch": 49.022869022869024, + "grad_norm": 0.00017910836322698742, + "learning_rate": 0.10838896796962669, + "loss": 0.2597, + "num_input_tokens_seen": 17983496, + "step": 23580 + }, + { + "epoch": 49.03326403326403, + "grad_norm": 0.00041906916885636747, + "learning_rate": 0.1083323781841459, + "loss": 0.2715, + "num_input_tokens_seen": 17987176, + "step": 23585 + }, + { + "epoch": 49.04365904365904, + "grad_norm": 0.0002511486236471683, + "learning_rate": 0.10827579482433607, + "loss": 0.2717, + "num_input_tokens_seen": 17991112, + "step": 23590 + }, + { + "epoch": 49.054054054054056, + "grad_norm": 5.151348523213528e-05, + "learning_rate": 0.10821921789892304, + "loss": 0.2707, + "num_input_tokens_seen": 17994920, + "step": 23595 + }, + { + "epoch": 49.064449064449065, + "grad_norm": 0.0002489004691597074, + "learning_rate": 0.10816264741663158, + "loss": 0.2467, + "num_input_tokens_seen": 17998760, + "step": 23600 + }, + { + "epoch": 49.064449064449065, + "eval_loss": 0.2500828802585602, + "eval_runtime": 13.3946, + "eval_samples_per_second": 63.906, + "eval_steps_per_second": 15.977, + "num_input_tokens_seen": 17998760, + "step": 23600 + }, + { + "epoch": 49.07484407484407, + "grad_norm": 0.0001256687828572467, + "learning_rate": 0.10810608338618573, + "loss": 0.2684, + "num_input_tokens_seen": 18002600, + "step": 23605 + }, + { + "epoch": 49.08523908523909, + "grad_norm": 0.000493725819978863, + "learning_rate": 0.10804952581630821, + "loss": 0.2689, + "num_input_tokens_seen": 18006536, + "step": 23610 + }, + { + "epoch": 49.0956340956341, + "grad_norm": 0.00014630363148171455, + "learning_rate": 0.10799297471572102, + "loss": 0.2493, + "num_input_tokens_seen": 18010504, + "step": 23615 + }, + { + "epoch": 49.106029106029105, + "grad_norm": 4.367139263194986e-05, + "learning_rate": 0.10793643009314507, + "loss": 0.2658, + "num_input_tokens_seen": 18014152, + "step": 23620 + }, + { + "epoch": 49.11642411642411, + "grad_norm": 6.575254519702867e-05, + "learning_rate": 0.10787989195730015, + "loss": 0.2576, + "num_input_tokens_seen": 18018120, + "step": 23625 + }, + { + "epoch": 49.12681912681913, + "grad_norm": 0.0002481926931068301, + "learning_rate": 0.10782336031690525, + "loss": 0.2816, + "num_input_tokens_seen": 18021928, + "step": 23630 + }, + { + "epoch": 49.13721413721414, + "grad_norm": 0.00010424883657833561, + "learning_rate": 0.10776683518067821, + "loss": 0.2593, + "num_input_tokens_seen": 18025864, + "step": 23635 + }, + { + "epoch": 49.147609147609145, + "grad_norm": 0.00010949529678327963, + "learning_rate": 0.10771031655733587, + "loss": 0.2535, + "num_input_tokens_seen": 18029672, + "step": 23640 + }, + { + "epoch": 49.15800415800416, + "grad_norm": 0.00022108136909082532, + "learning_rate": 0.10765380445559422, + "loss": 0.2812, + "num_input_tokens_seen": 18033672, + "step": 23645 + }, + { + "epoch": 49.16839916839917, + "grad_norm": 0.00029403407825157046, + "learning_rate": 0.10759729888416801, + "loss": 0.2787, + "num_input_tokens_seen": 18037576, + "step": 23650 + }, + { + "epoch": 49.17879417879418, + "grad_norm": 9.798179962672293e-05, + "learning_rate": 0.10754079985177119, + "loss": 0.2593, + "num_input_tokens_seen": 18041288, + "step": 23655 + }, + { + "epoch": 49.189189189189186, + "grad_norm": 0.00022263005666900426, + "learning_rate": 0.10748430736711667, + "loss": 0.2696, + "num_input_tokens_seen": 18045128, + "step": 23660 + }, + { + "epoch": 49.1995841995842, + "grad_norm": 0.0002532023936510086, + "learning_rate": 0.10742782143891623, + "loss": 0.2423, + "num_input_tokens_seen": 18048968, + "step": 23665 + }, + { + "epoch": 49.20997920997921, + "grad_norm": 0.00018222650396637619, + "learning_rate": 0.10737134207588069, + "loss": 0.2709, + "num_input_tokens_seen": 18052808, + "step": 23670 + }, + { + "epoch": 49.22037422037422, + "grad_norm": 0.00020528025925159454, + "learning_rate": 0.10731486928671992, + "loss": 0.2744, + "num_input_tokens_seen": 18056680, + "step": 23675 + }, + { + "epoch": 49.23076923076923, + "grad_norm": 7.971689046826214e-05, + "learning_rate": 0.10725840308014269, + "loss": 0.2529, + "num_input_tokens_seen": 18060488, + "step": 23680 + }, + { + "epoch": 49.24116424116424, + "grad_norm": 8.92926545930095e-05, + "learning_rate": 0.10720194346485688, + "loss": 0.2646, + "num_input_tokens_seen": 18064136, + "step": 23685 + }, + { + "epoch": 49.25155925155925, + "grad_norm": 0.00013779282744508237, + "learning_rate": 0.10714549044956918, + "loss": 0.2721, + "num_input_tokens_seen": 18067912, + "step": 23690 + }, + { + "epoch": 49.26195426195426, + "grad_norm": 0.00032711660605855286, + "learning_rate": 0.10708904404298542, + "loss": 0.2601, + "num_input_tokens_seen": 18071848, + "step": 23695 + }, + { + "epoch": 49.272349272349274, + "grad_norm": 8.303031791001558e-05, + "learning_rate": 0.1070326042538103, + "loss": 0.258, + "num_input_tokens_seen": 18075656, + "step": 23700 + }, + { + "epoch": 49.28274428274428, + "grad_norm": 7.937040936667472e-05, + "learning_rate": 0.10697617109074758, + "loss": 0.2478, + "num_input_tokens_seen": 18079464, + "step": 23705 + }, + { + "epoch": 49.29313929313929, + "grad_norm": 4.8353205784223974e-05, + "learning_rate": 0.10691974456249999, + "loss": 0.2765, + "num_input_tokens_seen": 18083240, + "step": 23710 + }, + { + "epoch": 49.303534303534306, + "grad_norm": 0.0004748390056192875, + "learning_rate": 0.10686332467776909, + "loss": 0.2424, + "num_input_tokens_seen": 18087112, + "step": 23715 + }, + { + "epoch": 49.313929313929314, + "grad_norm": 0.00023989638430066407, + "learning_rate": 0.10680691144525563, + "loss": 0.262, + "num_input_tokens_seen": 18090920, + "step": 23720 + }, + { + "epoch": 49.32432432432432, + "grad_norm": 9.276289347326383e-05, + "learning_rate": 0.10675050487365928, + "loss": 0.2614, + "num_input_tokens_seen": 18094760, + "step": 23725 + }, + { + "epoch": 49.33471933471934, + "grad_norm": 0.00012683146633207798, + "learning_rate": 0.10669410497167851, + "loss": 0.298, + "num_input_tokens_seen": 18098536, + "step": 23730 + }, + { + "epoch": 49.34511434511435, + "grad_norm": 0.0004436123126652092, + "learning_rate": 0.10663771174801102, + "loss": 0.2703, + "num_input_tokens_seen": 18102312, + "step": 23735 + }, + { + "epoch": 49.355509355509355, + "grad_norm": 0.0002383785613346845, + "learning_rate": 0.10658132521135329, + "loss": 0.279, + "num_input_tokens_seen": 18106056, + "step": 23740 + }, + { + "epoch": 49.36590436590436, + "grad_norm": 0.0001367150543956086, + "learning_rate": 0.10652494537040084, + "loss": 0.2703, + "num_input_tokens_seen": 18109928, + "step": 23745 + }, + { + "epoch": 49.37629937629938, + "grad_norm": 0.00024961400777101517, + "learning_rate": 0.1064685722338482, + "loss": 0.263, + "num_input_tokens_seen": 18113896, + "step": 23750 + }, + { + "epoch": 49.38669438669439, + "grad_norm": 0.00012926464842166752, + "learning_rate": 0.10641220581038871, + "loss": 0.219, + "num_input_tokens_seen": 18117768, + "step": 23755 + }, + { + "epoch": 49.397089397089395, + "grad_norm": 0.00019978737691417336, + "learning_rate": 0.10635584610871483, + "loss": 0.2577, + "num_input_tokens_seen": 18121512, + "step": 23760 + }, + { + "epoch": 49.40748440748441, + "grad_norm": 0.0003638410125859082, + "learning_rate": 0.10629949313751803, + "loss": 0.2554, + "num_input_tokens_seen": 18125288, + "step": 23765 + }, + { + "epoch": 49.41787941787942, + "grad_norm": 0.00018110027303919196, + "learning_rate": 0.10624314690548849, + "loss": 0.2733, + "num_input_tokens_seen": 18129064, + "step": 23770 + }, + { + "epoch": 49.42827442827443, + "grad_norm": 9.28209483390674e-05, + "learning_rate": 0.1061868074213156, + "loss": 0.263, + "num_input_tokens_seen": 18133064, + "step": 23775 + }, + { + "epoch": 49.438669438669436, + "grad_norm": 2.908603892137762e-05, + "learning_rate": 0.10613047469368765, + "loss": 0.2746, + "num_input_tokens_seen": 18136712, + "step": 23780 + }, + { + "epoch": 49.44906444906445, + "grad_norm": 8.971554780146107e-05, + "learning_rate": 0.10607414873129171, + "loss": 0.2628, + "num_input_tokens_seen": 18140552, + "step": 23785 + }, + { + "epoch": 49.45945945945946, + "grad_norm": 0.0001949443540070206, + "learning_rate": 0.10601782954281413, + "loss": 0.2558, + "num_input_tokens_seen": 18144328, + "step": 23790 + }, + { + "epoch": 49.46985446985447, + "grad_norm": 0.0002551834040787071, + "learning_rate": 0.1059615171369399, + "loss": 0.2715, + "num_input_tokens_seen": 18148200, + "step": 23795 + }, + { + "epoch": 49.48024948024948, + "grad_norm": 0.00017766660312190652, + "learning_rate": 0.10590521152235312, + "loss": 0.2768, + "num_input_tokens_seen": 18152072, + "step": 23800 + }, + { + "epoch": 49.48024948024948, + "eval_loss": 0.25095635652542114, + "eval_runtime": 13.3843, + "eval_samples_per_second": 63.956, + "eval_steps_per_second": 15.989, + "num_input_tokens_seen": 18152072, + "step": 23800 + }, + { + "epoch": 49.49064449064449, + "grad_norm": 0.0003702202229760587, + "learning_rate": 0.1058489127077369, + "loss": 0.2769, + "num_input_tokens_seen": 18155880, + "step": 23805 + }, + { + "epoch": 49.5010395010395, + "grad_norm": 0.00026751047698780894, + "learning_rate": 0.1057926207017732, + "loss": 0.2739, + "num_input_tokens_seen": 18159848, + "step": 23810 + }, + { + "epoch": 49.51143451143451, + "grad_norm": 0.00024904997553676367, + "learning_rate": 0.10573633551314285, + "loss": 0.2642, + "num_input_tokens_seen": 18163752, + "step": 23815 + }, + { + "epoch": 49.521829521829524, + "grad_norm": 0.00016037505702115595, + "learning_rate": 0.1056800571505259, + "loss": 0.2463, + "num_input_tokens_seen": 18167528, + "step": 23820 + }, + { + "epoch": 49.53222453222453, + "grad_norm": 0.00023132209025789052, + "learning_rate": 0.10562378562260105, + "loss": 0.2672, + "num_input_tokens_seen": 18171400, + "step": 23825 + }, + { + "epoch": 49.54261954261954, + "grad_norm": 0.0004048071859870106, + "learning_rate": 0.10556752093804615, + "loss": 0.275, + "num_input_tokens_seen": 18175304, + "step": 23830 + }, + { + "epoch": 49.553014553014556, + "grad_norm": 0.00011635488772299141, + "learning_rate": 0.10551126310553786, + "loss": 0.2309, + "num_input_tokens_seen": 18179176, + "step": 23835 + }, + { + "epoch": 49.563409563409564, + "grad_norm": 3.719219967024401e-05, + "learning_rate": 0.10545501213375187, + "loss": 0.2566, + "num_input_tokens_seen": 18182984, + "step": 23840 + }, + { + "epoch": 49.57380457380457, + "grad_norm": 0.0003588319232221693, + "learning_rate": 0.10539876803136287, + "loss": 0.2588, + "num_input_tokens_seen": 18186536, + "step": 23845 + }, + { + "epoch": 49.58419958419958, + "grad_norm": 0.00022439012536779046, + "learning_rate": 0.10534253080704428, + "loss": 0.2709, + "num_input_tokens_seen": 18190152, + "step": 23850 + }, + { + "epoch": 49.5945945945946, + "grad_norm": 0.0003745346039067954, + "learning_rate": 0.10528630046946862, + "loss": 0.2608, + "num_input_tokens_seen": 18193864, + "step": 23855 + }, + { + "epoch": 49.604989604989605, + "grad_norm": 0.0004120338417124003, + "learning_rate": 0.1052300770273074, + "loss": 0.2655, + "num_input_tokens_seen": 18197672, + "step": 23860 + }, + { + "epoch": 49.61538461538461, + "grad_norm": 0.00016493379371240735, + "learning_rate": 0.10517386048923086, + "loss": 0.2806, + "num_input_tokens_seen": 18201480, + "step": 23865 + }, + { + "epoch": 49.62577962577963, + "grad_norm": 0.00013868477253708988, + "learning_rate": 0.10511765086390841, + "loss": 0.2803, + "num_input_tokens_seen": 18205352, + "step": 23870 + }, + { + "epoch": 49.63617463617464, + "grad_norm": 0.00024250400019809604, + "learning_rate": 0.10506144816000816, + "loss": 0.2726, + "num_input_tokens_seen": 18208968, + "step": 23875 + }, + { + "epoch": 49.646569646569645, + "grad_norm": 0.00012117517326259986, + "learning_rate": 0.10500525238619736, + "loss": 0.2755, + "num_input_tokens_seen": 18212840, + "step": 23880 + }, + { + "epoch": 49.656964656964654, + "grad_norm": 0.0004627651069313288, + "learning_rate": 0.10494906355114209, + "loss": 0.2742, + "num_input_tokens_seen": 18216744, + "step": 23885 + }, + { + "epoch": 49.66735966735967, + "grad_norm": 0.0004094117321074009, + "learning_rate": 0.10489288166350737, + "loss": 0.2823, + "num_input_tokens_seen": 18220680, + "step": 23890 + }, + { + "epoch": 49.67775467775468, + "grad_norm": 0.0003043843898922205, + "learning_rate": 0.10483670673195711, + "loss": 0.267, + "num_input_tokens_seen": 18224360, + "step": 23895 + }, + { + "epoch": 49.688149688149686, + "grad_norm": 0.0003859902499243617, + "learning_rate": 0.10478053876515431, + "loss": 0.2857, + "num_input_tokens_seen": 18228072, + "step": 23900 + }, + { + "epoch": 49.6985446985447, + "grad_norm": 0.0004133705806452781, + "learning_rate": 0.10472437777176061, + "loss": 0.2688, + "num_input_tokens_seen": 18231784, + "step": 23905 + }, + { + "epoch": 49.70893970893971, + "grad_norm": 0.0002666472573764622, + "learning_rate": 0.1046682237604369, + "loss": 0.2527, + "num_input_tokens_seen": 18235464, + "step": 23910 + }, + { + "epoch": 49.71933471933472, + "grad_norm": 0.00042969704372808337, + "learning_rate": 0.1046120767398427, + "loss": 0.2693, + "num_input_tokens_seen": 18239336, + "step": 23915 + }, + { + "epoch": 49.729729729729726, + "grad_norm": 0.0004417806339915842, + "learning_rate": 0.10455593671863667, + "loss": 0.2628, + "num_input_tokens_seen": 18243016, + "step": 23920 + }, + { + "epoch": 49.74012474012474, + "grad_norm": 0.0004789374943356961, + "learning_rate": 0.1044998037054763, + "loss": 0.2797, + "num_input_tokens_seen": 18246792, + "step": 23925 + }, + { + "epoch": 49.75051975051975, + "grad_norm": 5.336150206858292e-05, + "learning_rate": 0.10444367770901794, + "loss": 0.2755, + "num_input_tokens_seen": 18250600, + "step": 23930 + }, + { + "epoch": 49.76091476091476, + "grad_norm": 7.816852303221822e-05, + "learning_rate": 0.10438755873791698, + "loss": 0.2893, + "num_input_tokens_seen": 18254440, + "step": 23935 + }, + { + "epoch": 49.771309771309774, + "grad_norm": 0.00030390999745577574, + "learning_rate": 0.10433144680082775, + "loss": 0.2582, + "num_input_tokens_seen": 18258184, + "step": 23940 + }, + { + "epoch": 49.78170478170478, + "grad_norm": 7.809425733285025e-05, + "learning_rate": 0.10427534190640322, + "loss": 0.2695, + "num_input_tokens_seen": 18262024, + "step": 23945 + }, + { + "epoch": 49.79209979209979, + "grad_norm": 0.00015017240366432816, + "learning_rate": 0.10421924406329568, + "loss": 0.2543, + "num_input_tokens_seen": 18265960, + "step": 23950 + }, + { + "epoch": 49.802494802494806, + "grad_norm": 0.0005150067736394703, + "learning_rate": 0.10416315328015598, + "loss": 0.2356, + "num_input_tokens_seen": 18269800, + "step": 23955 + }, + { + "epoch": 49.812889812889814, + "grad_norm": 0.0002945914748124778, + "learning_rate": 0.10410706956563402, + "loss": 0.2788, + "num_input_tokens_seen": 18273576, + "step": 23960 + }, + { + "epoch": 49.82328482328482, + "grad_norm": 0.0002123496524291113, + "learning_rate": 0.10405099292837874, + "loss": 0.275, + "num_input_tokens_seen": 18277448, + "step": 23965 + }, + { + "epoch": 49.83367983367983, + "grad_norm": 0.00026722936308942735, + "learning_rate": 0.10399492337703771, + "loss": 0.2791, + "num_input_tokens_seen": 18281160, + "step": 23970 + }, + { + "epoch": 49.84407484407485, + "grad_norm": 0.00010798217408591881, + "learning_rate": 0.10393886092025764, + "loss": 0.2626, + "num_input_tokens_seen": 18285064, + "step": 23975 + }, + { + "epoch": 49.854469854469855, + "grad_norm": 0.0001463006337871775, + "learning_rate": 0.10388280556668412, + "loss": 0.2686, + "num_input_tokens_seen": 18288904, + "step": 23980 + }, + { + "epoch": 49.86486486486486, + "grad_norm": 0.00013373060210142285, + "learning_rate": 0.10382675732496145, + "loss": 0.2536, + "num_input_tokens_seen": 18292712, + "step": 23985 + }, + { + "epoch": 49.87525987525988, + "grad_norm": 0.00023828244593460113, + "learning_rate": 0.10377071620373311, + "loss": 0.2169, + "num_input_tokens_seen": 18296520, + "step": 23990 + }, + { + "epoch": 49.88565488565489, + "grad_norm": 0.0002944665611721575, + "learning_rate": 0.10371468221164128, + "loss": 0.2938, + "num_input_tokens_seen": 18300328, + "step": 23995 + }, + { + "epoch": 49.896049896049895, + "grad_norm": 0.00022032791457604617, + "learning_rate": 0.10365865535732706, + "loss": 0.2944, + "num_input_tokens_seen": 18304072, + "step": 24000 + }, + { + "epoch": 49.896049896049895, + "eval_loss": 0.24930372834205627, + "eval_runtime": 13.4028, + "eval_samples_per_second": 63.867, + "eval_steps_per_second": 15.967, + "num_input_tokens_seen": 18304072, + "step": 24000 + }, + { + "epoch": 49.906444906444904, + "grad_norm": 0.00022074466687627137, + "learning_rate": 0.10360263564943062, + "loss": 0.2732, + "num_input_tokens_seen": 18307816, + "step": 24005 + }, + { + "epoch": 49.91683991683992, + "grad_norm": 8.974806405603886e-05, + "learning_rate": 0.10354662309659075, + "loss": 0.2649, + "num_input_tokens_seen": 18311624, + "step": 24010 + }, + { + "epoch": 49.92723492723493, + "grad_norm": 0.0003212092851754278, + "learning_rate": 0.10349061770744537, + "loss": 0.2644, + "num_input_tokens_seen": 18315496, + "step": 24015 + }, + { + "epoch": 49.937629937629936, + "grad_norm": 0.0001214043004438281, + "learning_rate": 0.10343461949063128, + "loss": 0.2611, + "num_input_tokens_seen": 18319208, + "step": 24020 + }, + { + "epoch": 49.94802494802495, + "grad_norm": 0.0001060393187799491, + "learning_rate": 0.103378628454784, + "loss": 0.2634, + "num_input_tokens_seen": 18322952, + "step": 24025 + }, + { + "epoch": 49.95841995841996, + "grad_norm": 0.00013607065193355083, + "learning_rate": 0.10332264460853811, + "loss": 0.2866, + "num_input_tokens_seen": 18326600, + "step": 24030 + }, + { + "epoch": 49.96881496881497, + "grad_norm": 8.524490112904459e-05, + "learning_rate": 0.10326666796052701, + "loss": 0.2383, + "num_input_tokens_seen": 18330440, + "step": 24035 + }, + { + "epoch": 49.979209979209976, + "grad_norm": 0.0001402928028255701, + "learning_rate": 0.10321069851938296, + "loss": 0.2609, + "num_input_tokens_seen": 18334184, + "step": 24040 + }, + { + "epoch": 49.98960498960499, + "grad_norm": 0.00022541174257639796, + "learning_rate": 0.10315473629373724, + "loss": 0.2506, + "num_input_tokens_seen": 18338120, + "step": 24045 + }, + { + "epoch": 50.0, + "grad_norm": 0.00013395698624663055, + "learning_rate": 0.10309878129221982, + "loss": 0.2529, + "num_input_tokens_seen": 18341840, + "step": 24050 + }, + { + "epoch": 50.01039501039501, + "grad_norm": 4.0829749195836484e-05, + "learning_rate": 0.10304283352345973, + "loss": 0.2513, + "num_input_tokens_seen": 18345680, + "step": 24055 + }, + { + "epoch": 50.020790020790024, + "grad_norm": 0.0005817682249471545, + "learning_rate": 0.10298689299608486, + "loss": 0.2938, + "num_input_tokens_seen": 18349520, + "step": 24060 + }, + { + "epoch": 50.03118503118503, + "grad_norm": 0.0003067734360229224, + "learning_rate": 0.10293095971872188, + "loss": 0.2642, + "num_input_tokens_seen": 18353264, + "step": 24065 + }, + { + "epoch": 50.04158004158004, + "grad_norm": 0.0002419666707282886, + "learning_rate": 0.10287503369999645, + "loss": 0.2704, + "num_input_tokens_seen": 18357040, + "step": 24070 + }, + { + "epoch": 50.05197505197505, + "grad_norm": 0.0002429165760986507, + "learning_rate": 0.10281911494853295, + "loss": 0.2767, + "num_input_tokens_seen": 18360976, + "step": 24075 + }, + { + "epoch": 50.062370062370064, + "grad_norm": 0.00031340416171588004, + "learning_rate": 0.10276320347295485, + "loss": 0.268, + "num_input_tokens_seen": 18364816, + "step": 24080 + }, + { + "epoch": 50.07276507276507, + "grad_norm": 0.0001533517352072522, + "learning_rate": 0.10270729928188446, + "loss": 0.2824, + "num_input_tokens_seen": 18368624, + "step": 24085 + }, + { + "epoch": 50.08316008316008, + "grad_norm": 0.0001746221532812342, + "learning_rate": 0.10265140238394276, + "loss": 0.2634, + "num_input_tokens_seen": 18372336, + "step": 24090 + }, + { + "epoch": 50.093555093555096, + "grad_norm": 0.00046772032510489225, + "learning_rate": 0.10259551278774988, + "loss": 0.283, + "num_input_tokens_seen": 18376016, + "step": 24095 + }, + { + "epoch": 50.103950103950105, + "grad_norm": 0.00032474062754772604, + "learning_rate": 0.10253963050192462, + "loss": 0.2502, + "num_input_tokens_seen": 18379952, + "step": 24100 + }, + { + "epoch": 50.11434511434511, + "grad_norm": 0.0002057932724710554, + "learning_rate": 0.10248375553508478, + "loss": 0.2604, + "num_input_tokens_seen": 18383664, + "step": 24105 + }, + { + "epoch": 50.12474012474012, + "grad_norm": 0.00039557606214657426, + "learning_rate": 0.102427887895847, + "loss": 0.2682, + "num_input_tokens_seen": 18387536, + "step": 24110 + }, + { + "epoch": 50.13513513513514, + "grad_norm": 0.0002889862225856632, + "learning_rate": 0.10237202759282668, + "loss": 0.2564, + "num_input_tokens_seen": 18391472, + "step": 24115 + }, + { + "epoch": 50.145530145530145, + "grad_norm": 8.735974552109838e-05, + "learning_rate": 0.10231617463463821, + "loss": 0.2796, + "num_input_tokens_seen": 18395152, + "step": 24120 + }, + { + "epoch": 50.15592515592515, + "grad_norm": 0.0003275749913882464, + "learning_rate": 0.10226032902989492, + "loss": 0.2474, + "num_input_tokens_seen": 18398896, + "step": 24125 + }, + { + "epoch": 50.16632016632017, + "grad_norm": 0.0003949719830416143, + "learning_rate": 0.10220449078720877, + "loss": 0.2806, + "num_input_tokens_seen": 18402640, + "step": 24130 + }, + { + "epoch": 50.17671517671518, + "grad_norm": 0.0005629869410768151, + "learning_rate": 0.1021486599151908, + "loss": 0.2764, + "num_input_tokens_seen": 18406448, + "step": 24135 + }, + { + "epoch": 50.187110187110186, + "grad_norm": 5.7437307987129316e-05, + "learning_rate": 0.10209283642245084, + "loss": 0.272, + "num_input_tokens_seen": 18410256, + "step": 24140 + }, + { + "epoch": 50.197505197505194, + "grad_norm": 0.00025217654183506966, + "learning_rate": 0.10203702031759748, + "loss": 0.2737, + "num_input_tokens_seen": 18414096, + "step": 24145 + }, + { + "epoch": 50.20790020790021, + "grad_norm": 0.00018602272029966116, + "learning_rate": 0.1019812116092384, + "loss": 0.2707, + "num_input_tokens_seen": 18417776, + "step": 24150 + }, + { + "epoch": 50.21829521829522, + "grad_norm": 0.0003643235540948808, + "learning_rate": 0.10192541030597986, + "loss": 0.2472, + "num_input_tokens_seen": 18421456, + "step": 24155 + }, + { + "epoch": 50.228690228690226, + "grad_norm": 9.985128417611122e-05, + "learning_rate": 0.1018696164164272, + "loss": 0.2588, + "num_input_tokens_seen": 18425296, + "step": 24160 + }, + { + "epoch": 50.23908523908524, + "grad_norm": 0.0001157584338216111, + "learning_rate": 0.10181382994918459, + "loss": 0.2687, + "num_input_tokens_seen": 18429200, + "step": 24165 + }, + { + "epoch": 50.24948024948025, + "grad_norm": 0.0002264860668219626, + "learning_rate": 0.10175805091285492, + "loss": 0.2634, + "num_input_tokens_seen": 18433136, + "step": 24170 + }, + { + "epoch": 50.25987525987526, + "grad_norm": 0.0002869514573831111, + "learning_rate": 0.10170227931603999, + "loss": 0.2133, + "num_input_tokens_seen": 18436944, + "step": 24175 + }, + { + "epoch": 50.270270270270274, + "grad_norm": 0.00016675608640071005, + "learning_rate": 0.10164651516734062, + "loss": 0.2774, + "num_input_tokens_seen": 18440880, + "step": 24180 + }, + { + "epoch": 50.28066528066528, + "grad_norm": 0.00010287578334100544, + "learning_rate": 0.1015907584753562, + "loss": 0.261, + "num_input_tokens_seen": 18444464, + "step": 24185 + }, + { + "epoch": 50.29106029106029, + "grad_norm": 0.0006566105876117945, + "learning_rate": 0.10153500924868523, + "loss": 0.2259, + "num_input_tokens_seen": 18448272, + "step": 24190 + }, + { + "epoch": 50.3014553014553, + "grad_norm": 0.00014778364857193083, + "learning_rate": 0.10147926749592483, + "loss": 0.2436, + "num_input_tokens_seen": 18451984, + "step": 24195 + }, + { + "epoch": 50.311850311850314, + "grad_norm": 0.0002861306711565703, + "learning_rate": 0.10142353322567112, + "loss": 0.2994, + "num_input_tokens_seen": 18455696, + "step": 24200 + }, + { + "epoch": 50.311850311850314, + "eval_loss": 0.24659645557403564, + "eval_runtime": 13.4121, + "eval_samples_per_second": 63.823, + "eval_steps_per_second": 15.956, + "num_input_tokens_seen": 18455696, + "step": 24200 + }, + { + "epoch": 50.32224532224532, + "grad_norm": 6.527492223540321e-05, + "learning_rate": 0.1013678064465191, + "loss": 0.2523, + "num_input_tokens_seen": 18459568, + "step": 24205 + }, + { + "epoch": 50.33264033264033, + "grad_norm": 0.0002931687922682613, + "learning_rate": 0.10131208716706244, + "loss": 0.2834, + "num_input_tokens_seen": 18463504, + "step": 24210 + }, + { + "epoch": 50.343035343035346, + "grad_norm": 0.0001677976833889261, + "learning_rate": 0.10125637539589379, + "loss": 0.2649, + "num_input_tokens_seen": 18467408, + "step": 24215 + }, + { + "epoch": 50.353430353430355, + "grad_norm": 0.00028112702420912683, + "learning_rate": 0.10120067114160464, + "loss": 0.2583, + "num_input_tokens_seen": 18471120, + "step": 24220 + }, + { + "epoch": 50.36382536382536, + "grad_norm": 0.00014877118519507349, + "learning_rate": 0.10114497441278517, + "loss": 0.2674, + "num_input_tokens_seen": 18475024, + "step": 24225 + }, + { + "epoch": 50.37422037422037, + "grad_norm": 0.00014939019456505775, + "learning_rate": 0.10108928521802468, + "loss": 0.2738, + "num_input_tokens_seen": 18478768, + "step": 24230 + }, + { + "epoch": 50.38461538461539, + "grad_norm": 9.070548549061641e-05, + "learning_rate": 0.101033603565911, + "loss": 0.2655, + "num_input_tokens_seen": 18482672, + "step": 24235 + }, + { + "epoch": 50.395010395010395, + "grad_norm": 0.0005833710310980678, + "learning_rate": 0.10097792946503102, + "loss": 0.2737, + "num_input_tokens_seen": 18486480, + "step": 24240 + }, + { + "epoch": 50.4054054054054, + "grad_norm": 0.00020180588762741536, + "learning_rate": 0.10092226292397039, + "loss": 0.2677, + "num_input_tokens_seen": 18490384, + "step": 24245 + }, + { + "epoch": 50.41580041580042, + "grad_norm": 0.00026685497141443193, + "learning_rate": 0.10086660395131354, + "loss": 0.257, + "num_input_tokens_seen": 18494224, + "step": 24250 + }, + { + "epoch": 50.42619542619543, + "grad_norm": 0.00023248510842677206, + "learning_rate": 0.10081095255564385, + "loss": 0.2524, + "num_input_tokens_seen": 18498000, + "step": 24255 + }, + { + "epoch": 50.436590436590436, + "grad_norm": 0.00020622876763809472, + "learning_rate": 0.10075530874554335, + "loss": 0.2937, + "num_input_tokens_seen": 18501936, + "step": 24260 + }, + { + "epoch": 50.446985446985444, + "grad_norm": 0.0003202626539859921, + "learning_rate": 0.10069967252959311, + "loss": 0.2566, + "num_input_tokens_seen": 18505648, + "step": 24265 + }, + { + "epoch": 50.45738045738046, + "grad_norm": 0.00010146533895749599, + "learning_rate": 0.10064404391637297, + "loss": 0.2592, + "num_input_tokens_seen": 18509360, + "step": 24270 + }, + { + "epoch": 50.46777546777547, + "grad_norm": 0.00012425774184521288, + "learning_rate": 0.10058842291446145, + "loss": 0.2716, + "num_input_tokens_seen": 18513136, + "step": 24275 + }, + { + "epoch": 50.478170478170476, + "grad_norm": 9.50094181462191e-05, + "learning_rate": 0.10053280953243608, + "loss": 0.2339, + "num_input_tokens_seen": 18516944, + "step": 24280 + }, + { + "epoch": 50.48856548856549, + "grad_norm": 0.00015902260201983154, + "learning_rate": 0.10047720377887315, + "loss": 0.2449, + "num_input_tokens_seen": 18520816, + "step": 24285 + }, + { + "epoch": 50.4989604989605, + "grad_norm": 7.598899537697434e-05, + "learning_rate": 0.10042160566234767, + "loss": 0.2353, + "num_input_tokens_seen": 18524528, + "step": 24290 + }, + { + "epoch": 50.50935550935551, + "grad_norm": 0.0002526042517274618, + "learning_rate": 0.10036601519143372, + "loss": 0.2917, + "num_input_tokens_seen": 18528432, + "step": 24295 + }, + { + "epoch": 50.51975051975052, + "grad_norm": 8.480740507366136e-05, + "learning_rate": 0.1003104323747039, + "loss": 0.2817, + "num_input_tokens_seen": 18532080, + "step": 24300 + }, + { + "epoch": 50.53014553014553, + "grad_norm": 0.00013210528413765132, + "learning_rate": 0.10025485722072984, + "loss": 0.2593, + "num_input_tokens_seen": 18535984, + "step": 24305 + }, + { + "epoch": 50.54054054054054, + "grad_norm": 4.810529571841471e-05, + "learning_rate": 0.10019928973808201, + "loss": 0.2702, + "num_input_tokens_seen": 18539856, + "step": 24310 + }, + { + "epoch": 50.55093555093555, + "grad_norm": 0.00028146570548415184, + "learning_rate": 0.10014372993532945, + "loss": 0.2543, + "num_input_tokens_seen": 18543696, + "step": 24315 + }, + { + "epoch": 50.561330561330564, + "grad_norm": 0.00033793141483329237, + "learning_rate": 0.1000881778210403, + "loss": 0.2446, + "num_input_tokens_seen": 18547664, + "step": 24320 + }, + { + "epoch": 50.57172557172557, + "grad_norm": 0.00026973572676070035, + "learning_rate": 0.10003263340378142, + "loss": 0.2638, + "num_input_tokens_seen": 18551568, + "step": 24325 + }, + { + "epoch": 50.58212058212058, + "grad_norm": 0.00018194993026554585, + "learning_rate": 0.09997709669211834, + "loss": 0.2591, + "num_input_tokens_seen": 18555536, + "step": 24330 + }, + { + "epoch": 50.59251559251559, + "grad_norm": 0.00017239435692317784, + "learning_rate": 0.0999215676946156, + "loss": 0.269, + "num_input_tokens_seen": 18559472, + "step": 24335 + }, + { + "epoch": 50.602910602910605, + "grad_norm": 0.00017497414955869317, + "learning_rate": 0.0998660464198364, + "loss": 0.3024, + "num_input_tokens_seen": 18563248, + "step": 24340 + }, + { + "epoch": 50.61330561330561, + "grad_norm": 0.0004112626484129578, + "learning_rate": 0.09981053287634288, + "loss": 0.2512, + "num_input_tokens_seen": 18567088, + "step": 24345 + }, + { + "epoch": 50.62370062370062, + "grad_norm": 9.315647912444547e-05, + "learning_rate": 0.09975502707269596, + "loss": 0.2677, + "num_input_tokens_seen": 18571088, + "step": 24350 + }, + { + "epoch": 50.63409563409564, + "grad_norm": 9.203373338095844e-05, + "learning_rate": 0.09969952901745524, + "loss": 0.2588, + "num_input_tokens_seen": 18574704, + "step": 24355 + }, + { + "epoch": 50.644490644490645, + "grad_norm": 6.929399387445301e-05, + "learning_rate": 0.09964403871917925, + "loss": 0.2486, + "num_input_tokens_seen": 18578480, + "step": 24360 + }, + { + "epoch": 50.65488565488565, + "grad_norm": 0.0001429767144145444, + "learning_rate": 0.09958855618642536, + "loss": 0.2645, + "num_input_tokens_seen": 18582288, + "step": 24365 + }, + { + "epoch": 50.66528066528066, + "grad_norm": 0.00015283816901501268, + "learning_rate": 0.09953308142774955, + "loss": 0.2614, + "num_input_tokens_seen": 18586000, + "step": 24370 + }, + { + "epoch": 50.67567567567568, + "grad_norm": 0.00022856822761241347, + "learning_rate": 0.09947761445170686, + "loss": 0.258, + "num_input_tokens_seen": 18589904, + "step": 24375 + }, + { + "epoch": 50.686070686070686, + "grad_norm": 0.00017272579134441912, + "learning_rate": 0.09942215526685086, + "loss": 0.2782, + "num_input_tokens_seen": 18593712, + "step": 24380 + }, + { + "epoch": 50.696465696465694, + "grad_norm": 0.0003653337771538645, + "learning_rate": 0.09936670388173414, + "loss": 0.2632, + "num_input_tokens_seen": 18597456, + "step": 24385 + }, + { + "epoch": 50.70686070686071, + "grad_norm": 0.0001519051002105698, + "learning_rate": 0.09931126030490799, + "loss": 0.2514, + "num_input_tokens_seen": 18601360, + "step": 24390 + }, + { + "epoch": 50.71725571725572, + "grad_norm": 0.00014558061957359314, + "learning_rate": 0.0992558245449225, + "loss": 0.2444, + "num_input_tokens_seen": 18605136, + "step": 24395 + }, + { + "epoch": 50.727650727650726, + "grad_norm": 0.00031312459032051265, + "learning_rate": 0.09920039661032651, + "loss": 0.2625, + "num_input_tokens_seen": 18608976, + "step": 24400 + }, + { + "epoch": 50.727650727650726, + "eval_loss": 0.24898014962673187, + "eval_runtime": 13.3853, + "eval_samples_per_second": 63.951, + "eval_steps_per_second": 15.988, + "num_input_tokens_seen": 18608976, + "step": 24400 + }, + { + "epoch": 50.73804573804574, + "grad_norm": 0.00031856808345764875, + "learning_rate": 0.09914497650966782, + "loss": 0.2275, + "num_input_tokens_seen": 18612848, + "step": 24405 + }, + { + "epoch": 50.74844074844075, + "grad_norm": 0.0007534954929724336, + "learning_rate": 0.09908956425149276, + "loss": 0.2838, + "num_input_tokens_seen": 18616688, + "step": 24410 + }, + { + "epoch": 50.75883575883576, + "grad_norm": 0.00045412409235723317, + "learning_rate": 0.09903415984434677, + "loss": 0.2603, + "num_input_tokens_seen": 18620496, + "step": 24415 + }, + { + "epoch": 50.76923076923077, + "grad_norm": 0.00018290500156581402, + "learning_rate": 0.09897876329677373, + "loss": 0.2755, + "num_input_tokens_seen": 18624400, + "step": 24420 + }, + { + "epoch": 50.77962577962578, + "grad_norm": 0.0003129853284917772, + "learning_rate": 0.09892337461731658, + "loss": 0.2644, + "num_input_tokens_seen": 18628336, + "step": 24425 + }, + { + "epoch": 50.79002079002079, + "grad_norm": 0.0002791692386381328, + "learning_rate": 0.09886799381451693, + "loss": 0.2916, + "num_input_tokens_seen": 18632208, + "step": 24430 + }, + { + "epoch": 50.8004158004158, + "grad_norm": 0.0002489391481503844, + "learning_rate": 0.09881262089691521, + "loss": 0.2733, + "num_input_tokens_seen": 18635984, + "step": 24435 + }, + { + "epoch": 50.810810810810814, + "grad_norm": 0.0005622346652671695, + "learning_rate": 0.09875725587305059, + "loss": 0.2782, + "num_input_tokens_seen": 18639952, + "step": 24440 + }, + { + "epoch": 50.82120582120582, + "grad_norm": 4.364913183962926e-05, + "learning_rate": 0.09870189875146111, + "loss": 0.2752, + "num_input_tokens_seen": 18643632, + "step": 24445 + }, + { + "epoch": 50.83160083160083, + "grad_norm": 0.00022214869386516511, + "learning_rate": 0.09864654954068346, + "loss": 0.2765, + "num_input_tokens_seen": 18647408, + "step": 24450 + }, + { + "epoch": 50.84199584199584, + "grad_norm": 0.00011121843272121623, + "learning_rate": 0.09859120824925326, + "loss": 0.2581, + "num_input_tokens_seen": 18651216, + "step": 24455 + }, + { + "epoch": 50.852390852390855, + "grad_norm": 0.00023151130881160498, + "learning_rate": 0.09853587488570474, + "loss": 0.2602, + "num_input_tokens_seen": 18654960, + "step": 24460 + }, + { + "epoch": 50.86278586278586, + "grad_norm": 0.0005612490349449217, + "learning_rate": 0.09848054945857107, + "loss": 0.2782, + "num_input_tokens_seen": 18658768, + "step": 24465 + }, + { + "epoch": 50.87318087318087, + "grad_norm": 0.0002154849935323, + "learning_rate": 0.09842523197638416, + "loss": 0.2645, + "num_input_tokens_seen": 18662480, + "step": 24470 + }, + { + "epoch": 50.88357588357589, + "grad_norm": 0.0001781236642273143, + "learning_rate": 0.09836992244767452, + "loss": 0.2617, + "num_input_tokens_seen": 18666256, + "step": 24475 + }, + { + "epoch": 50.893970893970895, + "grad_norm": 5.580757715506479e-05, + "learning_rate": 0.09831462088097168, + "loss": 0.2845, + "num_input_tokens_seen": 18669968, + "step": 24480 + }, + { + "epoch": 50.9043659043659, + "grad_norm": 0.0002857065701391548, + "learning_rate": 0.09825932728480385, + "loss": 0.2729, + "num_input_tokens_seen": 18673712, + "step": 24485 + }, + { + "epoch": 50.91476091476091, + "grad_norm": 9.086565842153504e-05, + "learning_rate": 0.09820404166769794, + "loss": 0.2726, + "num_input_tokens_seen": 18677360, + "step": 24490 + }, + { + "epoch": 50.92515592515593, + "grad_norm": 7.799208106007427e-05, + "learning_rate": 0.09814876403817978, + "loss": 0.2623, + "num_input_tokens_seen": 18681264, + "step": 24495 + }, + { + "epoch": 50.935550935550935, + "grad_norm": 0.00014397941413335502, + "learning_rate": 0.09809349440477376, + "loss": 0.2507, + "num_input_tokens_seen": 18685072, + "step": 24500 + }, + { + "epoch": 50.945945945945944, + "grad_norm": 0.00013089526328258216, + "learning_rate": 0.09803823277600317, + "loss": 0.2749, + "num_input_tokens_seen": 18688848, + "step": 24505 + }, + { + "epoch": 50.95634095634096, + "grad_norm": 0.0005686940276063979, + "learning_rate": 0.09798297916039014, + "loss": 0.2742, + "num_input_tokens_seen": 18692624, + "step": 24510 + }, + { + "epoch": 50.96673596673597, + "grad_norm": 0.00027545986813493073, + "learning_rate": 0.09792773356645534, + "loss": 0.2579, + "num_input_tokens_seen": 18696304, + "step": 24515 + }, + { + "epoch": 50.977130977130976, + "grad_norm": 4.643796637537889e-05, + "learning_rate": 0.09787249600271843, + "loss": 0.2594, + "num_input_tokens_seen": 18700080, + "step": 24520 + }, + { + "epoch": 50.987525987525984, + "grad_norm": 0.00015421421267092228, + "learning_rate": 0.09781726647769776, + "loss": 0.2594, + "num_input_tokens_seen": 18703920, + "step": 24525 + }, + { + "epoch": 50.997920997921, + "grad_norm": 9.456792759010568e-05, + "learning_rate": 0.0977620449999103, + "loss": 0.2673, + "num_input_tokens_seen": 18707696, + "step": 24530 + }, + { + "epoch": 51.00831600831601, + "grad_norm": 0.00024646814563311636, + "learning_rate": 0.09770683157787204, + "loss": 0.2596, + "num_input_tokens_seen": 18711488, + "step": 24535 + }, + { + "epoch": 51.018711018711016, + "grad_norm": 0.00024192185082938522, + "learning_rate": 0.09765162622009745, + "loss": 0.2469, + "num_input_tokens_seen": 18715296, + "step": 24540 + }, + { + "epoch": 51.02910602910603, + "grad_norm": 7.681120769120753e-05, + "learning_rate": 0.09759642893509995, + "loss": 0.2298, + "num_input_tokens_seen": 18719008, + "step": 24545 + }, + { + "epoch": 51.03950103950104, + "grad_norm": 0.0002607563219498843, + "learning_rate": 0.09754123973139169, + "loss": 0.2435, + "num_input_tokens_seen": 18722848, + "step": 24550 + }, + { + "epoch": 51.04989604989605, + "grad_norm": 8.269328827736899e-05, + "learning_rate": 0.09748605861748345, + "loss": 0.305, + "num_input_tokens_seen": 18726752, + "step": 24555 + }, + { + "epoch": 51.06029106029106, + "grad_norm": 0.00023772777058184147, + "learning_rate": 0.0974308856018849, + "loss": 0.2769, + "num_input_tokens_seen": 18730560, + "step": 24560 + }, + { + "epoch": 51.07068607068607, + "grad_norm": 0.00020780455088242888, + "learning_rate": 0.09737572069310449, + "loss": 0.2697, + "num_input_tokens_seen": 18734272, + "step": 24565 + }, + { + "epoch": 51.08108108108108, + "grad_norm": 0.0003766508889384568, + "learning_rate": 0.09732056389964922, + "loss": 0.2632, + "num_input_tokens_seen": 18738080, + "step": 24570 + }, + { + "epoch": 51.09147609147609, + "grad_norm": 0.00036014526267535985, + "learning_rate": 0.097265415230025, + "loss": 0.2798, + "num_input_tokens_seen": 18741888, + "step": 24575 + }, + { + "epoch": 51.101871101871104, + "grad_norm": 0.0001421941997250542, + "learning_rate": 0.09721027469273648, + "loss": 0.2686, + "num_input_tokens_seen": 18745664, + "step": 24580 + }, + { + "epoch": 51.11226611226611, + "grad_norm": 0.00047153487685136497, + "learning_rate": 0.09715514229628695, + "loss": 0.2858, + "num_input_tokens_seen": 18749472, + "step": 24585 + }, + { + "epoch": 51.12266112266112, + "grad_norm": 0.00030557592981494963, + "learning_rate": 0.09710001804917864, + "loss": 0.2868, + "num_input_tokens_seen": 18753312, + "step": 24590 + }, + { + "epoch": 51.13305613305613, + "grad_norm": 0.00013361706805881113, + "learning_rate": 0.09704490195991226, + "loss": 0.2687, + "num_input_tokens_seen": 18757152, + "step": 24595 + }, + { + "epoch": 51.143451143451145, + "grad_norm": 0.00010781670425785705, + "learning_rate": 0.09698979403698753, + "loss": 0.2745, + "num_input_tokens_seen": 18760928, + "step": 24600 + }, + { + "epoch": 51.143451143451145, + "eval_loss": 0.26018255949020386, + "eval_runtime": 13.3973, + "eval_samples_per_second": 63.893, + "eval_steps_per_second": 15.973, + "num_input_tokens_seen": 18760928, + "step": 24600 + }, + { + "epoch": 51.15384615384615, + "grad_norm": 0.0002476339286658913, + "learning_rate": 0.0969346942889027, + "loss": 0.2898, + "num_input_tokens_seen": 18764864, + "step": 24605 + }, + { + "epoch": 51.16424116424116, + "grad_norm": 0.000101707068097312, + "learning_rate": 0.09687960272415487, + "loss": 0.2651, + "num_input_tokens_seen": 18768672, + "step": 24610 + }, + { + "epoch": 51.17463617463618, + "grad_norm": 0.0004897169419564307, + "learning_rate": 0.0968245193512399, + "loss": 0.2623, + "num_input_tokens_seen": 18772480, + "step": 24615 + }, + { + "epoch": 51.185031185031185, + "grad_norm": 0.0001092800303013064, + "learning_rate": 0.09676944417865221, + "loss": 0.2812, + "num_input_tokens_seen": 18776448, + "step": 24620 + }, + { + "epoch": 51.195426195426194, + "grad_norm": 0.0002611938398331404, + "learning_rate": 0.09671437721488517, + "loss": 0.273, + "num_input_tokens_seen": 18780128, + "step": 24625 + }, + { + "epoch": 51.20582120582121, + "grad_norm": 0.00036381211248226464, + "learning_rate": 0.09665931846843086, + "loss": 0.261, + "num_input_tokens_seen": 18783840, + "step": 24630 + }, + { + "epoch": 51.21621621621622, + "grad_norm": 0.0002876933431252837, + "learning_rate": 0.0966042679477799, + "loss": 0.259, + "num_input_tokens_seen": 18787680, + "step": 24635 + }, + { + "epoch": 51.226611226611226, + "grad_norm": 6.897662387927994e-05, + "learning_rate": 0.09654922566142186, + "loss": 0.2686, + "num_input_tokens_seen": 18791552, + "step": 24640 + }, + { + "epoch": 51.237006237006234, + "grad_norm": 0.00016861112089827657, + "learning_rate": 0.09649419161784498, + "loss": 0.2784, + "num_input_tokens_seen": 18795392, + "step": 24645 + }, + { + "epoch": 51.24740124740125, + "grad_norm": 0.00033378094667568803, + "learning_rate": 0.09643916582553606, + "loss": 0.2494, + "num_input_tokens_seen": 18799424, + "step": 24650 + }, + { + "epoch": 51.25779625779626, + "grad_norm": 0.0004326729103922844, + "learning_rate": 0.09638414829298093, + "loss": 0.2654, + "num_input_tokens_seen": 18803200, + "step": 24655 + }, + { + "epoch": 51.268191268191266, + "grad_norm": 0.00012549009989015758, + "learning_rate": 0.09632913902866386, + "loss": 0.279, + "num_input_tokens_seen": 18807040, + "step": 24660 + }, + { + "epoch": 51.27858627858628, + "grad_norm": 0.00017018205835483968, + "learning_rate": 0.096274138041068, + "loss": 0.2668, + "num_input_tokens_seen": 18810944, + "step": 24665 + }, + { + "epoch": 51.28898128898129, + "grad_norm": 0.00022028469538781792, + "learning_rate": 0.09621914533867527, + "loss": 0.2204, + "num_input_tokens_seen": 18814848, + "step": 24670 + }, + { + "epoch": 51.2993762993763, + "grad_norm": 0.0002884832792915404, + "learning_rate": 0.09616416092996616, + "loss": 0.2897, + "num_input_tokens_seen": 18818688, + "step": 24675 + }, + { + "epoch": 51.30977130977131, + "grad_norm": 0.0003661253722384572, + "learning_rate": 0.09610918482342, + "loss": 0.2712, + "num_input_tokens_seen": 18822560, + "step": 24680 + }, + { + "epoch": 51.32016632016632, + "grad_norm": 9.568790119374171e-05, + "learning_rate": 0.09605421702751478, + "loss": 0.2677, + "num_input_tokens_seen": 18826272, + "step": 24685 + }, + { + "epoch": 51.33056133056133, + "grad_norm": 0.00018907897174358368, + "learning_rate": 0.09599925755072718, + "loss": 0.264, + "num_input_tokens_seen": 18830176, + "step": 24690 + }, + { + "epoch": 51.34095634095634, + "grad_norm": 0.0005880626849830151, + "learning_rate": 0.09594430640153273, + "loss": 0.2473, + "num_input_tokens_seen": 18833984, + "step": 24695 + }, + { + "epoch": 51.351351351351354, + "grad_norm": 0.0003065942437388003, + "learning_rate": 0.09588936358840547, + "loss": 0.2439, + "num_input_tokens_seen": 18837664, + "step": 24700 + }, + { + "epoch": 51.36174636174636, + "grad_norm": 0.0001648537872824818, + "learning_rate": 0.09583442911981836, + "loss": 0.2807, + "num_input_tokens_seen": 18841376, + "step": 24705 + }, + { + "epoch": 51.37214137214137, + "grad_norm": 3.6155055568087846e-05, + "learning_rate": 0.09577950300424302, + "loss": 0.2468, + "num_input_tokens_seen": 18845088, + "step": 24710 + }, + { + "epoch": 51.38253638253638, + "grad_norm": 0.0001401686604367569, + "learning_rate": 0.09572458525014967, + "loss": 0.2475, + "num_input_tokens_seen": 18848864, + "step": 24715 + }, + { + "epoch": 51.392931392931395, + "grad_norm": 8.26112664071843e-05, + "learning_rate": 0.0956696758660073, + "loss": 0.2979, + "num_input_tokens_seen": 18852544, + "step": 24720 + }, + { + "epoch": 51.4033264033264, + "grad_norm": 5.821938248118386e-05, + "learning_rate": 0.09561477486028373, + "loss": 0.2471, + "num_input_tokens_seen": 18856512, + "step": 24725 + }, + { + "epoch": 51.41372141372141, + "grad_norm": 0.00045093640801496804, + "learning_rate": 0.09555988224144528, + "loss": 0.2762, + "num_input_tokens_seen": 18860288, + "step": 24730 + }, + { + "epoch": 51.42411642411643, + "grad_norm": 0.0002846041170414537, + "learning_rate": 0.09550499801795717, + "loss": 0.2629, + "num_input_tokens_seen": 18864224, + "step": 24735 + }, + { + "epoch": 51.434511434511435, + "grad_norm": 0.00010604729322949424, + "learning_rate": 0.09545012219828314, + "loss": 0.2438, + "num_input_tokens_seen": 18867904, + "step": 24740 + }, + { + "epoch": 51.444906444906444, + "grad_norm": 0.0002491792547516525, + "learning_rate": 0.09539525479088577, + "loss": 0.3055, + "num_input_tokens_seen": 18871744, + "step": 24745 + }, + { + "epoch": 51.45530145530145, + "grad_norm": 0.00022967495897319168, + "learning_rate": 0.0953403958042264, + "loss": 0.2648, + "num_input_tokens_seen": 18875648, + "step": 24750 + }, + { + "epoch": 51.46569646569647, + "grad_norm": 9.365454752696678e-05, + "learning_rate": 0.09528554524676484, + "loss": 0.259, + "num_input_tokens_seen": 18879584, + "step": 24755 + }, + { + "epoch": 51.476091476091476, + "grad_norm": 7.376411667792127e-05, + "learning_rate": 0.09523070312695978, + "loss": 0.2777, + "num_input_tokens_seen": 18883520, + "step": 24760 + }, + { + "epoch": 51.486486486486484, + "grad_norm": 0.00016310378850903362, + "learning_rate": 0.09517586945326863, + "loss": 0.2785, + "num_input_tokens_seen": 18887360, + "step": 24765 + }, + { + "epoch": 51.4968814968815, + "grad_norm": 0.00025557290064170957, + "learning_rate": 0.0951210442341473, + "loss": 0.2668, + "num_input_tokens_seen": 18891168, + "step": 24770 + }, + { + "epoch": 51.50727650727651, + "grad_norm": 0.00027795921778306365, + "learning_rate": 0.09506622747805066, + "loss": 0.2609, + "num_input_tokens_seen": 18894912, + "step": 24775 + }, + { + "epoch": 51.517671517671516, + "grad_norm": 9.248357673641294e-05, + "learning_rate": 0.09501141919343203, + "loss": 0.2531, + "num_input_tokens_seen": 18898816, + "step": 24780 + }, + { + "epoch": 51.528066528066525, + "grad_norm": 6.760465475963429e-05, + "learning_rate": 0.09495661938874361, + "loss": 0.2784, + "num_input_tokens_seen": 18902624, + "step": 24785 + }, + { + "epoch": 51.53846153846154, + "grad_norm": 0.0001204007858177647, + "learning_rate": 0.0949018280724362, + "loss": 0.2584, + "num_input_tokens_seen": 18906400, + "step": 24790 + }, + { + "epoch": 51.54885654885655, + "grad_norm": 0.00013308181951288134, + "learning_rate": 0.09484704525295934, + "loss": 0.2505, + "num_input_tokens_seen": 18910240, + "step": 24795 + }, + { + "epoch": 51.55925155925156, + "grad_norm": 0.00047808833187446, + "learning_rate": 0.09479227093876112, + "loss": 0.2414, + "num_input_tokens_seen": 18913856, + "step": 24800 + }, + { + "epoch": 51.55925155925156, + "eval_loss": 0.24719470739364624, + "eval_runtime": 13.3894, + "eval_samples_per_second": 63.931, + "eval_steps_per_second": 15.983, + "num_input_tokens_seen": 18913856, + "step": 24800 + }, + { + "epoch": 51.56964656964657, + "grad_norm": 0.00014398552593775094, + "learning_rate": 0.0947375051382886, + "loss": 0.2605, + "num_input_tokens_seen": 18917600, + "step": 24805 + }, + { + "epoch": 51.58004158004158, + "grad_norm": 0.00012486219929996878, + "learning_rate": 0.09468274785998718, + "loss": 0.2462, + "num_input_tokens_seen": 18921344, + "step": 24810 + }, + { + "epoch": 51.59043659043659, + "grad_norm": 0.00010473545989952981, + "learning_rate": 0.09462799911230127, + "loss": 0.2746, + "num_input_tokens_seen": 18925248, + "step": 24815 + }, + { + "epoch": 51.6008316008316, + "grad_norm": 0.0002909967442974448, + "learning_rate": 0.0945732589036737, + "loss": 0.2749, + "num_input_tokens_seen": 18929056, + "step": 24820 + }, + { + "epoch": 51.61122661122661, + "grad_norm": 0.00019620853709056973, + "learning_rate": 0.09451852724254614, + "loss": 0.2697, + "num_input_tokens_seen": 18932864, + "step": 24825 + }, + { + "epoch": 51.62162162162162, + "grad_norm": 0.00010172414476983249, + "learning_rate": 0.09446380413735894, + "loss": 0.2552, + "num_input_tokens_seen": 18936736, + "step": 24830 + }, + { + "epoch": 51.63201663201663, + "grad_norm": 0.00010708453191909939, + "learning_rate": 0.09440908959655099, + "loss": 0.249, + "num_input_tokens_seen": 18940608, + "step": 24835 + }, + { + "epoch": 51.642411642411645, + "grad_norm": 0.00036251608980819583, + "learning_rate": 0.09435438362856004, + "loss": 0.2804, + "num_input_tokens_seen": 18944448, + "step": 24840 + }, + { + "epoch": 51.65280665280665, + "grad_norm": 0.0007139127119444311, + "learning_rate": 0.0942996862418225, + "loss": 0.2722, + "num_input_tokens_seen": 18948352, + "step": 24845 + }, + { + "epoch": 51.66320166320166, + "grad_norm": 0.00022203786647878587, + "learning_rate": 0.09424499744477322, + "loss": 0.2721, + "num_input_tokens_seen": 18952160, + "step": 24850 + }, + { + "epoch": 51.67359667359668, + "grad_norm": 0.00028811063384637237, + "learning_rate": 0.09419031724584608, + "loss": 0.2486, + "num_input_tokens_seen": 18955840, + "step": 24855 + }, + { + "epoch": 51.683991683991685, + "grad_norm": 0.00017130660125985742, + "learning_rate": 0.09413564565347331, + "loss": 0.2615, + "num_input_tokens_seen": 18959712, + "step": 24860 + }, + { + "epoch": 51.694386694386694, + "grad_norm": 0.0005064566503278911, + "learning_rate": 0.094080982676086, + "loss": 0.2501, + "num_input_tokens_seen": 18963616, + "step": 24865 + }, + { + "epoch": 51.7047817047817, + "grad_norm": 3.131589983240701e-05, + "learning_rate": 0.09402632832211395, + "loss": 0.2653, + "num_input_tokens_seen": 18967520, + "step": 24870 + }, + { + "epoch": 51.71517671517672, + "grad_norm": 8.942348358687013e-05, + "learning_rate": 0.09397168259998541, + "loss": 0.2633, + "num_input_tokens_seen": 18971168, + "step": 24875 + }, + { + "epoch": 51.725571725571726, + "grad_norm": 0.0004058416816405952, + "learning_rate": 0.09391704551812759, + "loss": 0.2659, + "num_input_tokens_seen": 18974944, + "step": 24880 + }, + { + "epoch": 51.735966735966734, + "grad_norm": 0.00015918316785246134, + "learning_rate": 0.09386241708496605, + "loss": 0.2931, + "num_input_tokens_seen": 18978880, + "step": 24885 + }, + { + "epoch": 51.74636174636175, + "grad_norm": 0.0003292606270406395, + "learning_rate": 0.09380779730892527, + "loss": 0.2582, + "num_input_tokens_seen": 18982656, + "step": 24890 + }, + { + "epoch": 51.75675675675676, + "grad_norm": 0.00020529964240267873, + "learning_rate": 0.09375318619842836, + "loss": 0.2664, + "num_input_tokens_seen": 18986432, + "step": 24895 + }, + { + "epoch": 51.767151767151766, + "grad_norm": 0.000501190428622067, + "learning_rate": 0.09369858376189696, + "loss": 0.2668, + "num_input_tokens_seen": 18990304, + "step": 24900 + }, + { + "epoch": 51.777546777546775, + "grad_norm": 0.0001369376404909417, + "learning_rate": 0.09364399000775143, + "loss": 0.271, + "num_input_tokens_seen": 18994112, + "step": 24905 + }, + { + "epoch": 51.78794178794179, + "grad_norm": 9.09310911083594e-05, + "learning_rate": 0.09358940494441093, + "loss": 0.2555, + "num_input_tokens_seen": 18998272, + "step": 24910 + }, + { + "epoch": 51.7983367983368, + "grad_norm": 0.00047601928235962987, + "learning_rate": 0.09353482858029301, + "loss": 0.2285, + "num_input_tokens_seen": 19002144, + "step": 24915 + }, + { + "epoch": 51.80873180873181, + "grad_norm": 0.00019199545204173774, + "learning_rate": 0.09348026092381419, + "loss": 0.2494, + "num_input_tokens_seen": 19006016, + "step": 24920 + }, + { + "epoch": 51.81912681912682, + "grad_norm": 0.0001717114937491715, + "learning_rate": 0.09342570198338931, + "loss": 0.2846, + "num_input_tokens_seen": 19009856, + "step": 24925 + }, + { + "epoch": 51.82952182952183, + "grad_norm": 7.368923252215609e-05, + "learning_rate": 0.0933711517674322, + "loss": 0.2853, + "num_input_tokens_seen": 19013696, + "step": 24930 + }, + { + "epoch": 51.83991683991684, + "grad_norm": 0.00044414735748432577, + "learning_rate": 0.09331661028435513, + "loss": 0.2561, + "num_input_tokens_seen": 19017504, + "step": 24935 + }, + { + "epoch": 51.85031185031185, + "grad_norm": 0.0004908243427053094, + "learning_rate": 0.09326207754256909, + "loss": 0.2656, + "num_input_tokens_seen": 19021280, + "step": 24940 + }, + { + "epoch": 51.86070686070686, + "grad_norm": 0.00013243909052107483, + "learning_rate": 0.09320755355048366, + "loss": 0.263, + "num_input_tokens_seen": 19025120, + "step": 24945 + }, + { + "epoch": 51.87110187110187, + "grad_norm": 0.00020346205565147102, + "learning_rate": 0.09315303831650722, + "loss": 0.273, + "num_input_tokens_seen": 19028896, + "step": 24950 + }, + { + "epoch": 51.88149688149688, + "grad_norm": 0.0003335009969305247, + "learning_rate": 0.09309853184904661, + "loss": 0.2524, + "num_input_tokens_seen": 19032832, + "step": 24955 + }, + { + "epoch": 51.891891891891895, + "grad_norm": 0.0003491482057143003, + "learning_rate": 0.09304403415650753, + "loss": 0.2586, + "num_input_tokens_seen": 19036608, + "step": 24960 + }, + { + "epoch": 51.9022869022869, + "grad_norm": 0.00011952204658882692, + "learning_rate": 0.09298954524729405, + "loss": 0.2572, + "num_input_tokens_seen": 19040384, + "step": 24965 + }, + { + "epoch": 51.91268191268191, + "grad_norm": 0.00010351989476475865, + "learning_rate": 0.09293506512980916, + "loss": 0.2685, + "num_input_tokens_seen": 19044128, + "step": 24970 + }, + { + "epoch": 51.92307692307692, + "grad_norm": 0.00025333426310680807, + "learning_rate": 0.0928805938124544, + "loss": 0.2752, + "num_input_tokens_seen": 19047936, + "step": 24975 + }, + { + "epoch": 51.933471933471935, + "grad_norm": 0.0004497557529248297, + "learning_rate": 0.09282613130362982, + "loss": 0.2653, + "num_input_tokens_seen": 19051712, + "step": 24980 + }, + { + "epoch": 51.943866943866944, + "grad_norm": 0.00020699689048342407, + "learning_rate": 0.09277167761173427, + "loss": 0.2794, + "num_input_tokens_seen": 19055392, + "step": 24985 + }, + { + "epoch": 51.95426195426195, + "grad_norm": 0.00016919725749175996, + "learning_rate": 0.0927172327451653, + "loss": 0.2742, + "num_input_tokens_seen": 19059136, + "step": 24990 + }, + { + "epoch": 51.96465696465697, + "grad_norm": 0.00012547911319416016, + "learning_rate": 0.09266279671231882, + "loss": 0.2621, + "num_input_tokens_seen": 19062880, + "step": 24995 + }, + { + "epoch": 51.975051975051976, + "grad_norm": 8.184587204596028e-05, + "learning_rate": 0.09260836952158967, + "loss": 0.2668, + "num_input_tokens_seen": 19066528, + "step": 25000 + }, + { + "epoch": 51.975051975051976, + "eval_loss": 0.24882346391677856, + "eval_runtime": 13.3923, + "eval_samples_per_second": 63.918, + "eval_steps_per_second": 15.979, + "num_input_tokens_seen": 19066528, + "step": 25000 + }, + { + "epoch": 51.985446985446984, + "grad_norm": 3.0351191526278853e-05, + "learning_rate": 0.09255395118137114, + "loss": 0.2454, + "num_input_tokens_seen": 19070368, + "step": 25005 + }, + { + "epoch": 51.99584199584199, + "grad_norm": 0.00014118723629508168, + "learning_rate": 0.09249954170005527, + "loss": 0.2351, + "num_input_tokens_seen": 19074368, + "step": 25010 + }, + { + "epoch": 52.00623700623701, + "grad_norm": 5.4178875871002674e-05, + "learning_rate": 0.0924451410860327, + "loss": 0.2854, + "num_input_tokens_seen": 19078008, + "step": 25015 + }, + { + "epoch": 52.016632016632016, + "grad_norm": 0.0001686487375991419, + "learning_rate": 0.09239074934769258, + "loss": 0.2552, + "num_input_tokens_seen": 19081944, + "step": 25020 + }, + { + "epoch": 52.027027027027025, + "grad_norm": 0.00012034373503411189, + "learning_rate": 0.09233636649342288, + "loss": 0.2626, + "num_input_tokens_seen": 19085592, + "step": 25025 + }, + { + "epoch": 52.03742203742204, + "grad_norm": 0.00013611235772259533, + "learning_rate": 0.09228199253161017, + "loss": 0.253, + "num_input_tokens_seen": 19089496, + "step": 25030 + }, + { + "epoch": 52.04781704781705, + "grad_norm": 0.00010463312355568632, + "learning_rate": 0.09222762747063949, + "loss": 0.2752, + "num_input_tokens_seen": 19093368, + "step": 25035 + }, + { + "epoch": 52.05821205821206, + "grad_norm": 0.00023305570357479155, + "learning_rate": 0.09217327131889473, + "loss": 0.2671, + "num_input_tokens_seen": 19097048, + "step": 25040 + }, + { + "epoch": 52.06860706860707, + "grad_norm": 0.0006485827616415918, + "learning_rate": 0.09211892408475818, + "loss": 0.2694, + "num_input_tokens_seen": 19100792, + "step": 25045 + }, + { + "epoch": 52.07900207900208, + "grad_norm": 0.00048761325888335705, + "learning_rate": 0.09206458577661089, + "loss": 0.2725, + "num_input_tokens_seen": 19104568, + "step": 25050 + }, + { + "epoch": 52.08939708939709, + "grad_norm": 0.00032891842420212924, + "learning_rate": 0.09201025640283263, + "loss": 0.2671, + "num_input_tokens_seen": 19108632, + "step": 25055 + }, + { + "epoch": 52.0997920997921, + "grad_norm": 0.0001725823531160131, + "learning_rate": 0.09195593597180148, + "loss": 0.2589, + "num_input_tokens_seen": 19112376, + "step": 25060 + }, + { + "epoch": 52.11018711018711, + "grad_norm": 8.355977479368448e-05, + "learning_rate": 0.09190162449189444, + "loss": 0.2598, + "num_input_tokens_seen": 19116184, + "step": 25065 + }, + { + "epoch": 52.12058212058212, + "grad_norm": 0.00017736456356942654, + "learning_rate": 0.09184732197148705, + "loss": 0.2458, + "num_input_tokens_seen": 19119960, + "step": 25070 + }, + { + "epoch": 52.13097713097713, + "grad_norm": 0.00014506731531582773, + "learning_rate": 0.09179302841895343, + "loss": 0.2838, + "num_input_tokens_seen": 19123672, + "step": 25075 + }, + { + "epoch": 52.141372141372145, + "grad_norm": 0.00021496537374332547, + "learning_rate": 0.09173874384266625, + "loss": 0.2552, + "num_input_tokens_seen": 19127608, + "step": 25080 + }, + { + "epoch": 52.15176715176715, + "grad_norm": 0.0004486608086153865, + "learning_rate": 0.09168446825099695, + "loss": 0.2744, + "num_input_tokens_seen": 19131352, + "step": 25085 + }, + { + "epoch": 52.16216216216216, + "grad_norm": 0.0001984643458854407, + "learning_rate": 0.09163020165231545, + "loss": 0.2679, + "num_input_tokens_seen": 19135256, + "step": 25090 + }, + { + "epoch": 52.17255717255717, + "grad_norm": 0.0003029173822142184, + "learning_rate": 0.09157594405499044, + "loss": 0.2552, + "num_input_tokens_seen": 19139000, + "step": 25095 + }, + { + "epoch": 52.182952182952185, + "grad_norm": 0.0006888463976792991, + "learning_rate": 0.09152169546738899, + "loss": 0.2671, + "num_input_tokens_seen": 19142744, + "step": 25100 + }, + { + "epoch": 52.19334719334719, + "grad_norm": 0.0003807439934462309, + "learning_rate": 0.09146745589787698, + "loss": 0.2717, + "num_input_tokens_seen": 19146616, + "step": 25105 + }, + { + "epoch": 52.2037422037422, + "grad_norm": 0.0003893484245054424, + "learning_rate": 0.09141322535481891, + "loss": 0.286, + "num_input_tokens_seen": 19150328, + "step": 25110 + }, + { + "epoch": 52.21413721413722, + "grad_norm": 8.591903315391392e-05, + "learning_rate": 0.0913590038465777, + "loss": 0.2721, + "num_input_tokens_seen": 19154104, + "step": 25115 + }, + { + "epoch": 52.224532224532226, + "grad_norm": 0.00023204648459795862, + "learning_rate": 0.09130479138151505, + "loss": 0.2782, + "num_input_tokens_seen": 19157816, + "step": 25120 + }, + { + "epoch": 52.234927234927234, + "grad_norm": 0.0004267179756425321, + "learning_rate": 0.09125058796799114, + "loss": 0.2735, + "num_input_tokens_seen": 19161496, + "step": 25125 + }, + { + "epoch": 52.24532224532224, + "grad_norm": 0.00018816835654433817, + "learning_rate": 0.09119639361436485, + "loss": 0.2618, + "num_input_tokens_seen": 19165240, + "step": 25130 + }, + { + "epoch": 52.25571725571726, + "grad_norm": 0.0002517510438337922, + "learning_rate": 0.09114220832899368, + "loss": 0.2521, + "num_input_tokens_seen": 19169240, + "step": 25135 + }, + { + "epoch": 52.266112266112266, + "grad_norm": 0.00038964030682109296, + "learning_rate": 0.0910880321202336, + "loss": 0.2596, + "num_input_tokens_seen": 19172952, + "step": 25140 + }, + { + "epoch": 52.276507276507274, + "grad_norm": 0.00012928710202686489, + "learning_rate": 0.09103386499643933, + "loss": 0.2559, + "num_input_tokens_seen": 19176856, + "step": 25145 + }, + { + "epoch": 52.28690228690229, + "grad_norm": 0.0004698314005509019, + "learning_rate": 0.09097970696596407, + "loss": 0.2959, + "num_input_tokens_seen": 19180600, + "step": 25150 + }, + { + "epoch": 52.2972972972973, + "grad_norm": 0.00024322712852153927, + "learning_rate": 0.09092555803715971, + "loss": 0.2792, + "num_input_tokens_seen": 19184408, + "step": 25155 + }, + { + "epoch": 52.30769230769231, + "grad_norm": 0.0001793637202354148, + "learning_rate": 0.0908714182183767, + "loss": 0.2761, + "num_input_tokens_seen": 19188312, + "step": 25160 + }, + { + "epoch": 52.318087318087315, + "grad_norm": 0.0005349906859919429, + "learning_rate": 0.090817287517964, + "loss": 0.2593, + "num_input_tokens_seen": 19191992, + "step": 25165 + }, + { + "epoch": 52.32848232848233, + "grad_norm": 0.00027267204131931067, + "learning_rate": 0.09076316594426931, + "loss": 0.292, + "num_input_tokens_seen": 19195896, + "step": 25170 + }, + { + "epoch": 52.33887733887734, + "grad_norm": 9.870494977803901e-05, + "learning_rate": 0.09070905350563888, + "loss": 0.2545, + "num_input_tokens_seen": 19199736, + "step": 25175 + }, + { + "epoch": 52.34927234927235, + "grad_norm": 0.00015784376591909677, + "learning_rate": 0.09065495021041745, + "loss": 0.247, + "num_input_tokens_seen": 19203544, + "step": 25180 + }, + { + "epoch": 52.35966735966736, + "grad_norm": 0.0001943733950611204, + "learning_rate": 0.09060085606694851, + "loss": 0.2453, + "num_input_tokens_seen": 19207288, + "step": 25185 + }, + { + "epoch": 52.37006237006237, + "grad_norm": 0.0005111717619001865, + "learning_rate": 0.09054677108357405, + "loss": 0.2551, + "num_input_tokens_seen": 19211128, + "step": 25190 + }, + { + "epoch": 52.38045738045738, + "grad_norm": 0.0001630577608011663, + "learning_rate": 0.09049269526863457, + "loss": 0.287, + "num_input_tokens_seen": 19214904, + "step": 25195 + }, + { + "epoch": 52.39085239085239, + "grad_norm": 0.00021629508410114795, + "learning_rate": 0.09043862863046935, + "loss": 0.2498, + "num_input_tokens_seen": 19218616, + "step": 25200 + }, + { + "epoch": 52.39085239085239, + "eval_loss": 0.251054972410202, + "eval_runtime": 13.4132, + "eval_samples_per_second": 63.818, + "eval_steps_per_second": 15.954, + "num_input_tokens_seen": 19218616, + "step": 25200 + }, + { + "epoch": 52.4012474012474, + "grad_norm": 0.00021273968741297722, + "learning_rate": 0.09038457117741602, + "loss": 0.2796, + "num_input_tokens_seen": 19222488, + "step": 25205 + }, + { + "epoch": 52.41164241164241, + "grad_norm": 0.000733297667466104, + "learning_rate": 0.09033052291781099, + "loss": 0.246, + "num_input_tokens_seen": 19226296, + "step": 25210 + }, + { + "epoch": 52.42203742203742, + "grad_norm": 0.0006143064820207655, + "learning_rate": 0.09027648385998926, + "loss": 0.2878, + "num_input_tokens_seen": 19230104, + "step": 25215 + }, + { + "epoch": 52.432432432432435, + "grad_norm": 5.60740809305571e-05, + "learning_rate": 0.09022245401228417, + "loss": 0.2609, + "num_input_tokens_seen": 19233912, + "step": 25220 + }, + { + "epoch": 52.44282744282744, + "grad_norm": 0.00037995801540091634, + "learning_rate": 0.09016843338302792, + "loss": 0.2736, + "num_input_tokens_seen": 19237688, + "step": 25225 + }, + { + "epoch": 52.45322245322245, + "grad_norm": 0.00019657934899441898, + "learning_rate": 0.09011442198055115, + "loss": 0.2523, + "num_input_tokens_seen": 19241464, + "step": 25230 + }, + { + "epoch": 52.46361746361746, + "grad_norm": 0.0004387819499243051, + "learning_rate": 0.09006041981318305, + "loss": 0.2744, + "num_input_tokens_seen": 19245496, + "step": 25235 + }, + { + "epoch": 52.474012474012476, + "grad_norm": 0.0006431292276829481, + "learning_rate": 0.09000642688925149, + "loss": 0.2478, + "num_input_tokens_seen": 19249240, + "step": 25240 + }, + { + "epoch": 52.484407484407484, + "grad_norm": 0.00035885735996998847, + "learning_rate": 0.0899524432170828, + "loss": 0.2418, + "num_input_tokens_seen": 19252984, + "step": 25245 + }, + { + "epoch": 52.49480249480249, + "grad_norm": 0.00022899993928149343, + "learning_rate": 0.08989846880500196, + "loss": 0.2382, + "num_input_tokens_seen": 19256760, + "step": 25250 + }, + { + "epoch": 52.50519750519751, + "grad_norm": 0.00016251785564236343, + "learning_rate": 0.08984450366133256, + "loss": 0.2313, + "num_input_tokens_seen": 19260664, + "step": 25255 + }, + { + "epoch": 52.515592515592516, + "grad_norm": 0.00027021896676160395, + "learning_rate": 0.08979054779439664, + "loss": 0.2866, + "num_input_tokens_seen": 19264472, + "step": 25260 + }, + { + "epoch": 52.525987525987524, + "grad_norm": 0.00012837324175052345, + "learning_rate": 0.08973660121251485, + "loss": 0.3001, + "num_input_tokens_seen": 19268216, + "step": 25265 + }, + { + "epoch": 52.53638253638254, + "grad_norm": 0.0005488950992003083, + "learning_rate": 0.08968266392400655, + "loss": 0.2648, + "num_input_tokens_seen": 19272120, + "step": 25270 + }, + { + "epoch": 52.54677754677755, + "grad_norm": 0.00010601641406537965, + "learning_rate": 0.0896287359371894, + "loss": 0.2671, + "num_input_tokens_seen": 19275992, + "step": 25275 + }, + { + "epoch": 52.55717255717256, + "grad_norm": 0.0002984137099701911, + "learning_rate": 0.08957481726037989, + "loss": 0.2844, + "num_input_tokens_seen": 19279896, + "step": 25280 + }, + { + "epoch": 52.567567567567565, + "grad_norm": 7.243233267217875e-05, + "learning_rate": 0.08952090790189286, + "loss": 0.267, + "num_input_tokens_seen": 19283800, + "step": 25285 + }, + { + "epoch": 52.57796257796258, + "grad_norm": 0.00010885141819017008, + "learning_rate": 0.08946700787004187, + "loss": 0.2634, + "num_input_tokens_seen": 19287480, + "step": 25290 + }, + { + "epoch": 52.58835758835759, + "grad_norm": 0.00011178448767168447, + "learning_rate": 0.08941311717313899, + "loss": 0.2931, + "num_input_tokens_seen": 19291288, + "step": 25295 + }, + { + "epoch": 52.5987525987526, + "grad_norm": 0.00012702326057478786, + "learning_rate": 0.08935923581949483, + "loss": 0.2642, + "num_input_tokens_seen": 19295192, + "step": 25300 + }, + { + "epoch": 52.60914760914761, + "grad_norm": 6.52926682960242e-05, + "learning_rate": 0.0893053638174185, + "loss": 0.2841, + "num_input_tokens_seen": 19298872, + "step": 25305 + }, + { + "epoch": 52.61954261954262, + "grad_norm": 0.0003655454784166068, + "learning_rate": 0.0892515011752179, + "loss": 0.2569, + "num_input_tokens_seen": 19302584, + "step": 25310 + }, + { + "epoch": 52.62993762993763, + "grad_norm": 0.00011825864930870011, + "learning_rate": 0.08919764790119918, + "loss": 0.2391, + "num_input_tokens_seen": 19306328, + "step": 25315 + }, + { + "epoch": 52.64033264033264, + "grad_norm": 0.0002848030417226255, + "learning_rate": 0.08914380400366727, + "loss": 0.2526, + "num_input_tokens_seen": 19310328, + "step": 25320 + }, + { + "epoch": 52.65072765072765, + "grad_norm": 0.00033655259176157415, + "learning_rate": 0.08908996949092551, + "loss": 0.2618, + "num_input_tokens_seen": 19314104, + "step": 25325 + }, + { + "epoch": 52.66112266112266, + "grad_norm": 0.0003371915954630822, + "learning_rate": 0.08903614437127592, + "loss": 0.2577, + "num_input_tokens_seen": 19317880, + "step": 25330 + }, + { + "epoch": 52.67151767151767, + "grad_norm": 0.00027667218819260597, + "learning_rate": 0.088982328653019, + "loss": 0.2564, + "num_input_tokens_seen": 19321720, + "step": 25335 + }, + { + "epoch": 52.681912681912685, + "grad_norm": 8.618910942459479e-05, + "learning_rate": 0.0889285223444538, + "loss": 0.2631, + "num_input_tokens_seen": 19325560, + "step": 25340 + }, + { + "epoch": 52.69230769230769, + "grad_norm": 0.00023816869361326098, + "learning_rate": 0.08887472545387787, + "loss": 0.298, + "num_input_tokens_seen": 19329368, + "step": 25345 + }, + { + "epoch": 52.7027027027027, + "grad_norm": 0.0003238253411836922, + "learning_rate": 0.08882093798958751, + "loss": 0.2758, + "num_input_tokens_seen": 19333368, + "step": 25350 + }, + { + "epoch": 52.71309771309771, + "grad_norm": 0.0005789937567897141, + "learning_rate": 0.08876715995987726, + "loss": 0.2739, + "num_input_tokens_seen": 19337112, + "step": 25355 + }, + { + "epoch": 52.723492723492726, + "grad_norm": 0.0003477366117294878, + "learning_rate": 0.08871339137304052, + "loss": 0.2753, + "num_input_tokens_seen": 19340952, + "step": 25360 + }, + { + "epoch": 52.733887733887734, + "grad_norm": 0.00014633526734542102, + "learning_rate": 0.0886596322373689, + "loss": 0.2725, + "num_input_tokens_seen": 19344728, + "step": 25365 + }, + { + "epoch": 52.74428274428274, + "grad_norm": 0.0005999618442729115, + "learning_rate": 0.08860588256115293, + "loss": 0.2568, + "num_input_tokens_seen": 19348504, + "step": 25370 + }, + { + "epoch": 52.75467775467776, + "grad_norm": 0.00025870188255794346, + "learning_rate": 0.0885521423526814, + "loss": 0.2696, + "num_input_tokens_seen": 19352152, + "step": 25375 + }, + { + "epoch": 52.765072765072766, + "grad_norm": 0.0002299930201843381, + "learning_rate": 0.08849841162024165, + "loss": 0.2433, + "num_input_tokens_seen": 19355992, + "step": 25380 + }, + { + "epoch": 52.775467775467774, + "grad_norm": 0.00029891615849919617, + "learning_rate": 0.08844469037211973, + "loss": 0.2841, + "num_input_tokens_seen": 19359736, + "step": 25385 + }, + { + "epoch": 52.78586278586278, + "grad_norm": 0.0001081323207472451, + "learning_rate": 0.08839097861660014, + "loss": 0.2805, + "num_input_tokens_seen": 19363384, + "step": 25390 + }, + { + "epoch": 52.7962577962578, + "grad_norm": 8.824890392133966e-05, + "learning_rate": 0.08833727636196585, + "loss": 0.269, + "num_input_tokens_seen": 19367192, + "step": 25395 + }, + { + "epoch": 52.80665280665281, + "grad_norm": 0.0001952427119249478, + "learning_rate": 0.08828358361649848, + "loss": 0.2658, + "num_input_tokens_seen": 19370872, + "step": 25400 + }, + { + "epoch": 52.80665280665281, + "eval_loss": 0.2578355073928833, + "eval_runtime": 13.3869, + "eval_samples_per_second": 63.943, + "eval_steps_per_second": 15.986, + "num_input_tokens_seen": 19370872, + "step": 25400 + }, + { + "epoch": 52.817047817047815, + "grad_norm": 6.909188232384622e-05, + "learning_rate": 0.08822990038847807, + "loss": 0.2646, + "num_input_tokens_seen": 19374840, + "step": 25405 + }, + { + "epoch": 52.82744282744283, + "grad_norm": 0.00022669322788715363, + "learning_rate": 0.08817622668618325, + "loss": 0.2593, + "num_input_tokens_seen": 19378616, + "step": 25410 + }, + { + "epoch": 52.83783783783784, + "grad_norm": 0.0005334245506674051, + "learning_rate": 0.08812256251789125, + "loss": 0.2794, + "num_input_tokens_seen": 19382328, + "step": 25415 + }, + { + "epoch": 52.84823284823285, + "grad_norm": 0.00016129582945723087, + "learning_rate": 0.08806890789187766, + "loss": 0.2406, + "num_input_tokens_seen": 19385976, + "step": 25420 + }, + { + "epoch": 52.858627858627855, + "grad_norm": 0.00036462044226936996, + "learning_rate": 0.08801526281641672, + "loss": 0.2655, + "num_input_tokens_seen": 19389944, + "step": 25425 + }, + { + "epoch": 52.86902286902287, + "grad_norm": 0.0006434862734749913, + "learning_rate": 0.0879616272997813, + "loss": 0.2562, + "num_input_tokens_seen": 19393784, + "step": 25430 + }, + { + "epoch": 52.87941787941788, + "grad_norm": 0.00046026933705434203, + "learning_rate": 0.08790800135024247, + "loss": 0.2453, + "num_input_tokens_seen": 19397656, + "step": 25435 + }, + { + "epoch": 52.88981288981289, + "grad_norm": 0.00041225148015655577, + "learning_rate": 0.08785438497607023, + "loss": 0.3011, + "num_input_tokens_seen": 19401656, + "step": 25440 + }, + { + "epoch": 52.9002079002079, + "grad_norm": 0.0007981536327861249, + "learning_rate": 0.08780077818553277, + "loss": 0.2659, + "num_input_tokens_seen": 19405592, + "step": 25445 + }, + { + "epoch": 52.91060291060291, + "grad_norm": 0.000723455857951194, + "learning_rate": 0.0877471809868969, + "loss": 0.2676, + "num_input_tokens_seen": 19409496, + "step": 25450 + }, + { + "epoch": 52.92099792099792, + "grad_norm": 4.023011933895759e-05, + "learning_rate": 0.08769359338842811, + "loss": 0.2633, + "num_input_tokens_seen": 19413240, + "step": 25455 + }, + { + "epoch": 52.931392931392935, + "grad_norm": 0.00022103404626250267, + "learning_rate": 0.08764001539839016, + "loss": 0.2448, + "num_input_tokens_seen": 19417176, + "step": 25460 + }, + { + "epoch": 52.94178794178794, + "grad_norm": 0.00010065849346574396, + "learning_rate": 0.08758644702504548, + "loss": 0.253, + "num_input_tokens_seen": 19421016, + "step": 25465 + }, + { + "epoch": 52.95218295218295, + "grad_norm": 0.00019730512576643378, + "learning_rate": 0.0875328882766551, + "loss": 0.3083, + "num_input_tokens_seen": 19424760, + "step": 25470 + }, + { + "epoch": 52.96257796257796, + "grad_norm": 0.00032311183167621493, + "learning_rate": 0.08747933916147828, + "loss": 0.2687, + "num_input_tokens_seen": 19428568, + "step": 25475 + }, + { + "epoch": 52.972972972972975, + "grad_norm": 0.0005062875570729375, + "learning_rate": 0.0874257996877731, + "loss": 0.2447, + "num_input_tokens_seen": 19432344, + "step": 25480 + }, + { + "epoch": 52.983367983367984, + "grad_norm": 0.0002024264686042443, + "learning_rate": 0.08737226986379593, + "loss": 0.254, + "num_input_tokens_seen": 19436120, + "step": 25485 + }, + { + "epoch": 52.99376299376299, + "grad_norm": 7.391955296043307e-05, + "learning_rate": 0.08731874969780173, + "loss": 0.2742, + "num_input_tokens_seen": 19439960, + "step": 25490 + }, + { + "epoch": 53.00415800415801, + "grad_norm": 0.00022842263570055366, + "learning_rate": 0.08726523919804412, + "loss": 0.245, + "num_input_tokens_seen": 19443752, + "step": 25495 + }, + { + "epoch": 53.014553014553016, + "grad_norm": 0.0002269794058520347, + "learning_rate": 0.08721173837277492, + "loss": 0.2573, + "num_input_tokens_seen": 19447496, + "step": 25500 + }, + { + "epoch": 53.024948024948024, + "grad_norm": 0.0003161179192829877, + "learning_rate": 0.08715824723024479, + "loss": 0.2689, + "num_input_tokens_seen": 19451304, + "step": 25505 + }, + { + "epoch": 53.03534303534303, + "grad_norm": 0.00027071021031588316, + "learning_rate": 0.08710476577870258, + "loss": 0.2705, + "num_input_tokens_seen": 19455240, + "step": 25510 + }, + { + "epoch": 53.04573804573805, + "grad_norm": 0.0002087950415443629, + "learning_rate": 0.08705129402639587, + "loss": 0.2858, + "num_input_tokens_seen": 19458888, + "step": 25515 + }, + { + "epoch": 53.056133056133056, + "grad_norm": 0.0003871911612804979, + "learning_rate": 0.08699783198157078, + "loss": 0.2792, + "num_input_tokens_seen": 19462728, + "step": 25520 + }, + { + "epoch": 53.066528066528065, + "grad_norm": 0.00022111559519544244, + "learning_rate": 0.08694437965247163, + "loss": 0.2683, + "num_input_tokens_seen": 19466536, + "step": 25525 + }, + { + "epoch": 53.07692307692308, + "grad_norm": 0.00046337186358869076, + "learning_rate": 0.08689093704734165, + "loss": 0.2738, + "num_input_tokens_seen": 19470344, + "step": 25530 + }, + { + "epoch": 53.08731808731809, + "grad_norm": 0.00012477106065489352, + "learning_rate": 0.08683750417442222, + "loss": 0.2704, + "num_input_tokens_seen": 19474344, + "step": 25535 + }, + { + "epoch": 53.0977130977131, + "grad_norm": 0.00011562240979401395, + "learning_rate": 0.08678408104195334, + "loss": 0.2814, + "num_input_tokens_seen": 19478312, + "step": 25540 + }, + { + "epoch": 53.108108108108105, + "grad_norm": 0.0009324764832854271, + "learning_rate": 0.08673066765817365, + "loss": 0.2691, + "num_input_tokens_seen": 19482088, + "step": 25545 + }, + { + "epoch": 53.11850311850312, + "grad_norm": 0.00013382262841332704, + "learning_rate": 0.08667726403132005, + "loss": 0.2506, + "num_input_tokens_seen": 19485928, + "step": 25550 + }, + { + "epoch": 53.12889812889813, + "grad_norm": 8.289182733278722e-05, + "learning_rate": 0.0866238701696281, + "loss": 0.2534, + "num_input_tokens_seen": 19489736, + "step": 25555 + }, + { + "epoch": 53.13929313929314, + "grad_norm": 0.00021687426487915218, + "learning_rate": 0.08657048608133185, + "loss": 0.275, + "num_input_tokens_seen": 19493544, + "step": 25560 + }, + { + "epoch": 53.14968814968815, + "grad_norm": 9.0418616309762e-05, + "learning_rate": 0.08651711177466369, + "loss": 0.2662, + "num_input_tokens_seen": 19497416, + "step": 25565 + }, + { + "epoch": 53.16008316008316, + "grad_norm": 0.0001370221289107576, + "learning_rate": 0.08646374725785466, + "loss": 0.274, + "num_input_tokens_seen": 19501320, + "step": 25570 + }, + { + "epoch": 53.17047817047817, + "grad_norm": 0.0005844387924298644, + "learning_rate": 0.08641039253913434, + "loss": 0.2659, + "num_input_tokens_seen": 19505192, + "step": 25575 + }, + { + "epoch": 53.18087318087318, + "grad_norm": 0.0003213598974980414, + "learning_rate": 0.08635704762673052, + "loss": 0.2813, + "num_input_tokens_seen": 19508936, + "step": 25580 + }, + { + "epoch": 53.19126819126819, + "grad_norm": 0.00029099799576215446, + "learning_rate": 0.08630371252886981, + "loss": 0.2554, + "num_input_tokens_seen": 19512712, + "step": 25585 + }, + { + "epoch": 53.2016632016632, + "grad_norm": 0.0002532747748773545, + "learning_rate": 0.08625038725377704, + "loss": 0.2625, + "num_input_tokens_seen": 19516616, + "step": 25590 + }, + { + "epoch": 53.21205821205821, + "grad_norm": 0.0002777600020635873, + "learning_rate": 0.08619707180967566, + "loss": 0.2094, + "num_input_tokens_seen": 19520424, + "step": 25595 + }, + { + "epoch": 53.222453222453225, + "grad_norm": 0.0005434795166365802, + "learning_rate": 0.08614376620478768, + "loss": 0.2538, + "num_input_tokens_seen": 19524232, + "step": 25600 + }, + { + "epoch": 53.222453222453225, + "eval_loss": 0.25216010212898254, + "eval_runtime": 13.3884, + "eval_samples_per_second": 63.936, + "eval_steps_per_second": 15.984, + "num_input_tokens_seen": 19524232, + "step": 25600 + }, + { + "epoch": 53.232848232848234, + "grad_norm": 0.00014258103328756988, + "learning_rate": 0.08609047044733344, + "loss": 0.2688, + "num_input_tokens_seen": 19527944, + "step": 25605 + }, + { + "epoch": 53.24324324324324, + "grad_norm": 0.0005566255422309041, + "learning_rate": 0.08603718454553168, + "loss": 0.262, + "num_input_tokens_seen": 19531816, + "step": 25610 + }, + { + "epoch": 53.25363825363825, + "grad_norm": 0.0003099758177995682, + "learning_rate": 0.08598390850759997, + "loss": 0.2911, + "num_input_tokens_seen": 19535752, + "step": 25615 + }, + { + "epoch": 53.264033264033266, + "grad_norm": 0.00035618332913145423, + "learning_rate": 0.08593064234175397, + "loss": 0.2627, + "num_input_tokens_seen": 19539624, + "step": 25620 + }, + { + "epoch": 53.274428274428274, + "grad_norm": 0.00017494460917077959, + "learning_rate": 0.08587738605620815, + "loss": 0.2611, + "num_input_tokens_seen": 19543496, + "step": 25625 + }, + { + "epoch": 53.28482328482328, + "grad_norm": 0.0003928478399757296, + "learning_rate": 0.08582413965917512, + "loss": 0.2593, + "num_input_tokens_seen": 19547208, + "step": 25630 + }, + { + "epoch": 53.2952182952183, + "grad_norm": 5.883342237211764e-05, + "learning_rate": 0.08577090315886628, + "loss": 0.2401, + "num_input_tokens_seen": 19550920, + "step": 25635 + }, + { + "epoch": 53.305613305613306, + "grad_norm": 0.0002586435293778777, + "learning_rate": 0.08571767656349136, + "loss": 0.2549, + "num_input_tokens_seen": 19554824, + "step": 25640 + }, + { + "epoch": 53.316008316008315, + "grad_norm": 0.0002915515215136111, + "learning_rate": 0.08566445988125847, + "loss": 0.2923, + "num_input_tokens_seen": 19558536, + "step": 25645 + }, + { + "epoch": 53.32640332640332, + "grad_norm": 0.00010289955389453098, + "learning_rate": 0.08561125312037436, + "loss": 0.2683, + "num_input_tokens_seen": 19562440, + "step": 25650 + }, + { + "epoch": 53.33679833679834, + "grad_norm": 7.264727901201695e-05, + "learning_rate": 0.08555805628904424, + "loss": 0.2342, + "num_input_tokens_seen": 19566088, + "step": 25655 + }, + { + "epoch": 53.34719334719335, + "grad_norm": 0.0003991942503489554, + "learning_rate": 0.08550486939547161, + "loss": 0.258, + "num_input_tokens_seen": 19569960, + "step": 25660 + }, + { + "epoch": 53.357588357588355, + "grad_norm": 0.0003789020993281156, + "learning_rate": 0.08545169244785869, + "loss": 0.2841, + "num_input_tokens_seen": 19573672, + "step": 25665 + }, + { + "epoch": 53.36798336798337, + "grad_norm": 0.00028057151939719915, + "learning_rate": 0.08539852545440589, + "loss": 0.263, + "num_input_tokens_seen": 19577576, + "step": 25670 + }, + { + "epoch": 53.37837837837838, + "grad_norm": 7.342134631471708e-05, + "learning_rate": 0.08534536842331235, + "loss": 0.2592, + "num_input_tokens_seen": 19581288, + "step": 25675 + }, + { + "epoch": 53.38877338877339, + "grad_norm": 0.00014633193495683372, + "learning_rate": 0.08529222136277545, + "loss": 0.2775, + "num_input_tokens_seen": 19585128, + "step": 25680 + }, + { + "epoch": 53.3991683991684, + "grad_norm": 0.00021866269526071846, + "learning_rate": 0.08523908428099125, + "loss": 0.2767, + "num_input_tokens_seen": 19588904, + "step": 25685 + }, + { + "epoch": 53.40956340956341, + "grad_norm": 9.841060091275722e-05, + "learning_rate": 0.08518595718615402, + "loss": 0.2679, + "num_input_tokens_seen": 19592776, + "step": 25690 + }, + { + "epoch": 53.41995841995842, + "grad_norm": 0.00016996328486129642, + "learning_rate": 0.08513284008645675, + "loss": 0.2626, + "num_input_tokens_seen": 19596616, + "step": 25695 + }, + { + "epoch": 53.43035343035343, + "grad_norm": 0.000500997994095087, + "learning_rate": 0.08507973299009065, + "loss": 0.2711, + "num_input_tokens_seen": 19600328, + "step": 25700 + }, + { + "epoch": 53.44074844074844, + "grad_norm": 0.0004233003710396588, + "learning_rate": 0.08502663590524563, + "loss": 0.2369, + "num_input_tokens_seen": 19604072, + "step": 25705 + }, + { + "epoch": 53.45114345114345, + "grad_norm": 0.00017534276412334293, + "learning_rate": 0.08497354884010981, + "loss": 0.2773, + "num_input_tokens_seen": 19607816, + "step": 25710 + }, + { + "epoch": 53.46153846153846, + "grad_norm": 0.0002948506153188646, + "learning_rate": 0.0849204718028699, + "loss": 0.2856, + "num_input_tokens_seen": 19611752, + "step": 25715 + }, + { + "epoch": 53.471933471933475, + "grad_norm": 0.0010303632589057088, + "learning_rate": 0.08486740480171118, + "loss": 0.2692, + "num_input_tokens_seen": 19615400, + "step": 25720 + }, + { + "epoch": 53.482328482328484, + "grad_norm": 0.0006806841702200472, + "learning_rate": 0.08481434784481706, + "loss": 0.2632, + "num_input_tokens_seen": 19619240, + "step": 25725 + }, + { + "epoch": 53.49272349272349, + "grad_norm": 0.0002270428667543456, + "learning_rate": 0.08476130094036968, + "loss": 0.2682, + "num_input_tokens_seen": 19622888, + "step": 25730 + }, + { + "epoch": 53.5031185031185, + "grad_norm": 0.00016551902808714658, + "learning_rate": 0.08470826409654961, + "loss": 0.2563, + "num_input_tokens_seen": 19626632, + "step": 25735 + }, + { + "epoch": 53.513513513513516, + "grad_norm": 0.0002010891039390117, + "learning_rate": 0.08465523732153564, + "loss": 0.2673, + "num_input_tokens_seen": 19630536, + "step": 25740 + }, + { + "epoch": 53.523908523908524, + "grad_norm": 0.0003778178652282804, + "learning_rate": 0.08460222062350532, + "loss": 0.2742, + "num_input_tokens_seen": 19634376, + "step": 25745 + }, + { + "epoch": 53.53430353430353, + "grad_norm": 0.0005589865031652153, + "learning_rate": 0.08454921401063442, + "loss": 0.2604, + "num_input_tokens_seen": 19638312, + "step": 25750 + }, + { + "epoch": 53.54469854469855, + "grad_norm": 0.0009035980910994112, + "learning_rate": 0.08449621749109716, + "loss": 0.2738, + "num_input_tokens_seen": 19642056, + "step": 25755 + }, + { + "epoch": 53.555093555093556, + "grad_norm": 0.0005097401444800198, + "learning_rate": 0.08444323107306641, + "loss": 0.2582, + "num_input_tokens_seen": 19645832, + "step": 25760 + }, + { + "epoch": 53.565488565488565, + "grad_norm": 0.00041103025432676077, + "learning_rate": 0.0843902547647132, + "loss": 0.2737, + "num_input_tokens_seen": 19649672, + "step": 25765 + }, + { + "epoch": 53.57588357588357, + "grad_norm": 0.00017348640540149063, + "learning_rate": 0.0843372885742072, + "loss": 0.2818, + "num_input_tokens_seen": 19653576, + "step": 25770 + }, + { + "epoch": 53.58627858627859, + "grad_norm": 0.00044966634595766664, + "learning_rate": 0.08428433250971652, + "loss": 0.2423, + "num_input_tokens_seen": 19657256, + "step": 25775 + }, + { + "epoch": 53.5966735966736, + "grad_norm": 0.0002476478985045105, + "learning_rate": 0.08423138657940757, + "loss": 0.275, + "num_input_tokens_seen": 19661224, + "step": 25780 + }, + { + "epoch": 53.607068607068605, + "grad_norm": 0.00032057089265435934, + "learning_rate": 0.08417845079144536, + "loss": 0.279, + "num_input_tokens_seen": 19664968, + "step": 25785 + }, + { + "epoch": 53.61746361746362, + "grad_norm": 0.0005667418008670211, + "learning_rate": 0.08412552515399314, + "loss": 0.2553, + "num_input_tokens_seen": 19668840, + "step": 25790 + }, + { + "epoch": 53.62785862785863, + "grad_norm": 0.00021251654834486544, + "learning_rate": 0.08407260967521278, + "loss": 0.2203, + "num_input_tokens_seen": 19672616, + "step": 25795 + }, + { + "epoch": 53.63825363825364, + "grad_norm": 7.873615686548874e-05, + "learning_rate": 0.08401970436326454, + "loss": 0.2282, + "num_input_tokens_seen": 19676456, + "step": 25800 + }, + { + "epoch": 53.63825363825364, + "eval_loss": 0.24929235875606537, + "eval_runtime": 13.389, + "eval_samples_per_second": 63.933, + "eval_steps_per_second": 15.983, + "num_input_tokens_seen": 19676456, + "step": 25800 + }, + { + "epoch": 53.648648648648646, + "grad_norm": 4.673874718719162e-05, + "learning_rate": 0.08396680922630702, + "loss": 0.2573, + "num_input_tokens_seen": 19680200, + "step": 25805 + }, + { + "epoch": 53.65904365904366, + "grad_norm": 0.00033500452991575, + "learning_rate": 0.08391392427249732, + "loss": 0.2324, + "num_input_tokens_seen": 19683944, + "step": 25810 + }, + { + "epoch": 53.66943866943867, + "grad_norm": 0.00012405934103298932, + "learning_rate": 0.08386104950999107, + "loss": 0.2596, + "num_input_tokens_seen": 19687720, + "step": 25815 + }, + { + "epoch": 53.67983367983368, + "grad_norm": 0.00015227448602672666, + "learning_rate": 0.0838081849469421, + "loss": 0.2871, + "num_input_tokens_seen": 19691560, + "step": 25820 + }, + { + "epoch": 53.69022869022869, + "grad_norm": 0.0002926836023107171, + "learning_rate": 0.08375533059150281, + "loss": 0.2692, + "num_input_tokens_seen": 19695272, + "step": 25825 + }, + { + "epoch": 53.7006237006237, + "grad_norm": 0.00046705154818482697, + "learning_rate": 0.08370248645182406, + "loss": 0.2599, + "num_input_tokens_seen": 19699176, + "step": 25830 + }, + { + "epoch": 53.71101871101871, + "grad_norm": 0.0003347960882820189, + "learning_rate": 0.083649652536055, + "loss": 0.2571, + "num_input_tokens_seen": 19702824, + "step": 25835 + }, + { + "epoch": 53.72141372141372, + "grad_norm": 0.00027674148441292346, + "learning_rate": 0.08359682885234339, + "loss": 0.2671, + "num_input_tokens_seen": 19706728, + "step": 25840 + }, + { + "epoch": 53.731808731808734, + "grad_norm": 0.0005954855587333441, + "learning_rate": 0.08354401540883516, + "loss": 0.2739, + "num_input_tokens_seen": 19710568, + "step": 25845 + }, + { + "epoch": 53.74220374220374, + "grad_norm": 9.156563464784995e-05, + "learning_rate": 0.0834912122136749, + "loss": 0.2569, + "num_input_tokens_seen": 19714440, + "step": 25850 + }, + { + "epoch": 53.75259875259875, + "grad_norm": 6.419842247851193e-05, + "learning_rate": 0.0834384192750056, + "loss": 0.2856, + "num_input_tokens_seen": 19718312, + "step": 25855 + }, + { + "epoch": 53.762993762993766, + "grad_norm": 0.000195767599507235, + "learning_rate": 0.08338563660096844, + "loss": 0.2562, + "num_input_tokens_seen": 19721992, + "step": 25860 + }, + { + "epoch": 53.773388773388774, + "grad_norm": 0.0002074228395940736, + "learning_rate": 0.08333286419970329, + "loss": 0.2688, + "num_input_tokens_seen": 19725800, + "step": 25865 + }, + { + "epoch": 53.78378378378378, + "grad_norm": 0.0003609775740187615, + "learning_rate": 0.08328010207934824, + "loss": 0.2678, + "num_input_tokens_seen": 19729672, + "step": 25870 + }, + { + "epoch": 53.79417879417879, + "grad_norm": 0.00026167414034716785, + "learning_rate": 0.08322735024803989, + "loss": 0.281, + "num_input_tokens_seen": 19733480, + "step": 25875 + }, + { + "epoch": 53.804573804573806, + "grad_norm": 0.00011447145516285673, + "learning_rate": 0.08317460871391331, + "loss": 0.2796, + "num_input_tokens_seen": 19737320, + "step": 25880 + }, + { + "epoch": 53.814968814968815, + "grad_norm": 0.00035589313483797014, + "learning_rate": 0.08312187748510179, + "loss": 0.2662, + "num_input_tokens_seen": 19741160, + "step": 25885 + }, + { + "epoch": 53.82536382536382, + "grad_norm": 0.0003252648457419127, + "learning_rate": 0.08306915656973726, + "loss": 0.2683, + "num_input_tokens_seen": 19745032, + "step": 25890 + }, + { + "epoch": 53.83575883575884, + "grad_norm": 0.0002471132611390203, + "learning_rate": 0.08301644597594988, + "loss": 0.2544, + "num_input_tokens_seen": 19748776, + "step": 25895 + }, + { + "epoch": 53.84615384615385, + "grad_norm": 0.00019129975407849997, + "learning_rate": 0.08296374571186826, + "loss": 0.2667, + "num_input_tokens_seen": 19752648, + "step": 25900 + }, + { + "epoch": 53.856548856548855, + "grad_norm": 7.480049680452794e-05, + "learning_rate": 0.08291105578561955, + "loss": 0.2548, + "num_input_tokens_seen": 19756488, + "step": 25905 + }, + { + "epoch": 53.86694386694387, + "grad_norm": 0.0005411587189882994, + "learning_rate": 0.08285837620532904, + "loss": 0.2989, + "num_input_tokens_seen": 19760232, + "step": 25910 + }, + { + "epoch": 53.87733887733888, + "grad_norm": 0.0002652540570124984, + "learning_rate": 0.0828057069791207, + "loss": 0.2668, + "num_input_tokens_seen": 19763944, + "step": 25915 + }, + { + "epoch": 53.88773388773389, + "grad_norm": 0.0003816888201981783, + "learning_rate": 0.0827530481151168, + "loss": 0.2575, + "num_input_tokens_seen": 19767848, + "step": 25920 + }, + { + "epoch": 53.898128898128896, + "grad_norm": 4.4268763303989545e-05, + "learning_rate": 0.08270039962143792, + "loss": 0.2658, + "num_input_tokens_seen": 19771688, + "step": 25925 + }, + { + "epoch": 53.90852390852391, + "grad_norm": 0.00022834762057755142, + "learning_rate": 0.08264776150620314, + "loss": 0.2639, + "num_input_tokens_seen": 19775368, + "step": 25930 + }, + { + "epoch": 53.91891891891892, + "grad_norm": 0.00016352102102246135, + "learning_rate": 0.08259513377753, + "loss": 0.2876, + "num_input_tokens_seen": 19779080, + "step": 25935 + }, + { + "epoch": 53.92931392931393, + "grad_norm": 7.525904948124662e-05, + "learning_rate": 0.08254251644353423, + "loss": 0.2647, + "num_input_tokens_seen": 19782888, + "step": 25940 + }, + { + "epoch": 53.93970893970894, + "grad_norm": 0.0002536469546612352, + "learning_rate": 0.08248990951233022, + "loss": 0.2605, + "num_input_tokens_seen": 19786760, + "step": 25945 + }, + { + "epoch": 53.95010395010395, + "grad_norm": 0.0002702054916881025, + "learning_rate": 0.08243731299203048, + "loss": 0.2754, + "num_input_tokens_seen": 19790568, + "step": 25950 + }, + { + "epoch": 53.96049896049896, + "grad_norm": 0.00038639953709207475, + "learning_rate": 0.08238472689074612, + "loss": 0.2575, + "num_input_tokens_seen": 19794344, + "step": 25955 + }, + { + "epoch": 53.97089397089397, + "grad_norm": 9.961640171241015e-05, + "learning_rate": 0.08233215121658666, + "loss": 0.2543, + "num_input_tokens_seen": 19798248, + "step": 25960 + }, + { + "epoch": 53.981288981288984, + "grad_norm": 0.00011329437984386459, + "learning_rate": 0.08227958597765982, + "loss": 0.2546, + "num_input_tokens_seen": 19802184, + "step": 25965 + }, + { + "epoch": 53.99168399168399, + "grad_norm": 0.0001849041727837175, + "learning_rate": 0.08222703118207181, + "loss": 0.2763, + "num_input_tokens_seen": 19805960, + "step": 25970 + }, + { + "epoch": 54.002079002079, + "grad_norm": 6.556616426678374e-05, + "learning_rate": 0.08217448683792734, + "loss": 0.2778, + "num_input_tokens_seen": 19809560, + "step": 25975 + }, + { + "epoch": 54.012474012474016, + "grad_norm": 0.00011560157872736454, + "learning_rate": 0.08212195295332926, + "loss": 0.2652, + "num_input_tokens_seen": 19813400, + "step": 25980 + }, + { + "epoch": 54.022869022869024, + "grad_norm": 0.0004670706985052675, + "learning_rate": 0.08206942953637915, + "loss": 0.2491, + "num_input_tokens_seen": 19817304, + "step": 25985 + }, + { + "epoch": 54.03326403326403, + "grad_norm": 0.0004275247920304537, + "learning_rate": 0.08201691659517658, + "loss": 0.2453, + "num_input_tokens_seen": 19821208, + "step": 25990 + }, + { + "epoch": 54.04365904365904, + "grad_norm": 0.00040047045331448317, + "learning_rate": 0.08196441413781981, + "loss": 0.2542, + "num_input_tokens_seen": 19824760, + "step": 25995 + }, + { + "epoch": 54.054054054054056, + "grad_norm": 0.0002530265774112195, + "learning_rate": 0.08191192217240544, + "loss": 0.2705, + "num_input_tokens_seen": 19828504, + "step": 26000 + }, + { + "epoch": 54.054054054054056, + "eval_loss": 0.24692478775978088, + "eval_runtime": 13.3784, + "eval_samples_per_second": 63.984, + "eval_steps_per_second": 15.996, + "num_input_tokens_seen": 19828504, + "step": 26000 + }, + { + "epoch": 54.064449064449065, + "grad_norm": 0.0002602216845843941, + "learning_rate": 0.08185944070702823, + "loss": 0.2624, + "num_input_tokens_seen": 19832376, + "step": 26005 + }, + { + "epoch": 54.07484407484407, + "grad_norm": 8.60114159877412e-05, + "learning_rate": 0.08180696974978159, + "loss": 0.2496, + "num_input_tokens_seen": 19836184, + "step": 26010 + }, + { + "epoch": 54.08523908523909, + "grad_norm": 0.00011810174328275025, + "learning_rate": 0.08175450930875724, + "loss": 0.2486, + "num_input_tokens_seen": 19839992, + "step": 26015 + }, + { + "epoch": 54.0956340956341, + "grad_norm": 0.00021255902538541704, + "learning_rate": 0.08170205939204513, + "loss": 0.2459, + "num_input_tokens_seen": 19843736, + "step": 26020 + }, + { + "epoch": 54.106029106029105, + "grad_norm": 0.0003656491171568632, + "learning_rate": 0.08164962000773379, + "loss": 0.2664, + "num_input_tokens_seen": 19847640, + "step": 26025 + }, + { + "epoch": 54.11642411642411, + "grad_norm": 6.77032585372217e-05, + "learning_rate": 0.08159719116390995, + "loss": 0.2793, + "num_input_tokens_seen": 19851416, + "step": 26030 + }, + { + "epoch": 54.12681912681913, + "grad_norm": 0.0002658611338119954, + "learning_rate": 0.08154477286865887, + "loss": 0.2725, + "num_input_tokens_seen": 19855320, + "step": 26035 + }, + { + "epoch": 54.13721413721414, + "grad_norm": 4.501608054852113e-05, + "learning_rate": 0.08149236513006404, + "loss": 0.2573, + "num_input_tokens_seen": 19859224, + "step": 26040 + }, + { + "epoch": 54.147609147609145, + "grad_norm": 0.00012734455231111497, + "learning_rate": 0.08143996795620746, + "loss": 0.2729, + "num_input_tokens_seen": 19863096, + "step": 26045 + }, + { + "epoch": 54.15800415800416, + "grad_norm": 0.00014130730414763093, + "learning_rate": 0.08138758135516938, + "loss": 0.2712, + "num_input_tokens_seen": 19866776, + "step": 26050 + }, + { + "epoch": 54.16839916839917, + "grad_norm": 0.00011177197302458808, + "learning_rate": 0.08133520533502851, + "loss": 0.27, + "num_input_tokens_seen": 19870584, + "step": 26055 + }, + { + "epoch": 54.17879417879418, + "grad_norm": 0.00037128428812138736, + "learning_rate": 0.08128283990386184, + "loss": 0.2837, + "num_input_tokens_seen": 19874616, + "step": 26060 + }, + { + "epoch": 54.189189189189186, + "grad_norm": 0.0002411420427961275, + "learning_rate": 0.08123048506974488, + "loss": 0.2656, + "num_input_tokens_seen": 19878488, + "step": 26065 + }, + { + "epoch": 54.1995841995842, + "grad_norm": 5.852436152053997e-05, + "learning_rate": 0.08117814084075124, + "loss": 0.26, + "num_input_tokens_seen": 19882232, + "step": 26070 + }, + { + "epoch": 54.20997920997921, + "grad_norm": 0.0003427154733799398, + "learning_rate": 0.08112580722495318, + "loss": 0.272, + "num_input_tokens_seen": 19885944, + "step": 26075 + }, + { + "epoch": 54.22037422037422, + "grad_norm": 0.00025025615468621254, + "learning_rate": 0.08107348423042122, + "loss": 0.2559, + "num_input_tokens_seen": 19889752, + "step": 26080 + }, + { + "epoch": 54.23076923076923, + "grad_norm": 0.00014690423267893493, + "learning_rate": 0.08102117186522413, + "loss": 0.2683, + "num_input_tokens_seen": 19893560, + "step": 26085 + }, + { + "epoch": 54.24116424116424, + "grad_norm": 0.0005931174382567406, + "learning_rate": 0.08096887013742916, + "loss": 0.2614, + "num_input_tokens_seen": 19897432, + "step": 26090 + }, + { + "epoch": 54.25155925155925, + "grad_norm": 7.394458225462586e-05, + "learning_rate": 0.08091657905510198, + "loss": 0.252, + "num_input_tokens_seen": 19901080, + "step": 26095 + }, + { + "epoch": 54.26195426195426, + "grad_norm": 0.0002644641208462417, + "learning_rate": 0.08086429862630642, + "loss": 0.2758, + "num_input_tokens_seen": 19904760, + "step": 26100 + }, + { + "epoch": 54.272349272349274, + "grad_norm": 0.00016311694344040006, + "learning_rate": 0.08081202885910488, + "loss": 0.2758, + "num_input_tokens_seen": 19908408, + "step": 26105 + }, + { + "epoch": 54.28274428274428, + "grad_norm": 0.00023390664136968553, + "learning_rate": 0.08075976976155795, + "loss": 0.2643, + "num_input_tokens_seen": 19912376, + "step": 26110 + }, + { + "epoch": 54.29313929313929, + "grad_norm": 0.00015235489991027862, + "learning_rate": 0.08070752134172461, + "loss": 0.249, + "num_input_tokens_seen": 19916344, + "step": 26115 + }, + { + "epoch": 54.303534303534306, + "grad_norm": 0.00010802914039231837, + "learning_rate": 0.08065528360766229, + "loss": 0.2594, + "num_input_tokens_seen": 19920088, + "step": 26120 + }, + { + "epoch": 54.313929313929314, + "grad_norm": 0.00020670176309067756, + "learning_rate": 0.08060305656742664, + "loss": 0.2683, + "num_input_tokens_seen": 19923832, + "step": 26125 + }, + { + "epoch": 54.32432432432432, + "grad_norm": 0.00014302424096968025, + "learning_rate": 0.08055084022907182, + "loss": 0.2989, + "num_input_tokens_seen": 19927544, + "step": 26130 + }, + { + "epoch": 54.33471933471934, + "grad_norm": 0.00019747289479710162, + "learning_rate": 0.08049863460065014, + "loss": 0.2665, + "num_input_tokens_seen": 19931480, + "step": 26135 + }, + { + "epoch": 54.34511434511435, + "grad_norm": 0.0002862780529540032, + "learning_rate": 0.0804464396902124, + "loss": 0.2799, + "num_input_tokens_seen": 19935288, + "step": 26140 + }, + { + "epoch": 54.355509355509355, + "grad_norm": 0.00042460119584575295, + "learning_rate": 0.08039425550580777, + "loss": 0.2616, + "num_input_tokens_seen": 19939128, + "step": 26145 + }, + { + "epoch": 54.36590436590436, + "grad_norm": 0.00024318067880813032, + "learning_rate": 0.08034208205548363, + "loss": 0.2836, + "num_input_tokens_seen": 19943000, + "step": 26150 + }, + { + "epoch": 54.37629937629938, + "grad_norm": 0.00029690409428440034, + "learning_rate": 0.08028991934728581, + "loss": 0.2662, + "num_input_tokens_seen": 19946904, + "step": 26155 + }, + { + "epoch": 54.38669438669439, + "grad_norm": 0.0005558753618970513, + "learning_rate": 0.0802377673892585, + "loss": 0.2582, + "num_input_tokens_seen": 19950712, + "step": 26160 + }, + { + "epoch": 54.397089397089395, + "grad_norm": 0.0004594610072672367, + "learning_rate": 0.0801856261894441, + "loss": 0.2712, + "num_input_tokens_seen": 19954456, + "step": 26165 + }, + { + "epoch": 54.40748440748441, + "grad_norm": 0.00047014711890369654, + "learning_rate": 0.08013349575588354, + "loss": 0.2929, + "num_input_tokens_seen": 19958296, + "step": 26170 + }, + { + "epoch": 54.41787941787942, + "grad_norm": 7.162798283388838e-05, + "learning_rate": 0.08008137609661586, + "loss": 0.2642, + "num_input_tokens_seen": 19961976, + "step": 26175 + }, + { + "epoch": 54.42827442827443, + "grad_norm": 6.433594535337761e-05, + "learning_rate": 0.08002926721967872, + "loss": 0.2598, + "num_input_tokens_seen": 19965720, + "step": 26180 + }, + { + "epoch": 54.438669438669436, + "grad_norm": 0.00031317290267907083, + "learning_rate": 0.07997716913310782, + "loss": 0.2704, + "num_input_tokens_seen": 19969560, + "step": 26185 + }, + { + "epoch": 54.44906444906445, + "grad_norm": 0.00010351963283028454, + "learning_rate": 0.07992508184493745, + "loss": 0.2582, + "num_input_tokens_seen": 19973336, + "step": 26190 + }, + { + "epoch": 54.45945945945946, + "grad_norm": 8.793236338533461e-05, + "learning_rate": 0.07987300536320001, + "loss": 0.2564, + "num_input_tokens_seen": 19977144, + "step": 26195 + }, + { + "epoch": 54.46985446985447, + "grad_norm": 8.558998524677008e-05, + "learning_rate": 0.07982093969592649, + "loss": 0.2621, + "num_input_tokens_seen": 19980856, + "step": 26200 + }, + { + "epoch": 54.46985446985447, + "eval_loss": 0.24724189937114716, + "eval_runtime": 13.3945, + "eval_samples_per_second": 63.907, + "eval_steps_per_second": 15.977, + "num_input_tokens_seen": 19980856, + "step": 26200 + }, + { + "epoch": 54.48024948024948, + "grad_norm": 0.00011386102414689958, + "learning_rate": 0.07976888485114592, + "loss": 0.2514, + "num_input_tokens_seen": 19984696, + "step": 26205 + }, + { + "epoch": 54.49064449064449, + "grad_norm": 0.0002468022285029292, + "learning_rate": 0.07971684083688595, + "loss": 0.2814, + "num_input_tokens_seen": 19988376, + "step": 26210 + }, + { + "epoch": 54.5010395010395, + "grad_norm": 0.00031022605253383517, + "learning_rate": 0.0796648076611723, + "loss": 0.2716, + "num_input_tokens_seen": 19992152, + "step": 26215 + }, + { + "epoch": 54.51143451143451, + "grad_norm": 0.00019124281243421137, + "learning_rate": 0.07961278533202922, + "loss": 0.2889, + "num_input_tokens_seen": 19996152, + "step": 26220 + }, + { + "epoch": 54.521829521829524, + "grad_norm": 0.00041830085683614016, + "learning_rate": 0.07956077385747919, + "loss": 0.2683, + "num_input_tokens_seen": 19999960, + "step": 26225 + }, + { + "epoch": 54.53222453222453, + "grad_norm": 0.00018451534560881555, + "learning_rate": 0.079508773245543, + "loss": 0.2732, + "num_input_tokens_seen": 20003640, + "step": 26230 + }, + { + "epoch": 54.54261954261954, + "grad_norm": 0.000328708701999858, + "learning_rate": 0.07945678350423982, + "loss": 0.233, + "num_input_tokens_seen": 20007320, + "step": 26235 + }, + { + "epoch": 54.553014553014556, + "grad_norm": 0.0002678821620065719, + "learning_rate": 0.07940480464158717, + "loss": 0.251, + "num_input_tokens_seen": 20011128, + "step": 26240 + }, + { + "epoch": 54.563409563409564, + "grad_norm": 0.00021444098092615604, + "learning_rate": 0.07935283666560076, + "loss": 0.2235, + "num_input_tokens_seen": 20014872, + "step": 26245 + }, + { + "epoch": 54.57380457380457, + "grad_norm": 0.00020633151871152222, + "learning_rate": 0.07930087958429478, + "loss": 0.3135, + "num_input_tokens_seen": 20018680, + "step": 26250 + }, + { + "epoch": 54.58419958419958, + "grad_norm": 0.0004138337098993361, + "learning_rate": 0.07924893340568159, + "loss": 0.2635, + "num_input_tokens_seen": 20022584, + "step": 26255 + }, + { + "epoch": 54.5945945945946, + "grad_norm": 0.00011721346527338028, + "learning_rate": 0.07919699813777205, + "loss": 0.2345, + "num_input_tokens_seen": 20026328, + "step": 26260 + }, + { + "epoch": 54.604989604989605, + "grad_norm": 0.00021726837439928204, + "learning_rate": 0.07914507378857515, + "loss": 0.2712, + "num_input_tokens_seen": 20030040, + "step": 26265 + }, + { + "epoch": 54.61538461538461, + "grad_norm": 0.00032105721766129136, + "learning_rate": 0.07909316036609822, + "loss": 0.2868, + "num_input_tokens_seen": 20033880, + "step": 26270 + }, + { + "epoch": 54.62577962577963, + "grad_norm": 0.00022261105186771601, + "learning_rate": 0.07904125787834704, + "loss": 0.2515, + "num_input_tokens_seen": 20037752, + "step": 26275 + }, + { + "epoch": 54.63617463617464, + "grad_norm": 0.0006170336273498833, + "learning_rate": 0.07898936633332569, + "loss": 0.2522, + "num_input_tokens_seen": 20041560, + "step": 26280 + }, + { + "epoch": 54.646569646569645, + "grad_norm": 0.00020674700499512255, + "learning_rate": 0.07893748573903635, + "loss": 0.2542, + "num_input_tokens_seen": 20045368, + "step": 26285 + }, + { + "epoch": 54.656964656964654, + "grad_norm": 0.00014733245188836008, + "learning_rate": 0.0788856161034798, + "loss": 0.2637, + "num_input_tokens_seen": 20049144, + "step": 26290 + }, + { + "epoch": 54.66735966735967, + "grad_norm": 0.00030363118276000023, + "learning_rate": 0.07883375743465487, + "loss": 0.2671, + "num_input_tokens_seen": 20052920, + "step": 26295 + }, + { + "epoch": 54.67775467775468, + "grad_norm": 0.0002180361479986459, + "learning_rate": 0.07878190974055888, + "loss": 0.2612, + "num_input_tokens_seen": 20056760, + "step": 26300 + }, + { + "epoch": 54.688149688149686, + "grad_norm": 0.0007007686654105783, + "learning_rate": 0.07873007302918746, + "loss": 0.2899, + "num_input_tokens_seen": 20060792, + "step": 26305 + }, + { + "epoch": 54.6985446985447, + "grad_norm": 0.00017218109860550612, + "learning_rate": 0.07867824730853433, + "loss": 0.2701, + "num_input_tokens_seen": 20064408, + "step": 26310 + }, + { + "epoch": 54.70893970893971, + "grad_norm": 0.00036265526432543993, + "learning_rate": 0.07862643258659176, + "loss": 0.2664, + "num_input_tokens_seen": 20068120, + "step": 26315 + }, + { + "epoch": 54.71933471933472, + "grad_norm": 0.0004432791320141405, + "learning_rate": 0.07857462887135026, + "loss": 0.2659, + "num_input_tokens_seen": 20072184, + "step": 26320 + }, + { + "epoch": 54.729729729729726, + "grad_norm": 0.0004398549208417535, + "learning_rate": 0.0785228361707986, + "loss": 0.2566, + "num_input_tokens_seen": 20076152, + "step": 26325 + }, + { + "epoch": 54.74012474012474, + "grad_norm": 0.00013934337766841054, + "learning_rate": 0.07847105449292378, + "loss": 0.2593, + "num_input_tokens_seen": 20080152, + "step": 26330 + }, + { + "epoch": 54.75051975051975, + "grad_norm": 0.00025779425050131977, + "learning_rate": 0.0784192838457113, + "loss": 0.2786, + "num_input_tokens_seen": 20083928, + "step": 26335 + }, + { + "epoch": 54.76091476091476, + "grad_norm": 0.00022268503380473703, + "learning_rate": 0.07836752423714473, + "loss": 0.2575, + "num_input_tokens_seen": 20087736, + "step": 26340 + }, + { + "epoch": 54.771309771309774, + "grad_norm": 0.0004874719015788287, + "learning_rate": 0.07831577567520616, + "loss": 0.2684, + "num_input_tokens_seen": 20091672, + "step": 26345 + }, + { + "epoch": 54.78170478170478, + "grad_norm": 0.0002922238491009921, + "learning_rate": 0.07826403816787579, + "loss": 0.2836, + "num_input_tokens_seen": 20095576, + "step": 26350 + }, + { + "epoch": 54.79209979209979, + "grad_norm": 0.00010126736742677167, + "learning_rate": 0.0782123117231322, + "loss": 0.2569, + "num_input_tokens_seen": 20099384, + "step": 26355 + }, + { + "epoch": 54.802494802494806, + "grad_norm": 0.0008457657531835139, + "learning_rate": 0.07816059634895237, + "loss": 0.2683, + "num_input_tokens_seen": 20103096, + "step": 26360 + }, + { + "epoch": 54.812889812889814, + "grad_norm": 0.00020537260570563376, + "learning_rate": 0.0781088920533113, + "loss": 0.2576, + "num_input_tokens_seen": 20106840, + "step": 26365 + }, + { + "epoch": 54.82328482328482, + "grad_norm": 0.00010219080286333337, + "learning_rate": 0.07805719884418257, + "loss": 0.2576, + "num_input_tokens_seen": 20110648, + "step": 26370 + }, + { + "epoch": 54.83367983367983, + "grad_norm": 0.00036513133090920746, + "learning_rate": 0.07800551672953779, + "loss": 0.2562, + "num_input_tokens_seen": 20114520, + "step": 26375 + }, + { + "epoch": 54.84407484407485, + "grad_norm": 0.00025904891663230956, + "learning_rate": 0.07795384571734709, + "loss": 0.2615, + "num_input_tokens_seen": 20118264, + "step": 26380 + }, + { + "epoch": 54.854469854469855, + "grad_norm": 0.0003789609472732991, + "learning_rate": 0.07790218581557883, + "loss": 0.2625, + "num_input_tokens_seen": 20122264, + "step": 26385 + }, + { + "epoch": 54.86486486486486, + "grad_norm": 0.00014257957809604704, + "learning_rate": 0.07785053703219949, + "loss": 0.2714, + "num_input_tokens_seen": 20126168, + "step": 26390 + }, + { + "epoch": 54.87525987525988, + "grad_norm": 0.00022269565670285374, + "learning_rate": 0.07779889937517409, + "loss": 0.2593, + "num_input_tokens_seen": 20130040, + "step": 26395 + }, + { + "epoch": 54.88565488565489, + "grad_norm": 0.00048069132026284933, + "learning_rate": 0.0777472728524657, + "loss": 0.263, + "num_input_tokens_seen": 20133784, + "step": 26400 + }, + { + "epoch": 54.88565488565489, + "eval_loss": 0.2537725865840912, + "eval_runtime": 13.3917, + "eval_samples_per_second": 63.92, + "eval_steps_per_second": 15.98, + "num_input_tokens_seen": 20133784, + "step": 26400 + }, + { + "epoch": 54.896049896049895, + "grad_norm": 0.00019268585310783237, + "learning_rate": 0.07769565747203584, + "loss": 0.2442, + "num_input_tokens_seen": 20137624, + "step": 26405 + }, + { + "epoch": 54.906444906444904, + "grad_norm": 0.0003926725185010582, + "learning_rate": 0.07764405324184427, + "loss": 0.2695, + "num_input_tokens_seen": 20141368, + "step": 26410 + }, + { + "epoch": 54.91683991683992, + "grad_norm": 0.0005147213232703507, + "learning_rate": 0.07759246016984889, + "loss": 0.2808, + "num_input_tokens_seen": 20145336, + "step": 26415 + }, + { + "epoch": 54.92723492723493, + "grad_norm": 0.00014873924374114722, + "learning_rate": 0.07754087826400609, + "loss": 0.278, + "num_input_tokens_seen": 20149304, + "step": 26420 + }, + { + "epoch": 54.937629937629936, + "grad_norm": 0.00027317943749949336, + "learning_rate": 0.0774893075322705, + "loss": 0.2538, + "num_input_tokens_seen": 20152888, + "step": 26425 + }, + { + "epoch": 54.94802494802495, + "grad_norm": 0.00012338340457063168, + "learning_rate": 0.07743774798259484, + "loss": 0.2552, + "num_input_tokens_seen": 20156568, + "step": 26430 + }, + { + "epoch": 54.95841995841996, + "grad_norm": 0.000618341495282948, + "learning_rate": 0.07738619962293032, + "loss": 0.264, + "num_input_tokens_seen": 20160376, + "step": 26435 + }, + { + "epoch": 54.96881496881497, + "grad_norm": 0.0002620588056743145, + "learning_rate": 0.0773346624612264, + "loss": 0.2899, + "num_input_tokens_seen": 20164248, + "step": 26440 + }, + { + "epoch": 54.979209979209976, + "grad_norm": 0.00026451688609085977, + "learning_rate": 0.07728313650543066, + "loss": 0.2411, + "num_input_tokens_seen": 20168088, + "step": 26445 + }, + { + "epoch": 54.98960498960499, + "grad_norm": 0.00017170059436466545, + "learning_rate": 0.07723162176348913, + "loss": 0.2803, + "num_input_tokens_seen": 20171864, + "step": 26450 + }, + { + "epoch": 55.0, + "grad_norm": 9.429289639228955e-05, + "learning_rate": 0.07718011824334593, + "loss": 0.269, + "num_input_tokens_seen": 20175464, + "step": 26455 + }, + { + "epoch": 55.01039501039501, + "grad_norm": 0.00012367004819680005, + "learning_rate": 0.07712862595294363, + "loss": 0.2558, + "num_input_tokens_seen": 20179496, + "step": 26460 + }, + { + "epoch": 55.020790020790024, + "grad_norm": 0.0006052608950994909, + "learning_rate": 0.07707714490022301, + "loss": 0.2508, + "num_input_tokens_seen": 20183272, + "step": 26465 + }, + { + "epoch": 55.03118503118503, + "grad_norm": 0.0002582585730124265, + "learning_rate": 0.07702567509312298, + "loss": 0.2645, + "num_input_tokens_seen": 20187112, + "step": 26470 + }, + { + "epoch": 55.04158004158004, + "grad_norm": 6.263565592234954e-05, + "learning_rate": 0.07697421653958098, + "loss": 0.2412, + "num_input_tokens_seen": 20190952, + "step": 26475 + }, + { + "epoch": 55.05197505197505, + "grad_norm": 0.00016367787611670792, + "learning_rate": 0.07692276924753247, + "loss": 0.2632, + "num_input_tokens_seen": 20194792, + "step": 26480 + }, + { + "epoch": 55.062370062370064, + "grad_norm": 0.0009348075254820287, + "learning_rate": 0.07687133322491124, + "loss": 0.269, + "num_input_tokens_seen": 20198600, + "step": 26485 + }, + { + "epoch": 55.07276507276507, + "grad_norm": 0.00048672521370463073, + "learning_rate": 0.07681990847964948, + "loss": 0.2336, + "num_input_tokens_seen": 20202312, + "step": 26490 + }, + { + "epoch": 55.08316008316008, + "grad_norm": 0.00015556183643639088, + "learning_rate": 0.0767684950196774, + "loss": 0.2398, + "num_input_tokens_seen": 20206088, + "step": 26495 + }, + { + "epoch": 55.093555093555096, + "grad_norm": 0.0006498659495264292, + "learning_rate": 0.0767170928529237, + "loss": 0.255, + "num_input_tokens_seen": 20210024, + "step": 26500 + }, + { + "epoch": 55.103950103950105, + "grad_norm": 0.000219001027289778, + "learning_rate": 0.07666570198731526, + "loss": 0.2469, + "num_input_tokens_seen": 20213704, + "step": 26505 + }, + { + "epoch": 55.11434511434511, + "grad_norm": 6.57796481391415e-05, + "learning_rate": 0.07661432243077708, + "loss": 0.2415, + "num_input_tokens_seen": 20217384, + "step": 26510 + }, + { + "epoch": 55.12474012474012, + "grad_norm": 5.6014323490671813e-05, + "learning_rate": 0.0765629541912326, + "loss": 0.2599, + "num_input_tokens_seen": 20221064, + "step": 26515 + }, + { + "epoch": 55.13513513513514, + "grad_norm": 6.651139847235754e-05, + "learning_rate": 0.07651159727660352, + "loss": 0.2588, + "num_input_tokens_seen": 20224712, + "step": 26520 + }, + { + "epoch": 55.145530145530145, + "grad_norm": 0.0002601959276944399, + "learning_rate": 0.07646025169480959, + "loss": 0.2716, + "num_input_tokens_seen": 20228488, + "step": 26525 + }, + { + "epoch": 55.15592515592515, + "grad_norm": 0.00014527332677971572, + "learning_rate": 0.07640891745376908, + "loss": 0.2596, + "num_input_tokens_seen": 20232456, + "step": 26530 + }, + { + "epoch": 55.16632016632017, + "grad_norm": 0.0006158179603517056, + "learning_rate": 0.07635759456139822, + "loss": 0.2517, + "num_input_tokens_seen": 20236232, + "step": 26535 + }, + { + "epoch": 55.17671517671518, + "grad_norm": 8.398152567679062e-05, + "learning_rate": 0.0763062830256118, + "loss": 0.2555, + "num_input_tokens_seen": 20240040, + "step": 26540 + }, + { + "epoch": 55.187110187110186, + "grad_norm": 0.00020210351794958115, + "learning_rate": 0.07625498285432258, + "loss": 0.289, + "num_input_tokens_seen": 20243944, + "step": 26545 + }, + { + "epoch": 55.197505197505194, + "grad_norm": 0.0005895509384572506, + "learning_rate": 0.07620369405544176, + "loss": 0.2764, + "num_input_tokens_seen": 20248040, + "step": 26550 + }, + { + "epoch": 55.20790020790021, + "grad_norm": 0.0005683311610482633, + "learning_rate": 0.07615241663687868, + "loss": 0.2808, + "num_input_tokens_seen": 20251976, + "step": 26555 + }, + { + "epoch": 55.21829521829522, + "grad_norm": 9.496301208855584e-05, + "learning_rate": 0.07610115060654106, + "loss": 0.2644, + "num_input_tokens_seen": 20255912, + "step": 26560 + }, + { + "epoch": 55.228690228690226, + "grad_norm": 9.120270260609686e-05, + "learning_rate": 0.07604989597233458, + "loss": 0.252, + "num_input_tokens_seen": 20259688, + "step": 26565 + }, + { + "epoch": 55.23908523908524, + "grad_norm": 0.0005651332903653383, + "learning_rate": 0.07599865274216352, + "loss": 0.2415, + "num_input_tokens_seen": 20263400, + "step": 26570 + }, + { + "epoch": 55.24948024948025, + "grad_norm": 0.0010400756727904081, + "learning_rate": 0.07594742092393013, + "loss": 0.3186, + "num_input_tokens_seen": 20267080, + "step": 26575 + }, + { + "epoch": 55.25987525987526, + "grad_norm": 0.0003462302847765386, + "learning_rate": 0.07589620052553503, + "loss": 0.2503, + "num_input_tokens_seen": 20270760, + "step": 26580 + }, + { + "epoch": 55.270270270270274, + "grad_norm": 0.00023988127941265702, + "learning_rate": 0.0758449915548771, + "loss": 0.2829, + "num_input_tokens_seen": 20274472, + "step": 26585 + }, + { + "epoch": 55.28066528066528, + "grad_norm": 0.00013170663441997021, + "learning_rate": 0.07579379401985332, + "loss": 0.2803, + "num_input_tokens_seen": 20278344, + "step": 26590 + }, + { + "epoch": 55.29106029106029, + "grad_norm": 0.00028670215397141874, + "learning_rate": 0.07574260792835905, + "loss": 0.2687, + "num_input_tokens_seen": 20282344, + "step": 26595 + }, + { + "epoch": 55.3014553014553, + "grad_norm": 5.830689406138845e-05, + "learning_rate": 0.07569143328828784, + "loss": 0.273, + "num_input_tokens_seen": 20286120, + "step": 26600 + }, + { + "epoch": 55.3014553014553, + "eval_loss": 0.2565113306045532, + "eval_runtime": 13.3937, + "eval_samples_per_second": 63.91, + "eval_steps_per_second": 15.978, + "num_input_tokens_seen": 20286120, + "step": 26600 + }, + { + "epoch": 55.311850311850314, + "grad_norm": 4.851643825531937e-05, + "learning_rate": 0.0756402701075314, + "loss": 0.2673, + "num_input_tokens_seen": 20289992, + "step": 26605 + }, + { + "epoch": 55.32224532224532, + "grad_norm": 0.0002664014755282551, + "learning_rate": 0.07558911839397982, + "loss": 0.2862, + "num_input_tokens_seen": 20293800, + "step": 26610 + }, + { + "epoch": 55.33264033264033, + "grad_norm": 0.0001437256287317723, + "learning_rate": 0.07553797815552123, + "loss": 0.2744, + "num_input_tokens_seen": 20297608, + "step": 26615 + }, + { + "epoch": 55.343035343035346, + "grad_norm": 0.00025565660325810313, + "learning_rate": 0.07548684940004222, + "loss": 0.2762, + "num_input_tokens_seen": 20301416, + "step": 26620 + }, + { + "epoch": 55.353430353430355, + "grad_norm": 0.0001771823881426826, + "learning_rate": 0.07543573213542744, + "loss": 0.262, + "num_input_tokens_seen": 20305128, + "step": 26625 + }, + { + "epoch": 55.36382536382536, + "grad_norm": 0.00010525508696446195, + "learning_rate": 0.0753846263695597, + "loss": 0.275, + "num_input_tokens_seen": 20309000, + "step": 26630 + }, + { + "epoch": 55.37422037422037, + "grad_norm": 0.0006201022188179195, + "learning_rate": 0.07533353211032029, + "loss": 0.2559, + "num_input_tokens_seen": 20312776, + "step": 26635 + }, + { + "epoch": 55.38461538461539, + "grad_norm": 7.772214303258806e-05, + "learning_rate": 0.07528244936558857, + "loss": 0.2662, + "num_input_tokens_seen": 20316552, + "step": 26640 + }, + { + "epoch": 55.395010395010395, + "grad_norm": 0.0005495261866599321, + "learning_rate": 0.07523137814324206, + "loss": 0.2699, + "num_input_tokens_seen": 20320296, + "step": 26645 + }, + { + "epoch": 55.4054054054054, + "grad_norm": 7.798885781085119e-05, + "learning_rate": 0.07518031845115672, + "loss": 0.2989, + "num_input_tokens_seen": 20324232, + "step": 26650 + }, + { + "epoch": 55.41580041580042, + "grad_norm": 8.477033406961709e-05, + "learning_rate": 0.07512927029720647, + "loss": 0.2782, + "num_input_tokens_seen": 20327912, + "step": 26655 + }, + { + "epoch": 55.42619542619543, + "grad_norm": 0.0002552078804001212, + "learning_rate": 0.0750782336892636, + "loss": 0.274, + "num_input_tokens_seen": 20331688, + "step": 26660 + }, + { + "epoch": 55.436590436590436, + "grad_norm": 0.0004291512304916978, + "learning_rate": 0.0750272086351987, + "loss": 0.2653, + "num_input_tokens_seen": 20335432, + "step": 26665 + }, + { + "epoch": 55.446985446985444, + "grad_norm": 0.000267739495029673, + "learning_rate": 0.07497619514288031, + "loss": 0.2799, + "num_input_tokens_seen": 20339112, + "step": 26670 + }, + { + "epoch": 55.45738045738046, + "grad_norm": 0.00014905829448252916, + "learning_rate": 0.07492519322017545, + "loss": 0.2807, + "num_input_tokens_seen": 20342920, + "step": 26675 + }, + { + "epoch": 55.46777546777547, + "grad_norm": 0.00010727380868047476, + "learning_rate": 0.0748742028749493, + "loss": 0.2651, + "num_input_tokens_seen": 20346920, + "step": 26680 + }, + { + "epoch": 55.478170478170476, + "grad_norm": 0.00040123899816535413, + "learning_rate": 0.0748232241150651, + "loss": 0.2419, + "num_input_tokens_seen": 20350728, + "step": 26685 + }, + { + "epoch": 55.48856548856549, + "grad_norm": 0.00024495512479916215, + "learning_rate": 0.07477225694838453, + "loss": 0.2704, + "num_input_tokens_seen": 20354600, + "step": 26690 + }, + { + "epoch": 55.4989604989605, + "grad_norm": 0.0003472850366961211, + "learning_rate": 0.07472130138276731, + "loss": 0.261, + "num_input_tokens_seen": 20358536, + "step": 26695 + }, + { + "epoch": 55.50935550935551, + "grad_norm": 6.669552385574207e-05, + "learning_rate": 0.07467035742607138, + "loss": 0.293, + "num_input_tokens_seen": 20362376, + "step": 26700 + }, + { + "epoch": 55.51975051975052, + "grad_norm": 0.00023361884814221412, + "learning_rate": 0.07461942508615303, + "loss": 0.2582, + "num_input_tokens_seen": 20366280, + "step": 26705 + }, + { + "epoch": 55.53014553014553, + "grad_norm": 0.00038345353095792234, + "learning_rate": 0.07456850437086657, + "loss": 0.2798, + "num_input_tokens_seen": 20370184, + "step": 26710 + }, + { + "epoch": 55.54054054054054, + "grad_norm": 0.0008303012582473457, + "learning_rate": 0.07451759528806468, + "loss": 0.2631, + "num_input_tokens_seen": 20373960, + "step": 26715 + }, + { + "epoch": 55.55093555093555, + "grad_norm": 0.00032848777482286096, + "learning_rate": 0.0744666978455982, + "loss": 0.2606, + "num_input_tokens_seen": 20377800, + "step": 26720 + }, + { + "epoch": 55.561330561330564, + "grad_norm": 0.0004507634730543941, + "learning_rate": 0.07441581205131609, + "loss": 0.2739, + "num_input_tokens_seen": 20381608, + "step": 26725 + }, + { + "epoch": 55.57172557172557, + "grad_norm": 0.0003685134288389236, + "learning_rate": 0.07436493791306566, + "loss": 0.2809, + "num_input_tokens_seen": 20385352, + "step": 26730 + }, + { + "epoch": 55.58212058212058, + "grad_norm": 6.376814417308196e-05, + "learning_rate": 0.07431407543869223, + "loss": 0.2255, + "num_input_tokens_seen": 20389224, + "step": 26735 + }, + { + "epoch": 55.59251559251559, + "grad_norm": 0.0004548740980681032, + "learning_rate": 0.0742632246360395, + "loss": 0.2675, + "num_input_tokens_seen": 20393000, + "step": 26740 + }, + { + "epoch": 55.602910602910605, + "grad_norm": 0.00047511718003079295, + "learning_rate": 0.07421238551294934, + "loss": 0.2092, + "num_input_tokens_seen": 20396904, + "step": 26745 + }, + { + "epoch": 55.61330561330561, + "grad_norm": 0.00021994025155436248, + "learning_rate": 0.07416155807726171, + "loss": 0.2737, + "num_input_tokens_seen": 20400712, + "step": 26750 + }, + { + "epoch": 55.62370062370062, + "grad_norm": 0.0002352344454266131, + "learning_rate": 0.07411074233681492, + "loss": 0.2523, + "num_input_tokens_seen": 20404648, + "step": 26755 + }, + { + "epoch": 55.63409563409564, + "grad_norm": 0.00011049176100641489, + "learning_rate": 0.07405993829944528, + "loss": 0.238, + "num_input_tokens_seen": 20408424, + "step": 26760 + }, + { + "epoch": 55.644490644490645, + "grad_norm": 0.0002080237027257681, + "learning_rate": 0.07400914597298755, + "loss": 0.2335, + "num_input_tokens_seen": 20412200, + "step": 26765 + }, + { + "epoch": 55.65488565488565, + "grad_norm": 0.00014888279838487506, + "learning_rate": 0.07395836536527445, + "loss": 0.2847, + "num_input_tokens_seen": 20416104, + "step": 26770 + }, + { + "epoch": 55.66528066528066, + "grad_norm": 8.534899825463071e-05, + "learning_rate": 0.07390759648413696, + "loss": 0.2651, + "num_input_tokens_seen": 20419912, + "step": 26775 + }, + { + "epoch": 55.67567567567568, + "grad_norm": 3.690117227961309e-05, + "learning_rate": 0.07385683933740435, + "loss": 0.2679, + "num_input_tokens_seen": 20423688, + "step": 26780 + }, + { + "epoch": 55.686070686070686, + "grad_norm": 0.00038660975405946374, + "learning_rate": 0.07380609393290402, + "loss": 0.2681, + "num_input_tokens_seen": 20427464, + "step": 26785 + }, + { + "epoch": 55.696465696465694, + "grad_norm": 0.00012417722609825432, + "learning_rate": 0.07375536027846147, + "loss": 0.274, + "num_input_tokens_seen": 20431272, + "step": 26790 + }, + { + "epoch": 55.70686070686071, + "grad_norm": 7.76619344833307e-05, + "learning_rate": 0.07370463838190057, + "loss": 0.2771, + "num_input_tokens_seen": 20435272, + "step": 26795 + }, + { + "epoch": 55.71725571725572, + "grad_norm": 0.00020915240747854114, + "learning_rate": 0.07365392825104317, + "loss": 0.2596, + "num_input_tokens_seen": 20439016, + "step": 26800 + }, + { + "epoch": 55.71725571725572, + "eval_loss": 0.2567959427833557, + "eval_runtime": 13.3934, + "eval_samples_per_second": 63.912, + "eval_steps_per_second": 15.978, + "num_input_tokens_seen": 20439016, + "step": 26800 + }, + { + "epoch": 55.727650727650726, + "grad_norm": 9.20339734875597e-05, + "learning_rate": 0.07360322989370945, + "loss": 0.2705, + "num_input_tokens_seen": 20442888, + "step": 26805 + }, + { + "epoch": 55.73804573804574, + "grad_norm": 0.00031852704705670476, + "learning_rate": 0.07355254331771781, + "loss": 0.2741, + "num_input_tokens_seen": 20446920, + "step": 26810 + }, + { + "epoch": 55.74844074844075, + "grad_norm": 0.00017445560661144555, + "learning_rate": 0.07350186853088461, + "loss": 0.2739, + "num_input_tokens_seen": 20450792, + "step": 26815 + }, + { + "epoch": 55.75883575883576, + "grad_norm": 0.00011392993474146351, + "learning_rate": 0.07345120554102462, + "loss": 0.2754, + "num_input_tokens_seen": 20454536, + "step": 26820 + }, + { + "epoch": 55.76923076923077, + "grad_norm": 0.0002019075327552855, + "learning_rate": 0.07340055435595079, + "loss": 0.2588, + "num_input_tokens_seen": 20458376, + "step": 26825 + }, + { + "epoch": 55.77962577962578, + "grad_norm": 0.0003578672476578504, + "learning_rate": 0.07334991498347401, + "loss": 0.271, + "num_input_tokens_seen": 20462184, + "step": 26830 + }, + { + "epoch": 55.79002079002079, + "grad_norm": 0.00017872618627734482, + "learning_rate": 0.07329928743140365, + "loss": 0.2413, + "num_input_tokens_seen": 20465992, + "step": 26835 + }, + { + "epoch": 55.8004158004158, + "grad_norm": 0.00015051898662932217, + "learning_rate": 0.07324867170754705, + "loss": 0.2568, + "num_input_tokens_seen": 20469928, + "step": 26840 + }, + { + "epoch": 55.810810810810814, + "grad_norm": 0.00038806095835752785, + "learning_rate": 0.07319806781970974, + "loss": 0.2182, + "num_input_tokens_seen": 20473576, + "step": 26845 + }, + { + "epoch": 55.82120582120582, + "grad_norm": 0.00018154567806050181, + "learning_rate": 0.07314747577569555, + "loss": 0.2844, + "num_input_tokens_seen": 20477288, + "step": 26850 + }, + { + "epoch": 55.83160083160083, + "grad_norm": 0.00016651175974402577, + "learning_rate": 0.07309689558330636, + "loss": 0.2558, + "num_input_tokens_seen": 20481160, + "step": 26855 + }, + { + "epoch": 55.84199584199584, + "grad_norm": 0.0004189663741271943, + "learning_rate": 0.0730463272503423, + "loss": 0.2799, + "num_input_tokens_seen": 20484968, + "step": 26860 + }, + { + "epoch": 55.852390852390855, + "grad_norm": 2.620306440803688e-05, + "learning_rate": 0.07299577078460168, + "loss": 0.2827, + "num_input_tokens_seen": 20488616, + "step": 26865 + }, + { + "epoch": 55.86278586278586, + "grad_norm": 0.0001188592505059205, + "learning_rate": 0.07294522619388083, + "loss": 0.2655, + "num_input_tokens_seen": 20492328, + "step": 26870 + }, + { + "epoch": 55.87318087318087, + "grad_norm": 5.945091470493935e-05, + "learning_rate": 0.07289469348597452, + "loss": 0.2696, + "num_input_tokens_seen": 20496104, + "step": 26875 + }, + { + "epoch": 55.88357588357589, + "grad_norm": 0.0005189462681300938, + "learning_rate": 0.07284417266867535, + "loss": 0.2742, + "num_input_tokens_seen": 20499912, + "step": 26880 + }, + { + "epoch": 55.893970893970895, + "grad_norm": 0.0003031924716196954, + "learning_rate": 0.07279366374977439, + "loss": 0.2684, + "num_input_tokens_seen": 20503656, + "step": 26885 + }, + { + "epoch": 55.9043659043659, + "grad_norm": 2.6656425688997842e-05, + "learning_rate": 0.07274316673706074, + "loss": 0.2665, + "num_input_tokens_seen": 20507400, + "step": 26890 + }, + { + "epoch": 55.91476091476091, + "grad_norm": 0.0003867803025059402, + "learning_rate": 0.07269268163832161, + "loss": 0.2593, + "num_input_tokens_seen": 20511272, + "step": 26895 + }, + { + "epoch": 55.92515592515593, + "grad_norm": 0.00018587293743621558, + "learning_rate": 0.07264220846134248, + "loss": 0.2763, + "num_input_tokens_seen": 20515176, + "step": 26900 + }, + { + "epoch": 55.935550935550935, + "grad_norm": 5.33304555574432e-05, + "learning_rate": 0.07259174721390699, + "loss": 0.2449, + "num_input_tokens_seen": 20518760, + "step": 26905 + }, + { + "epoch": 55.945945945945944, + "grad_norm": 5.91612551943399e-05, + "learning_rate": 0.07254129790379686, + "loss": 0.253, + "num_input_tokens_seen": 20522536, + "step": 26910 + }, + { + "epoch": 55.95634095634096, + "grad_norm": 0.0001117538777180016, + "learning_rate": 0.072490860538792, + "loss": 0.2807, + "num_input_tokens_seen": 20526344, + "step": 26915 + }, + { + "epoch": 55.96673596673597, + "grad_norm": 0.00021780496172141284, + "learning_rate": 0.07244043512667042, + "loss": 0.2514, + "num_input_tokens_seen": 20530312, + "step": 26920 + }, + { + "epoch": 55.977130977130976, + "grad_norm": 0.00018675463797990233, + "learning_rate": 0.07239002167520843, + "loss": 0.2696, + "num_input_tokens_seen": 20534088, + "step": 26925 + }, + { + "epoch": 55.987525987525984, + "grad_norm": 0.00045130591024644673, + "learning_rate": 0.07233962019218045, + "loss": 0.2777, + "num_input_tokens_seen": 20537896, + "step": 26930 + }, + { + "epoch": 55.997920997921, + "grad_norm": 4.9080379540100694e-05, + "learning_rate": 0.07228923068535892, + "loss": 0.2503, + "num_input_tokens_seen": 20541704, + "step": 26935 + }, + { + "epoch": 56.00831600831601, + "grad_norm": 0.00019629186135716736, + "learning_rate": 0.0722388531625146, + "loss": 0.2684, + "num_input_tokens_seen": 20545368, + "step": 26940 + }, + { + "epoch": 56.018711018711016, + "grad_norm": 0.00039805221604183316, + "learning_rate": 0.07218848763141639, + "loss": 0.2751, + "num_input_tokens_seen": 20549272, + "step": 26945 + }, + { + "epoch": 56.02910602910603, + "grad_norm": 0.0009759733220562339, + "learning_rate": 0.07213813409983118, + "loss": 0.2769, + "num_input_tokens_seen": 20552920, + "step": 26950 + }, + { + "epoch": 56.03950103950104, + "grad_norm": 0.00021742662647739053, + "learning_rate": 0.0720877925755242, + "loss": 0.2701, + "num_input_tokens_seen": 20556760, + "step": 26955 + }, + { + "epoch": 56.04989604989605, + "grad_norm": 0.00035511149326339364, + "learning_rate": 0.07203746306625866, + "loss": 0.2614, + "num_input_tokens_seen": 20560600, + "step": 26960 + }, + { + "epoch": 56.06029106029106, + "grad_norm": 3.565665247151628e-05, + "learning_rate": 0.07198714557979606, + "loss": 0.2676, + "num_input_tokens_seen": 20564344, + "step": 26965 + }, + { + "epoch": 56.07068607068607, + "grad_norm": 0.0006292808102443814, + "learning_rate": 0.07193684012389602, + "loss": 0.2265, + "num_input_tokens_seen": 20568312, + "step": 26970 + }, + { + "epoch": 56.08108108108108, + "grad_norm": 7.516855839639902e-05, + "learning_rate": 0.07188654670631621, + "loss": 0.2383, + "num_input_tokens_seen": 20572088, + "step": 26975 + }, + { + "epoch": 56.09147609147609, + "grad_norm": 0.0002188281068811193, + "learning_rate": 0.07183626533481258, + "loss": 0.2951, + "num_input_tokens_seen": 20575864, + "step": 26980 + }, + { + "epoch": 56.101871101871104, + "grad_norm": 0.00010026013478636742, + "learning_rate": 0.07178599601713909, + "loss": 0.229, + "num_input_tokens_seen": 20579704, + "step": 26985 + }, + { + "epoch": 56.11226611226611, + "grad_norm": 0.0007259357953444123, + "learning_rate": 0.07173573876104786, + "loss": 0.2956, + "num_input_tokens_seen": 20583608, + "step": 26990 + }, + { + "epoch": 56.12266112266112, + "grad_norm": 0.00037151671131141484, + "learning_rate": 0.0716854935742893, + "loss": 0.2671, + "num_input_tokens_seen": 20587480, + "step": 26995 + }, + { + "epoch": 56.13305613305613, + "grad_norm": 0.0004270995850674808, + "learning_rate": 0.07163526046461174, + "loss": 0.2786, + "num_input_tokens_seen": 20591320, + "step": 27000 + }, + { + "epoch": 56.13305613305613, + "eval_loss": 0.2581345736980438, + "eval_runtime": 13.3745, + "eval_samples_per_second": 64.002, + "eval_steps_per_second": 16.001, + "num_input_tokens_seen": 20591320, + "step": 27000 + }, + { + "epoch": 56.143451143451145, + "grad_norm": 0.0002719515177886933, + "learning_rate": 0.07158503943976181, + "loss": 0.2738, + "num_input_tokens_seen": 20594904, + "step": 27005 + }, + { + "epoch": 56.15384615384615, + "grad_norm": 0.00025673562777228653, + "learning_rate": 0.07153483050748427, + "loss": 0.2695, + "num_input_tokens_seen": 20598680, + "step": 27010 + }, + { + "epoch": 56.16424116424116, + "grad_norm": 0.0001781985629349947, + "learning_rate": 0.07148463367552188, + "loss": 0.2598, + "num_input_tokens_seen": 20602456, + "step": 27015 + }, + { + "epoch": 56.17463617463618, + "grad_norm": 0.0003269309236202389, + "learning_rate": 0.07143444895161565, + "loss": 0.2473, + "num_input_tokens_seen": 20606104, + "step": 27020 + }, + { + "epoch": 56.185031185031185, + "grad_norm": 0.0001319317234447226, + "learning_rate": 0.07138427634350476, + "loss": 0.2513, + "num_input_tokens_seen": 20609848, + "step": 27025 + }, + { + "epoch": 56.195426195426194, + "grad_norm": 5.980447167530656e-05, + "learning_rate": 0.07133411585892636, + "loss": 0.2628, + "num_input_tokens_seen": 20613816, + "step": 27030 + }, + { + "epoch": 56.20582120582121, + "grad_norm": 0.0001459171762689948, + "learning_rate": 0.07128396750561593, + "loss": 0.3004, + "num_input_tokens_seen": 20617720, + "step": 27035 + }, + { + "epoch": 56.21621621621622, + "grad_norm": 0.00025547415134496987, + "learning_rate": 0.07123383129130685, + "loss": 0.2705, + "num_input_tokens_seen": 20621432, + "step": 27040 + }, + { + "epoch": 56.226611226611226, + "grad_norm": 0.0007055096793919802, + "learning_rate": 0.07118370722373084, + "loss": 0.2568, + "num_input_tokens_seen": 20625208, + "step": 27045 + }, + { + "epoch": 56.237006237006234, + "grad_norm": 0.0001665985182626173, + "learning_rate": 0.07113359531061769, + "loss": 0.2587, + "num_input_tokens_seen": 20629016, + "step": 27050 + }, + { + "epoch": 56.24740124740125, + "grad_norm": 0.00030473803053610027, + "learning_rate": 0.07108349555969525, + "loss": 0.2744, + "num_input_tokens_seen": 20632920, + "step": 27055 + }, + { + "epoch": 56.25779625779626, + "grad_norm": 9.962715557776392e-05, + "learning_rate": 0.07103340797868944, + "loss": 0.2923, + "num_input_tokens_seen": 20636728, + "step": 27060 + }, + { + "epoch": 56.268191268191266, + "grad_norm": 0.00044138816883787513, + "learning_rate": 0.07098333257532453, + "loss": 0.2605, + "num_input_tokens_seen": 20640568, + "step": 27065 + }, + { + "epoch": 56.27858627858628, + "grad_norm": 0.00015792126941960305, + "learning_rate": 0.07093326935732269, + "loss": 0.2613, + "num_input_tokens_seen": 20644376, + "step": 27070 + }, + { + "epoch": 56.28898128898129, + "grad_norm": 0.00021895718236919492, + "learning_rate": 0.0708832183324044, + "loss": 0.2763, + "num_input_tokens_seen": 20648184, + "step": 27075 + }, + { + "epoch": 56.2993762993763, + "grad_norm": 0.0003913456457667053, + "learning_rate": 0.07083317950828799, + "loss": 0.2329, + "num_input_tokens_seen": 20651960, + "step": 27080 + }, + { + "epoch": 56.30977130977131, + "grad_norm": 0.000294084515189752, + "learning_rate": 0.0707831528926902, + "loss": 0.2805, + "num_input_tokens_seen": 20655768, + "step": 27085 + }, + { + "epoch": 56.32016632016632, + "grad_norm": 0.00021721870871260762, + "learning_rate": 0.07073313849332578, + "loss": 0.2741, + "num_input_tokens_seen": 20659544, + "step": 27090 + }, + { + "epoch": 56.33056133056133, + "grad_norm": 0.00021275108156260103, + "learning_rate": 0.07068313631790749, + "loss": 0.235, + "num_input_tokens_seen": 20663352, + "step": 27095 + }, + { + "epoch": 56.34095634095634, + "grad_norm": 6.678060890408233e-05, + "learning_rate": 0.07063314637414632, + "loss": 0.2458, + "num_input_tokens_seen": 20667160, + "step": 27100 + }, + { + "epoch": 56.351351351351354, + "grad_norm": 8.47870614961721e-05, + "learning_rate": 0.07058316866975144, + "loss": 0.2511, + "num_input_tokens_seen": 20670968, + "step": 27105 + }, + { + "epoch": 56.36174636174636, + "grad_norm": 0.00012517163122538477, + "learning_rate": 0.0705332032124299, + "loss": 0.2483, + "num_input_tokens_seen": 20674872, + "step": 27110 + }, + { + "epoch": 56.37214137214137, + "grad_norm": 0.00028426633798517287, + "learning_rate": 0.0704832500098871, + "loss": 0.2776, + "num_input_tokens_seen": 20678808, + "step": 27115 + }, + { + "epoch": 56.38253638253638, + "grad_norm": 8.44882961246185e-05, + "learning_rate": 0.07043330906982641, + "loss": 0.2561, + "num_input_tokens_seen": 20682776, + "step": 27120 + }, + { + "epoch": 56.392931392931395, + "grad_norm": 0.0004336985875852406, + "learning_rate": 0.07038338039994936, + "loss": 0.2751, + "num_input_tokens_seen": 20686552, + "step": 27125 + }, + { + "epoch": 56.4033264033264, + "grad_norm": 0.00011908068699995056, + "learning_rate": 0.07033346400795562, + "loss": 0.2467, + "num_input_tokens_seen": 20690296, + "step": 27130 + }, + { + "epoch": 56.41372141372141, + "grad_norm": 0.000276941922493279, + "learning_rate": 0.07028355990154282, + "loss": 0.2794, + "num_input_tokens_seen": 20694040, + "step": 27135 + }, + { + "epoch": 56.42411642411643, + "grad_norm": 0.00017509180179331452, + "learning_rate": 0.07023366808840685, + "loss": 0.2596, + "num_input_tokens_seen": 20697816, + "step": 27140 + }, + { + "epoch": 56.434511434511435, + "grad_norm": 0.00017745922377798706, + "learning_rate": 0.07018378857624172, + "loss": 0.2707, + "num_input_tokens_seen": 20701624, + "step": 27145 + }, + { + "epoch": 56.444906444906444, + "grad_norm": 0.0006116931908763945, + "learning_rate": 0.0701339213727394, + "loss": 0.2452, + "num_input_tokens_seen": 20705560, + "step": 27150 + }, + { + "epoch": 56.45530145530145, + "grad_norm": 0.00029468053253367543, + "learning_rate": 0.07008406648559008, + "loss": 0.2412, + "num_input_tokens_seen": 20709272, + "step": 27155 + }, + { + "epoch": 56.46569646569647, + "grad_norm": 0.0005130237550474703, + "learning_rate": 0.07003422392248196, + "loss": 0.2479, + "num_input_tokens_seen": 20713048, + "step": 27160 + }, + { + "epoch": 56.476091476091476, + "grad_norm": 0.0003463863395154476, + "learning_rate": 0.06998439369110142, + "loss": 0.2645, + "num_input_tokens_seen": 20717016, + "step": 27165 + }, + { + "epoch": 56.486486486486484, + "grad_norm": 0.0006893663667142391, + "learning_rate": 0.06993457579913295, + "loss": 0.3108, + "num_input_tokens_seen": 20720824, + "step": 27170 + }, + { + "epoch": 56.4968814968815, + "grad_norm": 0.0006446539191529155, + "learning_rate": 0.06988477025425903, + "loss": 0.2827, + "num_input_tokens_seen": 20724664, + "step": 27175 + }, + { + "epoch": 56.50727650727651, + "grad_norm": 0.00012618370237760246, + "learning_rate": 0.06983497706416032, + "loss": 0.2698, + "num_input_tokens_seen": 20728408, + "step": 27180 + }, + { + "epoch": 56.517671517671516, + "grad_norm": 0.00029114712378941476, + "learning_rate": 0.0697851962365156, + "loss": 0.2761, + "num_input_tokens_seen": 20732184, + "step": 27185 + }, + { + "epoch": 56.528066528066525, + "grad_norm": 0.00043628213461488485, + "learning_rate": 0.06973542777900163, + "loss": 0.271, + "num_input_tokens_seen": 20735896, + "step": 27190 + }, + { + "epoch": 56.53846153846154, + "grad_norm": 6.305332499323413e-05, + "learning_rate": 0.06968567169929342, + "loss": 0.2541, + "num_input_tokens_seen": 20739992, + "step": 27195 + }, + { + "epoch": 56.54885654885655, + "grad_norm": 0.00034865402267314494, + "learning_rate": 0.06963592800506392, + "loss": 0.2734, + "num_input_tokens_seen": 20743736, + "step": 27200 + }, + { + "epoch": 56.54885654885655, + "eval_loss": 0.24752461910247803, + "eval_runtime": 13.3921, + "eval_samples_per_second": 63.918, + "eval_steps_per_second": 15.98, + "num_input_tokens_seen": 20743736, + "step": 27200 + }, + { + "epoch": 56.55925155925156, + "grad_norm": 5.927007441641763e-05, + "learning_rate": 0.06958619670398417, + "loss": 0.2702, + "num_input_tokens_seen": 20747480, + "step": 27205 + }, + { + "epoch": 56.56964656964657, + "grad_norm": 0.00034109526313841343, + "learning_rate": 0.0695364778037235, + "loss": 0.2529, + "num_input_tokens_seen": 20751288, + "step": 27210 + }, + { + "epoch": 56.58004158004158, + "grad_norm": 0.0003387674514669925, + "learning_rate": 0.06948677131194907, + "loss": 0.2726, + "num_input_tokens_seen": 20755032, + "step": 27215 + }, + { + "epoch": 56.59043659043659, + "grad_norm": 0.0001852315617725253, + "learning_rate": 0.06943707723632629, + "loss": 0.2594, + "num_input_tokens_seen": 20758776, + "step": 27220 + }, + { + "epoch": 56.6008316008316, + "grad_norm": 6.984312494751066e-05, + "learning_rate": 0.06938739558451867, + "loss": 0.2551, + "num_input_tokens_seen": 20762776, + "step": 27225 + }, + { + "epoch": 56.61122661122661, + "grad_norm": 0.00020217348355799913, + "learning_rate": 0.06933772636418763, + "loss": 0.2554, + "num_input_tokens_seen": 20766520, + "step": 27230 + }, + { + "epoch": 56.62162162162162, + "grad_norm": 4.147472282056697e-05, + "learning_rate": 0.06928806958299293, + "loss": 0.2791, + "num_input_tokens_seen": 20770520, + "step": 27235 + }, + { + "epoch": 56.63201663201663, + "grad_norm": 0.0004289264907129109, + "learning_rate": 0.06923842524859211, + "loss": 0.2416, + "num_input_tokens_seen": 20774552, + "step": 27240 + }, + { + "epoch": 56.642411642411645, + "grad_norm": 0.0005177919520065188, + "learning_rate": 0.06918879336864105, + "loss": 0.2557, + "num_input_tokens_seen": 20778424, + "step": 27245 + }, + { + "epoch": 56.65280665280665, + "grad_norm": 0.0002976930409204215, + "learning_rate": 0.06913917395079362, + "loss": 0.2837, + "num_input_tokens_seen": 20782360, + "step": 27250 + }, + { + "epoch": 56.66320166320166, + "grad_norm": 8.254634303739294e-05, + "learning_rate": 0.0690895670027017, + "loss": 0.2559, + "num_input_tokens_seen": 20786328, + "step": 27255 + }, + { + "epoch": 56.67359667359668, + "grad_norm": 0.00041633605724200606, + "learning_rate": 0.06903997253201531, + "loss": 0.2689, + "num_input_tokens_seen": 20790008, + "step": 27260 + }, + { + "epoch": 56.683991683991685, + "grad_norm": 0.00038434506859630346, + "learning_rate": 0.06899039054638263, + "loss": 0.2771, + "num_input_tokens_seen": 20793752, + "step": 27265 + }, + { + "epoch": 56.694386694386694, + "grad_norm": 3.790081973420456e-05, + "learning_rate": 0.06894082105344976, + "loss": 0.2694, + "num_input_tokens_seen": 20797464, + "step": 27270 + }, + { + "epoch": 56.7047817047817, + "grad_norm": 0.00043999950867146254, + "learning_rate": 0.06889126406086087, + "loss": 0.2674, + "num_input_tokens_seen": 20801368, + "step": 27275 + }, + { + "epoch": 56.71517671517672, + "grad_norm": 0.0001274142268812284, + "learning_rate": 0.0688417195762584, + "loss": 0.2643, + "num_input_tokens_seen": 20805144, + "step": 27280 + }, + { + "epoch": 56.725571725571726, + "grad_norm": 0.00030875412630848587, + "learning_rate": 0.06879218760728262, + "loss": 0.2626, + "num_input_tokens_seen": 20808856, + "step": 27285 + }, + { + "epoch": 56.735966735966734, + "grad_norm": 0.00036022934364154935, + "learning_rate": 0.06874266816157207, + "loss": 0.2586, + "num_input_tokens_seen": 20812728, + "step": 27290 + }, + { + "epoch": 56.74636174636175, + "grad_norm": 0.00023577686806675047, + "learning_rate": 0.06869316124676321, + "loss": 0.2913, + "num_input_tokens_seen": 20816536, + "step": 27295 + }, + { + "epoch": 56.75675675675676, + "grad_norm": 0.00047496502520516515, + "learning_rate": 0.06864366687049062, + "loss": 0.2761, + "num_input_tokens_seen": 20820408, + "step": 27300 + }, + { + "epoch": 56.767151767151766, + "grad_norm": 5.42862580914516e-05, + "learning_rate": 0.06859418504038704, + "loss": 0.2577, + "num_input_tokens_seen": 20824216, + "step": 27305 + }, + { + "epoch": 56.777546777546775, + "grad_norm": 7.574332994408906e-05, + "learning_rate": 0.06854471576408311, + "loss": 0.2802, + "num_input_tokens_seen": 20827992, + "step": 27310 + }, + { + "epoch": 56.78794178794179, + "grad_norm": 0.0003212700830772519, + "learning_rate": 0.06849525904920767, + "loss": 0.268, + "num_input_tokens_seen": 20831832, + "step": 27315 + }, + { + "epoch": 56.7983367983368, + "grad_norm": 0.00012190680718049407, + "learning_rate": 0.06844581490338748, + "loss": 0.2547, + "num_input_tokens_seen": 20835672, + "step": 27320 + }, + { + "epoch": 56.80873180873181, + "grad_norm": 0.00013367677456699312, + "learning_rate": 0.06839638333424752, + "loss": 0.284, + "num_input_tokens_seen": 20839352, + "step": 27325 + }, + { + "epoch": 56.81912681912682, + "grad_norm": 0.0003036781563423574, + "learning_rate": 0.06834696434941082, + "loss": 0.2783, + "num_input_tokens_seen": 20843224, + "step": 27330 + }, + { + "epoch": 56.82952182952183, + "grad_norm": 0.00011597020784392953, + "learning_rate": 0.06829755795649824, + "loss": 0.2645, + "num_input_tokens_seen": 20847000, + "step": 27335 + }, + { + "epoch": 56.83991683991684, + "grad_norm": 6.796064553782344e-05, + "learning_rate": 0.06824816416312904, + "loss": 0.2656, + "num_input_tokens_seen": 20850872, + "step": 27340 + }, + { + "epoch": 56.85031185031185, + "grad_norm": 0.0002623731561470777, + "learning_rate": 0.06819878297692027, + "loss": 0.265, + "num_input_tokens_seen": 20854680, + "step": 27345 + }, + { + "epoch": 56.86070686070686, + "grad_norm": 0.00013267601025290787, + "learning_rate": 0.0681494144054871, + "loss": 0.2727, + "num_input_tokens_seen": 20858552, + "step": 27350 + }, + { + "epoch": 56.87110187110187, + "grad_norm": 0.0003577958850655705, + "learning_rate": 0.06810005845644286, + "loss": 0.2602, + "num_input_tokens_seen": 20862424, + "step": 27355 + }, + { + "epoch": 56.88149688149688, + "grad_norm": 0.00036278337938711047, + "learning_rate": 0.06805071513739878, + "loss": 0.271, + "num_input_tokens_seen": 20866200, + "step": 27360 + }, + { + "epoch": 56.891891891891895, + "grad_norm": 0.0002547994372434914, + "learning_rate": 0.06800138445596428, + "loss": 0.2598, + "num_input_tokens_seen": 20869944, + "step": 27365 + }, + { + "epoch": 56.9022869022869, + "grad_norm": 0.00022513128351420164, + "learning_rate": 0.06795206641974678, + "loss": 0.2625, + "num_input_tokens_seen": 20873752, + "step": 27370 + }, + { + "epoch": 56.91268191268191, + "grad_norm": 0.0002479232207406312, + "learning_rate": 0.06790276103635169, + "loss": 0.2462, + "num_input_tokens_seen": 20877432, + "step": 27375 + }, + { + "epoch": 56.92307692307692, + "grad_norm": 3.6150941014057025e-05, + "learning_rate": 0.0678534683133826, + "loss": 0.2662, + "num_input_tokens_seen": 20881080, + "step": 27380 + }, + { + "epoch": 56.933471933471935, + "grad_norm": 0.00012071291712345555, + "learning_rate": 0.06780418825844095, + "loss": 0.2615, + "num_input_tokens_seen": 20884856, + "step": 27385 + }, + { + "epoch": 56.943866943866944, + "grad_norm": 9.35002535698004e-05, + "learning_rate": 0.0677549208791264, + "loss": 0.2819, + "num_input_tokens_seen": 20888632, + "step": 27390 + }, + { + "epoch": 56.95426195426195, + "grad_norm": 0.00014534560614265501, + "learning_rate": 0.06770566618303668, + "loss": 0.2742, + "num_input_tokens_seen": 20892440, + "step": 27395 + }, + { + "epoch": 56.96465696465697, + "grad_norm": 7.129237928893417e-05, + "learning_rate": 0.06765642417776736, + "loss": 0.2701, + "num_input_tokens_seen": 20896184, + "step": 27400 + }, + { + "epoch": 56.96465696465697, + "eval_loss": 0.25731244683265686, + "eval_runtime": 13.412, + "eval_samples_per_second": 63.823, + "eval_steps_per_second": 15.956, + "num_input_tokens_seen": 20896184, + "step": 27400 + }, + { + "epoch": 56.975051975051976, + "grad_norm": 0.00025478401221334934, + "learning_rate": 0.0676071948709122, + "loss": 0.2744, + "num_input_tokens_seen": 20900088, + "step": 27405 + }, + { + "epoch": 56.985446985446984, + "grad_norm": 0.00022322667064145207, + "learning_rate": 0.06755797827006307, + "loss": 0.2725, + "num_input_tokens_seen": 20903960, + "step": 27410 + }, + { + "epoch": 56.99584199584199, + "grad_norm": 0.0004477265465538949, + "learning_rate": 0.06750877438280974, + "loss": 0.2495, + "num_input_tokens_seen": 20907864, + "step": 27415 + }, + { + "epoch": 57.00623700623701, + "grad_norm": 0.0004648904432542622, + "learning_rate": 0.06745958321673998, + "loss": 0.2387, + "num_input_tokens_seen": 20911816, + "step": 27420 + }, + { + "epoch": 57.016632016632016, + "grad_norm": 7.16396389179863e-05, + "learning_rate": 0.0674104047794398, + "loss": 0.2751, + "num_input_tokens_seen": 20915720, + "step": 27425 + }, + { + "epoch": 57.027027027027025, + "grad_norm": 0.00017544305592309684, + "learning_rate": 0.06736123907849303, + "loss": 0.2738, + "num_input_tokens_seen": 20919560, + "step": 27430 + }, + { + "epoch": 57.03742203742204, + "grad_norm": 3.616348112700507e-05, + "learning_rate": 0.06731208612148178, + "loss": 0.2253, + "num_input_tokens_seen": 20923336, + "step": 27435 + }, + { + "epoch": 57.04781704781705, + "grad_norm": 0.0003717425570357591, + "learning_rate": 0.0672629459159859, + "loss": 0.2718, + "num_input_tokens_seen": 20927144, + "step": 27440 + }, + { + "epoch": 57.05821205821206, + "grad_norm": 0.0003411647630855441, + "learning_rate": 0.0672138184695835, + "loss": 0.2692, + "num_input_tokens_seen": 20930952, + "step": 27445 + }, + { + "epoch": 57.06860706860707, + "grad_norm": 0.00027951670926995575, + "learning_rate": 0.0671647037898507, + "loss": 0.2835, + "num_input_tokens_seen": 20934792, + "step": 27450 + }, + { + "epoch": 57.07900207900208, + "grad_norm": 0.00012088056973880157, + "learning_rate": 0.0671156018843615, + "loss": 0.2633, + "num_input_tokens_seen": 20938632, + "step": 27455 + }, + { + "epoch": 57.08939708939709, + "grad_norm": 9.906120976665989e-05, + "learning_rate": 0.06706651276068812, + "loss": 0.2692, + "num_input_tokens_seen": 20942568, + "step": 27460 + }, + { + "epoch": 57.0997920997921, + "grad_norm": 0.00019783749303314835, + "learning_rate": 0.06701743642640064, + "loss": 0.2818, + "num_input_tokens_seen": 20946472, + "step": 27465 + }, + { + "epoch": 57.11018711018711, + "grad_norm": 0.00023209824576042593, + "learning_rate": 0.06696837288906729, + "loss": 0.2723, + "num_input_tokens_seen": 20950376, + "step": 27470 + }, + { + "epoch": 57.12058212058212, + "grad_norm": 0.00015431437350343913, + "learning_rate": 0.06691932215625432, + "loss": 0.2725, + "num_input_tokens_seen": 20954088, + "step": 27475 + }, + { + "epoch": 57.13097713097713, + "grad_norm": 0.0004136514035053551, + "learning_rate": 0.06687028423552589, + "loss": 0.2527, + "num_input_tokens_seen": 20957864, + "step": 27480 + }, + { + "epoch": 57.141372141372145, + "grad_norm": 0.0006636630860157311, + "learning_rate": 0.06682125913444435, + "loss": 0.282, + "num_input_tokens_seen": 20961736, + "step": 27485 + }, + { + "epoch": 57.15176715176715, + "grad_norm": 0.0002690414839889854, + "learning_rate": 0.0667722468605699, + "loss": 0.2491, + "num_input_tokens_seen": 20965448, + "step": 27490 + }, + { + "epoch": 57.16216216216216, + "grad_norm": 4.071547300554812e-05, + "learning_rate": 0.06672324742146094, + "loss": 0.2728, + "num_input_tokens_seen": 20969256, + "step": 27495 + }, + { + "epoch": 57.17255717255717, + "grad_norm": 5.4155843827174976e-05, + "learning_rate": 0.06667426082467373, + "loss": 0.2726, + "num_input_tokens_seen": 20973192, + "step": 27500 + }, + { + "epoch": 57.182952182952185, + "grad_norm": 0.0002416012721369043, + "learning_rate": 0.0666252870777626, + "loss": 0.2666, + "num_input_tokens_seen": 20977000, + "step": 27505 + }, + { + "epoch": 57.19334719334719, + "grad_norm": 0.0002885781868826598, + "learning_rate": 0.06657632618827995, + "loss": 0.2618, + "num_input_tokens_seen": 20980744, + "step": 27510 + }, + { + "epoch": 57.2037422037422, + "grad_norm": 0.00011953869397984818, + "learning_rate": 0.06652737816377623, + "loss": 0.2595, + "num_input_tokens_seen": 20984424, + "step": 27515 + }, + { + "epoch": 57.21413721413722, + "grad_norm": 0.0004012670833617449, + "learning_rate": 0.06647844301179971, + "loss": 0.2654, + "num_input_tokens_seen": 20988136, + "step": 27520 + }, + { + "epoch": 57.224532224532226, + "grad_norm": 0.00010758600546978414, + "learning_rate": 0.06642952073989689, + "loss": 0.2352, + "num_input_tokens_seen": 20992040, + "step": 27525 + }, + { + "epoch": 57.234927234927234, + "grad_norm": 0.0001168713642982766, + "learning_rate": 0.06638061135561223, + "loss": 0.247, + "num_input_tokens_seen": 20995912, + "step": 27530 + }, + { + "epoch": 57.24532224532224, + "grad_norm": 0.00017923676932696253, + "learning_rate": 0.06633171486648808, + "loss": 0.2436, + "num_input_tokens_seen": 20999784, + "step": 27535 + }, + { + "epoch": 57.25571725571726, + "grad_norm": 0.00033305975375697017, + "learning_rate": 0.06628283128006499, + "loss": 0.2826, + "num_input_tokens_seen": 21003784, + "step": 27540 + }, + { + "epoch": 57.266112266112266, + "grad_norm": 0.000500043504871428, + "learning_rate": 0.0662339606038813, + "loss": 0.2811, + "num_input_tokens_seen": 21007400, + "step": 27545 + }, + { + "epoch": 57.276507276507274, + "grad_norm": 0.00020137093088123947, + "learning_rate": 0.06618510284547358, + "loss": 0.2465, + "num_input_tokens_seen": 21011176, + "step": 27550 + }, + { + "epoch": 57.28690228690229, + "grad_norm": 0.0001073284656740725, + "learning_rate": 0.06613625801237633, + "loss": 0.2623, + "num_input_tokens_seen": 21014952, + "step": 27555 + }, + { + "epoch": 57.2972972972973, + "grad_norm": 0.00028848633519373834, + "learning_rate": 0.066087426112122, + "loss": 0.2781, + "num_input_tokens_seen": 21018696, + "step": 27560 + }, + { + "epoch": 57.30769230769231, + "grad_norm": 0.00014227887731976807, + "learning_rate": 0.06603860715224101, + "loss": 0.2664, + "num_input_tokens_seen": 21022472, + "step": 27565 + }, + { + "epoch": 57.318087318087315, + "grad_norm": 0.0007072836742736399, + "learning_rate": 0.06598980114026198, + "loss": 0.2625, + "num_input_tokens_seen": 21026280, + "step": 27570 + }, + { + "epoch": 57.32848232848233, + "grad_norm": 0.00016485484957229346, + "learning_rate": 0.06594100808371128, + "loss": 0.2643, + "num_input_tokens_seen": 21030088, + "step": 27575 + }, + { + "epoch": 57.33887733887734, + "grad_norm": 0.00019880793115589768, + "learning_rate": 0.06589222799011357, + "loss": 0.2513, + "num_input_tokens_seen": 21033800, + "step": 27580 + }, + { + "epoch": 57.34927234927235, + "grad_norm": 0.00030493922531604767, + "learning_rate": 0.0658434608669912, + "loss": 0.2898, + "num_input_tokens_seen": 21037640, + "step": 27585 + }, + { + "epoch": 57.35966735966736, + "grad_norm": 3.667144483188167e-05, + "learning_rate": 0.06579470672186473, + "loss": 0.2761, + "num_input_tokens_seen": 21041512, + "step": 27590 + }, + { + "epoch": 57.37006237006237, + "grad_norm": 0.0004897796316072345, + "learning_rate": 0.06574596556225275, + "loss": 0.2603, + "num_input_tokens_seen": 21045384, + "step": 27595 + }, + { + "epoch": 57.38045738045738, + "grad_norm": 4.648097456083633e-05, + "learning_rate": 0.06569723739567161, + "loss": 0.2462, + "num_input_tokens_seen": 21049160, + "step": 27600 + }, + { + "epoch": 57.38045738045738, + "eval_loss": 0.2511451244354248, + "eval_runtime": 13.392, + "eval_samples_per_second": 63.919, + "eval_steps_per_second": 15.98, + "num_input_tokens_seen": 21049160, + "step": 27600 + }, + { + "epoch": 57.39085239085239, + "grad_norm": 0.0002854251943062991, + "learning_rate": 0.06564852222963588, + "loss": 0.2724, + "num_input_tokens_seen": 21052904, + "step": 27605 + }, + { + "epoch": 57.4012474012474, + "grad_norm": 0.00017001866945065558, + "learning_rate": 0.06559982007165813, + "loss": 0.2779, + "num_input_tokens_seen": 21056648, + "step": 27610 + }, + { + "epoch": 57.41164241164241, + "grad_norm": 0.00018727932183537632, + "learning_rate": 0.06555113092924868, + "loss": 0.2693, + "num_input_tokens_seen": 21060424, + "step": 27615 + }, + { + "epoch": 57.42203742203742, + "grad_norm": 0.00014927737356629223, + "learning_rate": 0.06550245480991615, + "loss": 0.2625, + "num_input_tokens_seen": 21064200, + "step": 27620 + }, + { + "epoch": 57.432432432432435, + "grad_norm": 0.00011720373731805012, + "learning_rate": 0.0654537917211669, + "loss": 0.2515, + "num_input_tokens_seen": 21067944, + "step": 27625 + }, + { + "epoch": 57.44282744282744, + "grad_norm": 0.00030888401670381427, + "learning_rate": 0.0654051416705055, + "loss": 0.2688, + "num_input_tokens_seen": 21071944, + "step": 27630 + }, + { + "epoch": 57.45322245322245, + "grad_norm": 0.0005090531194582582, + "learning_rate": 0.06535650466543427, + "loss": 0.2869, + "num_input_tokens_seen": 21075752, + "step": 27635 + }, + { + "epoch": 57.46361746361746, + "grad_norm": 0.00022100505884736776, + "learning_rate": 0.0653078807134538, + "loss": 0.2763, + "num_input_tokens_seen": 21079560, + "step": 27640 + }, + { + "epoch": 57.474012474012476, + "grad_norm": 0.00019075334421359003, + "learning_rate": 0.06525926982206236, + "loss": 0.2653, + "num_input_tokens_seen": 21083464, + "step": 27645 + }, + { + "epoch": 57.484407484407484, + "grad_norm": 0.0005287841777317226, + "learning_rate": 0.06521067199875648, + "loss": 0.2757, + "num_input_tokens_seen": 21087432, + "step": 27650 + }, + { + "epoch": 57.49480249480249, + "grad_norm": 0.0003339296963531524, + "learning_rate": 0.06516208725103047, + "loss": 0.274, + "num_input_tokens_seen": 21091304, + "step": 27655 + }, + { + "epoch": 57.50519750519751, + "grad_norm": 0.00017390261928085238, + "learning_rate": 0.06511351558637678, + "loss": 0.2716, + "num_input_tokens_seen": 21095208, + "step": 27660 + }, + { + "epoch": 57.515592515592516, + "grad_norm": 0.00043048374936915934, + "learning_rate": 0.06506495701228569, + "loss": 0.2626, + "num_input_tokens_seen": 21099112, + "step": 27665 + }, + { + "epoch": 57.525987525987524, + "grad_norm": 0.00015417866234201938, + "learning_rate": 0.06501641153624559, + "loss": 0.2808, + "num_input_tokens_seen": 21102952, + "step": 27670 + }, + { + "epoch": 57.53638253638254, + "grad_norm": 0.00010996204218827188, + "learning_rate": 0.06496787916574286, + "loss": 0.268, + "num_input_tokens_seen": 21106728, + "step": 27675 + }, + { + "epoch": 57.54677754677755, + "grad_norm": 0.001330801984295249, + "learning_rate": 0.06491935990826168, + "loss": 0.273, + "num_input_tokens_seen": 21110536, + "step": 27680 + }, + { + "epoch": 57.55717255717256, + "grad_norm": 0.00032189779449254274, + "learning_rate": 0.0648708537712844, + "loss": 0.2689, + "num_input_tokens_seen": 21114504, + "step": 27685 + }, + { + "epoch": 57.567567567567565, + "grad_norm": 0.0006891913944855332, + "learning_rate": 0.06482236076229132, + "loss": 0.2425, + "num_input_tokens_seen": 21118248, + "step": 27690 + }, + { + "epoch": 57.57796257796258, + "grad_norm": 0.00024683194351382554, + "learning_rate": 0.06477388088876056, + "loss": 0.2482, + "num_input_tokens_seen": 21121992, + "step": 27695 + }, + { + "epoch": 57.58835758835759, + "grad_norm": 0.00025702352286316454, + "learning_rate": 0.06472541415816846, + "loss": 0.2942, + "num_input_tokens_seen": 21125736, + "step": 27700 + }, + { + "epoch": 57.5987525987526, + "grad_norm": 0.0001310812949668616, + "learning_rate": 0.06467696057798909, + "loss": 0.2584, + "num_input_tokens_seen": 21129704, + "step": 27705 + }, + { + "epoch": 57.60914760914761, + "grad_norm": 0.00045357918133959174, + "learning_rate": 0.0646285201556946, + "loss": 0.2937, + "num_input_tokens_seen": 21133512, + "step": 27710 + }, + { + "epoch": 57.61954261954262, + "grad_norm": 0.00026637132395990193, + "learning_rate": 0.06458009289875521, + "loss": 0.2398, + "num_input_tokens_seen": 21137224, + "step": 27715 + }, + { + "epoch": 57.62993762993763, + "grad_norm": 0.00027478273841552436, + "learning_rate": 0.0645316788146389, + "loss": 0.2668, + "num_input_tokens_seen": 21140936, + "step": 27720 + }, + { + "epoch": 57.64033264033264, + "grad_norm": 0.0002912095806095749, + "learning_rate": 0.06448327791081175, + "loss": 0.2398, + "num_input_tokens_seen": 21144680, + "step": 27725 + }, + { + "epoch": 57.65072765072765, + "grad_norm": 0.00016119316569529474, + "learning_rate": 0.0644348901947379, + "loss": 0.2493, + "num_input_tokens_seen": 21148520, + "step": 27730 + }, + { + "epoch": 57.66112266112266, + "grad_norm": 0.00014505573199130595, + "learning_rate": 0.06438651567387917, + "loss": 0.2706, + "num_input_tokens_seen": 21152488, + "step": 27735 + }, + { + "epoch": 57.67151767151767, + "grad_norm": 8.437263750238344e-05, + "learning_rate": 0.0643381543556957, + "loss": 0.2775, + "num_input_tokens_seen": 21156424, + "step": 27740 + }, + { + "epoch": 57.681912681912685, + "grad_norm": 0.00010920163185801357, + "learning_rate": 0.06428980624764526, + "loss": 0.2526, + "num_input_tokens_seen": 21160136, + "step": 27745 + }, + { + "epoch": 57.69230769230769, + "grad_norm": 0.00016672284982632846, + "learning_rate": 0.06424147135718378, + "loss": 0.2875, + "num_input_tokens_seen": 21163912, + "step": 27750 + }, + { + "epoch": 57.7027027027027, + "grad_norm": 0.000708926236256957, + "learning_rate": 0.06419314969176519, + "loss": 0.2712, + "num_input_tokens_seen": 21167848, + "step": 27755 + }, + { + "epoch": 57.71309771309771, + "grad_norm": 4.3382948206271976e-05, + "learning_rate": 0.06414484125884118, + "loss": 0.2731, + "num_input_tokens_seen": 21171592, + "step": 27760 + }, + { + "epoch": 57.723492723492726, + "grad_norm": 0.0003938140580430627, + "learning_rate": 0.06409654606586157, + "loss": 0.2694, + "num_input_tokens_seen": 21175432, + "step": 27765 + }, + { + "epoch": 57.733887733887734, + "grad_norm": 6.75722403684631e-05, + "learning_rate": 0.06404826412027415, + "loss": 0.2676, + "num_input_tokens_seen": 21179144, + "step": 27770 + }, + { + "epoch": 57.74428274428274, + "grad_norm": 0.0005064334254711866, + "learning_rate": 0.06399999542952453, + "loss": 0.276, + "num_input_tokens_seen": 21182792, + "step": 27775 + }, + { + "epoch": 57.75467775467776, + "grad_norm": 5.749869524152018e-05, + "learning_rate": 0.0639517400010563, + "loss": 0.2531, + "num_input_tokens_seen": 21186408, + "step": 27780 + }, + { + "epoch": 57.765072765072766, + "grad_norm": 4.6296434447867796e-05, + "learning_rate": 0.06390349784231118, + "loss": 0.2667, + "num_input_tokens_seen": 21190248, + "step": 27785 + }, + { + "epoch": 57.775467775467774, + "grad_norm": 0.0001912676525535062, + "learning_rate": 0.06385526896072859, + "loss": 0.2605, + "num_input_tokens_seen": 21194056, + "step": 27790 + }, + { + "epoch": 57.78586278586278, + "grad_norm": 8.812930173007771e-05, + "learning_rate": 0.06380705336374613, + "loss": 0.2776, + "num_input_tokens_seen": 21197832, + "step": 27795 + }, + { + "epoch": 57.7962577962578, + "grad_norm": 0.00041584664722904563, + "learning_rate": 0.06375885105879918, + "loss": 0.269, + "num_input_tokens_seen": 21201640, + "step": 27800 + }, + { + "epoch": 57.7962577962578, + "eval_loss": 0.2509770393371582, + "eval_runtime": 13.373, + "eval_samples_per_second": 64.009, + "eval_steps_per_second": 16.002, + "num_input_tokens_seen": 21201640, + "step": 27800 + }, + { + "epoch": 57.80665280665281, + "grad_norm": 0.00025830318918451667, + "learning_rate": 0.06371066205332115, + "loss": 0.2429, + "num_input_tokens_seen": 21205352, + "step": 27805 + }, + { + "epoch": 57.817047817047815, + "grad_norm": 5.51459270354826e-05, + "learning_rate": 0.06366248635474347, + "loss": 0.2515, + "num_input_tokens_seen": 21209064, + "step": 27810 + }, + { + "epoch": 57.82744282744283, + "grad_norm": 6.299043161561713e-05, + "learning_rate": 0.06361432397049532, + "loss": 0.2563, + "num_input_tokens_seen": 21212840, + "step": 27815 + }, + { + "epoch": 57.83783783783784, + "grad_norm": 0.0003149108379147947, + "learning_rate": 0.06356617490800408, + "loss": 0.2433, + "num_input_tokens_seen": 21216744, + "step": 27820 + }, + { + "epoch": 57.84823284823285, + "grad_norm": 0.0002543545560911298, + "learning_rate": 0.06351803917469478, + "loss": 0.3197, + "num_input_tokens_seen": 21220808, + "step": 27825 + }, + { + "epoch": 57.858627858627855, + "grad_norm": 0.0004248986078891903, + "learning_rate": 0.06346991677799067, + "loss": 0.2608, + "num_input_tokens_seen": 21224648, + "step": 27830 + }, + { + "epoch": 57.86902286902287, + "grad_norm": 0.00026921232347376645, + "learning_rate": 0.06342180772531283, + "loss": 0.2684, + "num_input_tokens_seen": 21228552, + "step": 27835 + }, + { + "epoch": 57.87941787941788, + "grad_norm": 0.00020852299348916858, + "learning_rate": 0.06337371202408021, + "loss": 0.261, + "num_input_tokens_seen": 21232392, + "step": 27840 + }, + { + "epoch": 57.88981288981289, + "grad_norm": 0.00013755708641838282, + "learning_rate": 0.06332562968170984, + "loss": 0.2454, + "num_input_tokens_seen": 21236168, + "step": 27845 + }, + { + "epoch": 57.9002079002079, + "grad_norm": 0.00022799032740294933, + "learning_rate": 0.06327756070561656, + "loss": 0.2665, + "num_input_tokens_seen": 21240232, + "step": 27850 + }, + { + "epoch": 57.91060291060291, + "grad_norm": 4.474133311305195e-05, + "learning_rate": 0.06322950510321329, + "loss": 0.2789, + "num_input_tokens_seen": 21243976, + "step": 27855 + }, + { + "epoch": 57.92099792099792, + "grad_norm": 0.0003047766222152859, + "learning_rate": 0.06318146288191076, + "loss": 0.282, + "num_input_tokens_seen": 21247720, + "step": 27860 + }, + { + "epoch": 57.931392931392935, + "grad_norm": 0.00011227883805986494, + "learning_rate": 0.06313343404911763, + "loss": 0.2559, + "num_input_tokens_seen": 21251528, + "step": 27865 + }, + { + "epoch": 57.94178794178794, + "grad_norm": 4.382493716548197e-05, + "learning_rate": 0.0630854186122406, + "loss": 0.2669, + "num_input_tokens_seen": 21255176, + "step": 27870 + }, + { + "epoch": 57.95218295218295, + "grad_norm": 0.00026669702492654324, + "learning_rate": 0.06303741657868431, + "loss": 0.2574, + "num_input_tokens_seen": 21258920, + "step": 27875 + }, + { + "epoch": 57.96257796257796, + "grad_norm": 7.003654900472611e-05, + "learning_rate": 0.06298942795585115, + "loss": 0.2672, + "num_input_tokens_seen": 21262696, + "step": 27880 + }, + { + "epoch": 57.972972972972975, + "grad_norm": 0.00035903751268051565, + "learning_rate": 0.06294145275114167, + "loss": 0.2514, + "num_input_tokens_seen": 21266344, + "step": 27885 + }, + { + "epoch": 57.983367983367984, + "grad_norm": 0.00019085004169028252, + "learning_rate": 0.06289349097195428, + "loss": 0.2372, + "num_input_tokens_seen": 21270152, + "step": 27890 + }, + { + "epoch": 57.99376299376299, + "grad_norm": 0.00032597981044091284, + "learning_rate": 0.06284554262568516, + "loss": 0.2674, + "num_input_tokens_seen": 21274024, + "step": 27895 + }, + { + "epoch": 58.00415800415801, + "grad_norm": 0.0001216758246300742, + "learning_rate": 0.06279760771972868, + "loss": 0.2545, + "num_input_tokens_seen": 21277792, + "step": 27900 + }, + { + "epoch": 58.014553014553016, + "grad_norm": 7.99032422946766e-05, + "learning_rate": 0.06274968626147688, + "loss": 0.2557, + "num_input_tokens_seen": 21281664, + "step": 27905 + }, + { + "epoch": 58.024948024948024, + "grad_norm": 0.00011803515371866524, + "learning_rate": 0.06270177825831993, + "loss": 0.2572, + "num_input_tokens_seen": 21285504, + "step": 27910 + }, + { + "epoch": 58.03534303534303, + "grad_norm": 0.00018923603056464344, + "learning_rate": 0.06265388371764587, + "loss": 0.2737, + "num_input_tokens_seen": 21289312, + "step": 27915 + }, + { + "epoch": 58.04573804573805, + "grad_norm": 0.0004680600541178137, + "learning_rate": 0.0626060026468406, + "loss": 0.2679, + "num_input_tokens_seen": 21293120, + "step": 27920 + }, + { + "epoch": 58.056133056133056, + "grad_norm": 0.0001631857012398541, + "learning_rate": 0.06255813505328794, + "loss": 0.2546, + "num_input_tokens_seen": 21297152, + "step": 27925 + }, + { + "epoch": 58.066528066528065, + "grad_norm": 0.0002768254198599607, + "learning_rate": 0.06251028094436978, + "loss": 0.2837, + "num_input_tokens_seen": 21300896, + "step": 27930 + }, + { + "epoch": 58.07692307692308, + "grad_norm": 0.0006342835840769112, + "learning_rate": 0.06246244032746568, + "loss": 0.2775, + "num_input_tokens_seen": 21304704, + "step": 27935 + }, + { + "epoch": 58.08731808731809, + "grad_norm": 0.00023807137040421367, + "learning_rate": 0.06241461320995342, + "loss": 0.2729, + "num_input_tokens_seen": 21308512, + "step": 27940 + }, + { + "epoch": 58.0977130977131, + "grad_norm": 0.0005684626521542668, + "learning_rate": 0.062366799599208426, + "loss": 0.2704, + "num_input_tokens_seen": 21312448, + "step": 27945 + }, + { + "epoch": 58.108108108108105, + "grad_norm": 6.058826329535805e-05, + "learning_rate": 0.06231899950260418, + "loss": 0.2559, + "num_input_tokens_seen": 21316288, + "step": 27950 + }, + { + "epoch": 58.11850311850312, + "grad_norm": 0.0005555640091188252, + "learning_rate": 0.06227121292751214, + "loss": 0.2737, + "num_input_tokens_seen": 21320128, + "step": 27955 + }, + { + "epoch": 58.12889812889813, + "grad_norm": 0.0012029018253087997, + "learning_rate": 0.062223439881301496, + "loss": 0.2763, + "num_input_tokens_seen": 21323904, + "step": 27960 + }, + { + "epoch": 58.13929313929314, + "grad_norm": 0.0002661730395630002, + "learning_rate": 0.06217568037133948, + "loss": 0.2586, + "num_input_tokens_seen": 21327584, + "step": 27965 + }, + { + "epoch": 58.14968814968815, + "grad_norm": 0.0002500272821635008, + "learning_rate": 0.06212793440499126, + "loss": 0.2553, + "num_input_tokens_seen": 21331392, + "step": 27970 + }, + { + "epoch": 58.16008316008316, + "grad_norm": 6.795930676162243e-05, + "learning_rate": 0.062080201989619783, + "loss": 0.2655, + "num_input_tokens_seen": 21335168, + "step": 27975 + }, + { + "epoch": 58.17047817047817, + "grad_norm": 9.237636550096795e-05, + "learning_rate": 0.062032483132586094, + "loss": 0.2771, + "num_input_tokens_seen": 21339104, + "step": 27980 + }, + { + "epoch": 58.18087318087318, + "grad_norm": 0.0004958044737577438, + "learning_rate": 0.0619847778412489, + "loss": 0.2817, + "num_input_tokens_seen": 21342880, + "step": 27985 + }, + { + "epoch": 58.19126819126819, + "grad_norm": 0.0006020687287673354, + "learning_rate": 0.06193708612296509, + "loss": 0.2499, + "num_input_tokens_seen": 21346624, + "step": 27990 + }, + { + "epoch": 58.2016632016632, + "grad_norm": 7.999097579158843e-05, + "learning_rate": 0.06188940798508923, + "loss": 0.2542, + "num_input_tokens_seen": 21350400, + "step": 27995 + }, + { + "epoch": 58.21205821205821, + "grad_norm": 0.00011698801245074719, + "learning_rate": 0.06184174343497397, + "loss": 0.2667, + "num_input_tokens_seen": 21354208, + "step": 28000 + }, + { + "epoch": 58.21205821205821, + "eval_loss": 0.2471611350774765, + "eval_runtime": 13.4136, + "eval_samples_per_second": 63.816, + "eval_steps_per_second": 15.954, + "num_input_tokens_seen": 21354208, + "step": 28000 + }, + { + "epoch": 58.222453222453225, + "grad_norm": 0.0003057768917642534, + "learning_rate": 0.061794092479969726, + "loss": 0.274, + "num_input_tokens_seen": 21357952, + "step": 28005 + }, + { + "epoch": 58.232848232848234, + "grad_norm": 0.0005550442729145288, + "learning_rate": 0.06174645512742485, + "loss": 0.2374, + "num_input_tokens_seen": 21361664, + "step": 28010 + }, + { + "epoch": 58.24324324324324, + "grad_norm": 0.0001592139306012541, + "learning_rate": 0.06169883138468565, + "loss": 0.2572, + "num_input_tokens_seen": 21365504, + "step": 28015 + }, + { + "epoch": 58.25363825363825, + "grad_norm": 4.640250699594617e-05, + "learning_rate": 0.06165122125909637, + "loss": 0.2446, + "num_input_tokens_seen": 21369120, + "step": 28020 + }, + { + "epoch": 58.264033264033266, + "grad_norm": 0.00048565768520347774, + "learning_rate": 0.061603624757998965, + "loss": 0.2686, + "num_input_tokens_seen": 21373024, + "step": 28025 + }, + { + "epoch": 58.274428274428274, + "grad_norm": 0.0001465724635636434, + "learning_rate": 0.0615560418887335, + "loss": 0.2691, + "num_input_tokens_seen": 21376928, + "step": 28030 + }, + { + "epoch": 58.28482328482328, + "grad_norm": 0.0010097576305270195, + "learning_rate": 0.06150847265863787, + "loss": 0.2655, + "num_input_tokens_seen": 21380640, + "step": 28035 + }, + { + "epoch": 58.2952182952183, + "grad_norm": 0.0001094237231882289, + "learning_rate": 0.061460917075047757, + "loss": 0.2572, + "num_input_tokens_seen": 21384544, + "step": 28040 + }, + { + "epoch": 58.305613305613306, + "grad_norm": 0.0004562342946883291, + "learning_rate": 0.06141337514529694, + "loss": 0.2551, + "num_input_tokens_seen": 21388480, + "step": 28045 + }, + { + "epoch": 58.316008316008315, + "grad_norm": 0.00030350618180818856, + "learning_rate": 0.06136584687671687, + "loss": 0.2546, + "num_input_tokens_seen": 21392256, + "step": 28050 + }, + { + "epoch": 58.32640332640332, + "grad_norm": 0.00025551021099090576, + "learning_rate": 0.061318332276637064, + "loss": 0.253, + "num_input_tokens_seen": 21396064, + "step": 28055 + }, + { + "epoch": 58.33679833679834, + "grad_norm": 0.00012534890265669674, + "learning_rate": 0.06127083135238491, + "loss": 0.2613, + "num_input_tokens_seen": 21399776, + "step": 28060 + }, + { + "epoch": 58.34719334719335, + "grad_norm": 0.0003287180734332651, + "learning_rate": 0.06122334411128555, + "loss": 0.2569, + "num_input_tokens_seen": 21403424, + "step": 28065 + }, + { + "epoch": 58.357588357588355, + "grad_norm": 0.00011033266491722316, + "learning_rate": 0.06117587056066223, + "loss": 0.2773, + "num_input_tokens_seen": 21407232, + "step": 28070 + }, + { + "epoch": 58.36798336798337, + "grad_norm": 0.00011290735710645095, + "learning_rate": 0.06112841070783589, + "loss": 0.2593, + "num_input_tokens_seen": 21411072, + "step": 28075 + }, + { + "epoch": 58.37837837837838, + "grad_norm": 0.00029994198121130466, + "learning_rate": 0.061080964560125406, + "loss": 0.2812, + "num_input_tokens_seen": 21415008, + "step": 28080 + }, + { + "epoch": 58.38877338877339, + "grad_norm": 0.000195614222320728, + "learning_rate": 0.06103353212484766, + "loss": 0.284, + "num_input_tokens_seen": 21418784, + "step": 28085 + }, + { + "epoch": 58.3991683991684, + "grad_norm": 0.0003068026271648705, + "learning_rate": 0.06098611340931722, + "loss": 0.2571, + "num_input_tokens_seen": 21422528, + "step": 28090 + }, + { + "epoch": 58.40956340956341, + "grad_norm": 0.0003253926697652787, + "learning_rate": 0.06093870842084672, + "loss": 0.2646, + "num_input_tokens_seen": 21426272, + "step": 28095 + }, + { + "epoch": 58.41995841995842, + "grad_norm": 0.0004531065351329744, + "learning_rate": 0.06089131716674666, + "loss": 0.2646, + "num_input_tokens_seen": 21430112, + "step": 28100 + }, + { + "epoch": 58.43035343035343, + "grad_norm": 0.00031820303411222994, + "learning_rate": 0.060843939654325226, + "loss": 0.247, + "num_input_tokens_seen": 21434048, + "step": 28105 + }, + { + "epoch": 58.44074844074844, + "grad_norm": 5.92818578297738e-05, + "learning_rate": 0.06079657589088873, + "loss": 0.2556, + "num_input_tokens_seen": 21437952, + "step": 28110 + }, + { + "epoch": 58.45114345114345, + "grad_norm": 0.00018259974604006857, + "learning_rate": 0.06074922588374126, + "loss": 0.2659, + "num_input_tokens_seen": 21441728, + "step": 28115 + }, + { + "epoch": 58.46153846153846, + "grad_norm": 0.0004484377568587661, + "learning_rate": 0.06070188964018472, + "loss": 0.2485, + "num_input_tokens_seen": 21445536, + "step": 28120 + }, + { + "epoch": 58.471933471933475, + "grad_norm": 0.00017381951329298317, + "learning_rate": 0.06065456716751902, + "loss": 0.2752, + "num_input_tokens_seen": 21449376, + "step": 28125 + }, + { + "epoch": 58.482328482328484, + "grad_norm": 0.0003474741824902594, + "learning_rate": 0.06060725847304182, + "loss": 0.2345, + "num_input_tokens_seen": 21453344, + "step": 28130 + }, + { + "epoch": 58.49272349272349, + "grad_norm": 0.00015435223758686334, + "learning_rate": 0.06055996356404877, + "loss": 0.2644, + "num_input_tokens_seen": 21457184, + "step": 28135 + }, + { + "epoch": 58.5031185031185, + "grad_norm": 0.0003369873738847673, + "learning_rate": 0.06051268244783327, + "loss": 0.272, + "num_input_tokens_seen": 21460992, + "step": 28140 + }, + { + "epoch": 58.513513513513516, + "grad_norm": 0.00045230990508571267, + "learning_rate": 0.06046541513168676, + "loss": 0.245, + "num_input_tokens_seen": 21464896, + "step": 28145 + }, + { + "epoch": 58.523908523908524, + "grad_norm": 0.0002394439943600446, + "learning_rate": 0.060418161622898356, + "loss": 0.27, + "num_input_tokens_seen": 21468640, + "step": 28150 + }, + { + "epoch": 58.53430353430353, + "grad_norm": 8.886944851838052e-05, + "learning_rate": 0.06037092192875521, + "loss": 0.2782, + "num_input_tokens_seen": 21472288, + "step": 28155 + }, + { + "epoch": 58.54469854469855, + "grad_norm": 0.00027448433684185147, + "learning_rate": 0.060323696056542225, + "loss": 0.2716, + "num_input_tokens_seen": 21476032, + "step": 28160 + }, + { + "epoch": 58.555093555093556, + "grad_norm": 0.0004599363310262561, + "learning_rate": 0.06027648401354229, + "loss": 0.2592, + "num_input_tokens_seen": 21479904, + "step": 28165 + }, + { + "epoch": 58.565488565488565, + "grad_norm": 5.780800711363554e-05, + "learning_rate": 0.06022928580703601, + "loss": 0.2674, + "num_input_tokens_seen": 21483776, + "step": 28170 + }, + { + "epoch": 58.57588357588357, + "grad_norm": 0.0006084410124458373, + "learning_rate": 0.060182101444301986, + "loss": 0.2877, + "num_input_tokens_seen": 21487744, + "step": 28175 + }, + { + "epoch": 58.58627858627859, + "grad_norm": 0.0003769618342630565, + "learning_rate": 0.06013493093261669, + "loss": 0.2543, + "num_input_tokens_seen": 21491456, + "step": 28180 + }, + { + "epoch": 58.5966735966736, + "grad_norm": 0.00043927814112976193, + "learning_rate": 0.06008777427925432, + "loss": 0.2603, + "num_input_tokens_seen": 21495520, + "step": 28185 + }, + { + "epoch": 58.607068607068605, + "grad_norm": 0.00027972384123131633, + "learning_rate": 0.06004063149148705, + "loss": 0.2649, + "num_input_tokens_seen": 21499296, + "step": 28190 + }, + { + "epoch": 58.61746361746362, + "grad_norm": 0.0005749806296080351, + "learning_rate": 0.05999350257658497, + "loss": 0.2436, + "num_input_tokens_seen": 21503104, + "step": 28195 + }, + { + "epoch": 58.62785862785863, + "grad_norm": 0.0001661076967138797, + "learning_rate": 0.05994638754181582, + "loss": 0.2785, + "num_input_tokens_seen": 21506752, + "step": 28200 + }, + { + "epoch": 58.62785862785863, + "eval_loss": 0.24804453551769257, + "eval_runtime": 13.4855, + "eval_samples_per_second": 63.476, + "eval_steps_per_second": 15.869, + "num_input_tokens_seen": 21506752, + "step": 28200 + }, + { + "epoch": 58.63825363825364, + "grad_norm": 8.766123210079968e-05, + "learning_rate": 0.059899286394445445, + "loss": 0.2677, + "num_input_tokens_seen": 21510592, + "step": 28205 + }, + { + "epoch": 58.648648648648646, + "grad_norm": 0.0002547898911871016, + "learning_rate": 0.059852199141737346, + "loss": 0.2834, + "num_input_tokens_seen": 21514560, + "step": 28210 + }, + { + "epoch": 58.65904365904366, + "grad_norm": 0.00027078064158558846, + "learning_rate": 0.05980512579095304, + "loss": 0.2383, + "num_input_tokens_seen": 21518240, + "step": 28215 + }, + { + "epoch": 58.66943866943867, + "grad_norm": 0.00023122024140320718, + "learning_rate": 0.05975806634935181, + "loss": 0.261, + "num_input_tokens_seen": 21522048, + "step": 28220 + }, + { + "epoch": 58.67983367983368, + "grad_norm": 0.00047529683797620237, + "learning_rate": 0.05971102082419076, + "loss": 0.2535, + "num_input_tokens_seen": 21525728, + "step": 28225 + }, + { + "epoch": 58.69022869022869, + "grad_norm": 6.772636697860435e-05, + "learning_rate": 0.05966398922272492, + "loss": 0.2898, + "num_input_tokens_seen": 21529632, + "step": 28230 + }, + { + "epoch": 58.7006237006237, + "grad_norm": 5.464633431984112e-05, + "learning_rate": 0.059616971552207236, + "loss": 0.258, + "num_input_tokens_seen": 21533440, + "step": 28235 + }, + { + "epoch": 58.71101871101871, + "grad_norm": 0.0004955774056725204, + "learning_rate": 0.059569967819888305, + "loss": 0.2572, + "num_input_tokens_seen": 21537312, + "step": 28240 + }, + { + "epoch": 58.72141372141372, + "grad_norm": 0.00041029872954823077, + "learning_rate": 0.05952297803301681, + "loss": 0.2752, + "num_input_tokens_seen": 21541056, + "step": 28245 + }, + { + "epoch": 58.731808731808734, + "grad_norm": 0.0002988222404383123, + "learning_rate": 0.059476002198839056, + "loss": 0.2683, + "num_input_tokens_seen": 21544832, + "step": 28250 + }, + { + "epoch": 58.74220374220374, + "grad_norm": 0.00037701110704801977, + "learning_rate": 0.05942904032459935, + "loss": 0.2838, + "num_input_tokens_seen": 21548640, + "step": 28255 + }, + { + "epoch": 58.75259875259875, + "grad_norm": 4.13027846661862e-05, + "learning_rate": 0.05938209241753987, + "loss": 0.2481, + "num_input_tokens_seen": 21552352, + "step": 28260 + }, + { + "epoch": 58.762993762993766, + "grad_norm": 0.0003292766341473907, + "learning_rate": 0.05933515848490046, + "loss": 0.2476, + "num_input_tokens_seen": 21556160, + "step": 28265 + }, + { + "epoch": 58.773388773388774, + "grad_norm": 0.0003277112846262753, + "learning_rate": 0.059288238533918985, + "loss": 0.2911, + "num_input_tokens_seen": 21559936, + "step": 28270 + }, + { + "epoch": 58.78378378378378, + "grad_norm": 0.00011082505079684779, + "learning_rate": 0.05924133257183113, + "loss": 0.2626, + "num_input_tokens_seen": 21563712, + "step": 28275 + }, + { + "epoch": 58.79417879417879, + "grad_norm": 0.000592010619584471, + "learning_rate": 0.059194440605870285, + "loss": 0.26, + "num_input_tokens_seen": 21567776, + "step": 28280 + }, + { + "epoch": 58.804573804573806, + "grad_norm": 0.00010882051719818264, + "learning_rate": 0.059147562643267884, + "loss": 0.2673, + "num_input_tokens_seen": 21571680, + "step": 28285 + }, + { + "epoch": 58.814968814968815, + "grad_norm": 0.0001854533766163513, + "learning_rate": 0.059100698691253055, + "loss": 0.2506, + "num_input_tokens_seen": 21575520, + "step": 28290 + }, + { + "epoch": 58.82536382536382, + "grad_norm": 0.00020167410548310727, + "learning_rate": 0.05905384875705273, + "loss": 0.2542, + "num_input_tokens_seen": 21579296, + "step": 28295 + }, + { + "epoch": 58.83575883575884, + "grad_norm": 0.0006994350114837289, + "learning_rate": 0.05900701284789189, + "loss": 0.2452, + "num_input_tokens_seen": 21583136, + "step": 28300 + }, + { + "epoch": 58.84615384615385, + "grad_norm": 0.00023889818112365901, + "learning_rate": 0.058960190970993115, + "loss": 0.2859, + "num_input_tokens_seen": 21586944, + "step": 28305 + }, + { + "epoch": 58.856548856548855, + "grad_norm": 0.0003332647029310465, + "learning_rate": 0.058913383133576955, + "loss": 0.2611, + "num_input_tokens_seen": 21590752, + "step": 28310 + }, + { + "epoch": 58.86694386694387, + "grad_norm": 0.0002455249195918441, + "learning_rate": 0.05886658934286185, + "loss": 0.2523, + "num_input_tokens_seen": 21594496, + "step": 28315 + }, + { + "epoch": 58.87733887733888, + "grad_norm": 0.0002897169906646013, + "learning_rate": 0.058819809606063846, + "loss": 0.2822, + "num_input_tokens_seen": 21598304, + "step": 28320 + }, + { + "epoch": 58.88773388773389, + "grad_norm": 0.00033308478305116296, + "learning_rate": 0.05877304393039711, + "loss": 0.2995, + "num_input_tokens_seen": 21602176, + "step": 28325 + }, + { + "epoch": 58.898128898128896, + "grad_norm": 0.00018867666949518025, + "learning_rate": 0.05872629232307338, + "loss": 0.2749, + "num_input_tokens_seen": 21606016, + "step": 28330 + }, + { + "epoch": 58.90852390852391, + "grad_norm": 0.0005097058019600809, + "learning_rate": 0.05867955479130239, + "loss": 0.282, + "num_input_tokens_seen": 21609792, + "step": 28335 + }, + { + "epoch": 58.91891891891892, + "grad_norm": 0.00013791833771392703, + "learning_rate": 0.058632831342291705, + "loss": 0.2692, + "num_input_tokens_seen": 21613664, + "step": 28340 + }, + { + "epoch": 58.92931392931393, + "grad_norm": 0.00035076276981271803, + "learning_rate": 0.05858612198324655, + "loss": 0.2796, + "num_input_tokens_seen": 21617504, + "step": 28345 + }, + { + "epoch": 58.93970893970894, + "grad_norm": 0.00025151160662062466, + "learning_rate": 0.05853942672137025, + "loss": 0.263, + "num_input_tokens_seen": 21621216, + "step": 28350 + }, + { + "epoch": 58.95010395010395, + "grad_norm": 0.00021297895000316203, + "learning_rate": 0.05849274556386363, + "loss": 0.242, + "num_input_tokens_seen": 21625024, + "step": 28355 + }, + { + "epoch": 58.96049896049896, + "grad_norm": 0.0002409277221886441, + "learning_rate": 0.05844607851792567, + "loss": 0.2574, + "num_input_tokens_seen": 21628928, + "step": 28360 + }, + { + "epoch": 58.97089397089397, + "grad_norm": 0.0004970143199898303, + "learning_rate": 0.058399425590752924, + "loss": 0.2959, + "num_input_tokens_seen": 21632800, + "step": 28365 + }, + { + "epoch": 58.981288981288984, + "grad_norm": 0.00031185749685391784, + "learning_rate": 0.05835278678953985, + "loss": 0.275, + "num_input_tokens_seen": 21636672, + "step": 28370 + }, + { + "epoch": 58.99168399168399, + "grad_norm": 0.0002325763343833387, + "learning_rate": 0.05830616212147874, + "loss": 0.2585, + "num_input_tokens_seen": 21640352, + "step": 28375 + }, + { + "epoch": 59.002079002079, + "grad_norm": 0.0002094060619128868, + "learning_rate": 0.058259551593759784, + "loss": 0.2815, + "num_input_tokens_seen": 21644240, + "step": 28380 + }, + { + "epoch": 59.012474012474016, + "grad_norm": 0.0004323652829043567, + "learning_rate": 0.058212955213570804, + "loss": 0.2699, + "num_input_tokens_seen": 21648176, + "step": 28385 + }, + { + "epoch": 59.022869022869024, + "grad_norm": 0.0009287319844588637, + "learning_rate": 0.0581663729880976, + "loss": 0.2752, + "num_input_tokens_seen": 21651984, + "step": 28390 + }, + { + "epoch": 59.03326403326403, + "grad_norm": 0.00040505407378077507, + "learning_rate": 0.05811980492452379, + "loss": 0.2769, + "num_input_tokens_seen": 21655856, + "step": 28395 + }, + { + "epoch": 59.04365904365904, + "grad_norm": 0.00010733300587162375, + "learning_rate": 0.058073251030030644, + "loss": 0.2669, + "num_input_tokens_seen": 21659696, + "step": 28400 + }, + { + "epoch": 59.04365904365904, + "eval_loss": 0.25756606459617615, + "eval_runtime": 13.4052, + "eval_samples_per_second": 63.856, + "eval_steps_per_second": 15.964, + "num_input_tokens_seen": 21659696, + "step": 28400 + }, + { + "epoch": 59.054054054054056, + "grad_norm": 0.0002945765445474535, + "learning_rate": 0.05802671131179747, + "loss": 0.2573, + "num_input_tokens_seen": 21663600, + "step": 28405 + }, + { + "epoch": 59.064449064449065, + "grad_norm": 4.221084600430913e-05, + "learning_rate": 0.057980185777001154, + "loss": 0.2706, + "num_input_tokens_seen": 21667376, + "step": 28410 + }, + { + "epoch": 59.07484407484407, + "grad_norm": 9.901568409986794e-05, + "learning_rate": 0.057933674432816606, + "loss": 0.28, + "num_input_tokens_seen": 21671312, + "step": 28415 + }, + { + "epoch": 59.08523908523909, + "grad_norm": 9.723401308292523e-05, + "learning_rate": 0.05788717728641648, + "loss": 0.2674, + "num_input_tokens_seen": 21675152, + "step": 28420 + }, + { + "epoch": 59.0956340956341, + "grad_norm": 2.7791116735897958e-05, + "learning_rate": 0.057840694344971126, + "loss": 0.2833, + "num_input_tokens_seen": 21678960, + "step": 28425 + }, + { + "epoch": 59.106029106029105, + "grad_norm": 5.818685895064846e-05, + "learning_rate": 0.0577942256156489, + "loss": 0.2644, + "num_input_tokens_seen": 21682864, + "step": 28430 + }, + { + "epoch": 59.11642411642411, + "grad_norm": 0.00011367109982529655, + "learning_rate": 0.057747771105615804, + "loss": 0.2633, + "num_input_tokens_seen": 21686640, + "step": 28435 + }, + { + "epoch": 59.12681912681913, + "grad_norm": 0.0004155321803409606, + "learning_rate": 0.05770133082203568, + "loss": 0.2426, + "num_input_tokens_seen": 21690416, + "step": 28440 + }, + { + "epoch": 59.13721413721414, + "grad_norm": 0.0002099186385748908, + "learning_rate": 0.0576549047720703, + "loss": 0.2735, + "num_input_tokens_seen": 21694128, + "step": 28445 + }, + { + "epoch": 59.147609147609145, + "grad_norm": 9.252870222553611e-05, + "learning_rate": 0.05760849296287902, + "loss": 0.2595, + "num_input_tokens_seen": 21697776, + "step": 28450 + }, + { + "epoch": 59.15800415800416, + "grad_norm": 6.353397475322708e-05, + "learning_rate": 0.05756209540161919, + "loss": 0.2932, + "num_input_tokens_seen": 21701552, + "step": 28455 + }, + { + "epoch": 59.16839916839917, + "grad_norm": 0.0002833329781424254, + "learning_rate": 0.05751571209544595, + "loss": 0.266, + "num_input_tokens_seen": 21705328, + "step": 28460 + }, + { + "epoch": 59.17879417879418, + "grad_norm": 0.00023166960454545915, + "learning_rate": 0.057469343051512085, + "loss": 0.2398, + "num_input_tokens_seen": 21708976, + "step": 28465 + }, + { + "epoch": 59.189189189189186, + "grad_norm": 0.00025007323711179197, + "learning_rate": 0.057422988276968324, + "loss": 0.2739, + "num_input_tokens_seen": 21712912, + "step": 28470 + }, + { + "epoch": 59.1995841995842, + "grad_norm": 0.00024293549358844757, + "learning_rate": 0.05737664777896323, + "loss": 0.2551, + "num_input_tokens_seen": 21716656, + "step": 28475 + }, + { + "epoch": 59.20997920997921, + "grad_norm": 0.00024286335974466056, + "learning_rate": 0.057330321564642975, + "loss": 0.249, + "num_input_tokens_seen": 21720432, + "step": 28480 + }, + { + "epoch": 59.22037422037422, + "grad_norm": 0.00023277421132661402, + "learning_rate": 0.05728400964115174, + "loss": 0.2523, + "num_input_tokens_seen": 21724240, + "step": 28485 + }, + { + "epoch": 59.23076923076923, + "grad_norm": 3.65491105185356e-05, + "learning_rate": 0.057237712015631305, + "loss": 0.2625, + "num_input_tokens_seen": 21728144, + "step": 28490 + }, + { + "epoch": 59.24116424116424, + "grad_norm": 4.9077800213126466e-05, + "learning_rate": 0.057191428695221425, + "loss": 0.2609, + "num_input_tokens_seen": 21731920, + "step": 28495 + }, + { + "epoch": 59.25155925155925, + "grad_norm": 8.089486800599843e-05, + "learning_rate": 0.05714515968705958, + "loss": 0.27, + "num_input_tokens_seen": 21735760, + "step": 28500 + }, + { + "epoch": 59.26195426195426, + "grad_norm": 0.0001673252700129524, + "learning_rate": 0.05709890499828099, + "loss": 0.2628, + "num_input_tokens_seen": 21739536, + "step": 28505 + }, + { + "epoch": 59.272349272349274, + "grad_norm": 0.00032226540497504175, + "learning_rate": 0.05705266463601868, + "loss": 0.2445, + "num_input_tokens_seen": 21743280, + "step": 28510 + }, + { + "epoch": 59.28274428274428, + "grad_norm": 0.0003142687492072582, + "learning_rate": 0.057006438607403565, + "loss": 0.241, + "num_input_tokens_seen": 21746992, + "step": 28515 + }, + { + "epoch": 59.29313929313929, + "grad_norm": 0.0002858842199202627, + "learning_rate": 0.056960226919564205, + "loss": 0.257, + "num_input_tokens_seen": 21750800, + "step": 28520 + }, + { + "epoch": 59.303534303534306, + "grad_norm": 0.0004515312029980123, + "learning_rate": 0.05691402957962713, + "loss": 0.2857, + "num_input_tokens_seen": 21754576, + "step": 28525 + }, + { + "epoch": 59.313929313929314, + "grad_norm": 0.0004459033953025937, + "learning_rate": 0.05686784659471642, + "loss": 0.2688, + "num_input_tokens_seen": 21758352, + "step": 28530 + }, + { + "epoch": 59.32432432432432, + "grad_norm": 0.0004037866892758757, + "learning_rate": 0.056821677971954136, + "loss": 0.252, + "num_input_tokens_seen": 21762224, + "step": 28535 + }, + { + "epoch": 59.33471933471934, + "grad_norm": 0.0002533178194426, + "learning_rate": 0.05677552371846012, + "loss": 0.2607, + "num_input_tokens_seen": 21766000, + "step": 28540 + }, + { + "epoch": 59.34511434511435, + "grad_norm": 8.1624835729599e-05, + "learning_rate": 0.05672938384135182, + "loss": 0.2619, + "num_input_tokens_seen": 21769936, + "step": 28545 + }, + { + "epoch": 59.355509355509355, + "grad_norm": 0.00014089929754845798, + "learning_rate": 0.05668325834774465, + "loss": 0.2649, + "num_input_tokens_seen": 21773776, + "step": 28550 + }, + { + "epoch": 59.36590436590436, + "grad_norm": 0.0004051189753226936, + "learning_rate": 0.05663714724475177, + "loss": 0.2502, + "num_input_tokens_seen": 21777616, + "step": 28555 + }, + { + "epoch": 59.37629937629938, + "grad_norm": 0.00048392615281045437, + "learning_rate": 0.05659105053948403, + "loss": 0.2574, + "num_input_tokens_seen": 21781360, + "step": 28560 + }, + { + "epoch": 59.38669438669439, + "grad_norm": 0.00015165570948738605, + "learning_rate": 0.056544968239050176, + "loss": 0.277, + "num_input_tokens_seen": 21785104, + "step": 28565 + }, + { + "epoch": 59.397089397089395, + "grad_norm": 0.0005822497187182307, + "learning_rate": 0.056498900350556616, + "loss": 0.249, + "num_input_tokens_seen": 21788816, + "step": 28570 + }, + { + "epoch": 59.40748440748441, + "grad_norm": 0.0005350075662136078, + "learning_rate": 0.05645284688110766, + "loss": 0.2766, + "num_input_tokens_seen": 21792592, + "step": 28575 + }, + { + "epoch": 59.41787941787942, + "grad_norm": 0.00042768902494572103, + "learning_rate": 0.05640680783780532, + "loss": 0.2606, + "num_input_tokens_seen": 21796400, + "step": 28580 + }, + { + "epoch": 59.42827442827443, + "grad_norm": 0.0002751356805674732, + "learning_rate": 0.056360783227749324, + "loss": 0.2648, + "num_input_tokens_seen": 21800080, + "step": 28585 + }, + { + "epoch": 59.438669438669436, + "grad_norm": 0.0008365960675291717, + "learning_rate": 0.05631477305803728, + "loss": 0.2638, + "num_input_tokens_seen": 21803728, + "step": 28590 + }, + { + "epoch": 59.44906444906445, + "grad_norm": 0.00010132075840374455, + "learning_rate": 0.05626877733576462, + "loss": 0.2327, + "num_input_tokens_seen": 21807728, + "step": 28595 + }, + { + "epoch": 59.45945945945946, + "grad_norm": 0.00025405475753359497, + "learning_rate": 0.05622279606802435, + "loss": 0.2614, + "num_input_tokens_seen": 21811600, + "step": 28600 + }, + { + "epoch": 59.45945945945946, + "eval_loss": 0.24791602790355682, + "eval_runtime": 13.4042, + "eval_samples_per_second": 63.861, + "eval_steps_per_second": 15.965, + "num_input_tokens_seen": 21811600, + "step": 28600 + }, + { + "epoch": 59.46985446985447, + "grad_norm": 0.00014457415090873837, + "learning_rate": 0.05617682926190744, + "loss": 0.2639, + "num_input_tokens_seen": 21815344, + "step": 28605 + }, + { + "epoch": 59.48024948024948, + "grad_norm": 0.00020768803369719535, + "learning_rate": 0.05613087692450248, + "loss": 0.2842, + "num_input_tokens_seen": 21819312, + "step": 28610 + }, + { + "epoch": 59.49064449064449, + "grad_norm": 0.00032393611036241055, + "learning_rate": 0.05608493906289592, + "loss": 0.2642, + "num_input_tokens_seen": 21823184, + "step": 28615 + }, + { + "epoch": 59.5010395010395, + "grad_norm": 0.00032815185841172934, + "learning_rate": 0.05603901568417201, + "loss": 0.2711, + "num_input_tokens_seen": 21827152, + "step": 28620 + }, + { + "epoch": 59.51143451143451, + "grad_norm": 0.000411802320741117, + "learning_rate": 0.055993106795412625, + "loss": 0.2442, + "num_input_tokens_seen": 21831184, + "step": 28625 + }, + { + "epoch": 59.521829521829524, + "grad_norm": 9.419520210940391e-05, + "learning_rate": 0.05594721240369759, + "loss": 0.273, + "num_input_tokens_seen": 21834992, + "step": 28630 + }, + { + "epoch": 59.53222453222453, + "grad_norm": 0.0004108079883735627, + "learning_rate": 0.055901332516104296, + "loss": 0.2675, + "num_input_tokens_seen": 21838896, + "step": 28635 + }, + { + "epoch": 59.54261954261954, + "grad_norm": 0.00042091155773960054, + "learning_rate": 0.05585546713970804, + "loss": 0.2484, + "num_input_tokens_seen": 21842992, + "step": 28640 + }, + { + "epoch": 59.553014553014556, + "grad_norm": 8.988139597931877e-05, + "learning_rate": 0.05580961628158189, + "loss": 0.2634, + "num_input_tokens_seen": 21846672, + "step": 28645 + }, + { + "epoch": 59.563409563409564, + "grad_norm": 9.53419366851449e-05, + "learning_rate": 0.05576377994879659, + "loss": 0.2711, + "num_input_tokens_seen": 21850480, + "step": 28650 + }, + { + "epoch": 59.57380457380457, + "grad_norm": 0.0003029521321877837, + "learning_rate": 0.05571795814842063, + "loss": 0.2816, + "num_input_tokens_seen": 21854288, + "step": 28655 + }, + { + "epoch": 59.58419958419958, + "grad_norm": 0.0004547022108454257, + "learning_rate": 0.05567215088752037, + "loss": 0.2727, + "num_input_tokens_seen": 21858064, + "step": 28660 + }, + { + "epoch": 59.5945945945946, + "grad_norm": 0.00023953290656208992, + "learning_rate": 0.05562635817315981, + "loss": 0.2757, + "num_input_tokens_seen": 21861808, + "step": 28665 + }, + { + "epoch": 59.604989604989605, + "grad_norm": 0.00022702181013301015, + "learning_rate": 0.05558058001240083, + "loss": 0.2668, + "num_input_tokens_seen": 21865744, + "step": 28670 + }, + { + "epoch": 59.61538461538461, + "grad_norm": 5.641702955472283e-05, + "learning_rate": 0.055534816412302915, + "loss": 0.2754, + "num_input_tokens_seen": 21869360, + "step": 28675 + }, + { + "epoch": 59.62577962577963, + "grad_norm": 0.00037678508670069277, + "learning_rate": 0.055489067379923436, + "loss": 0.2734, + "num_input_tokens_seen": 21872976, + "step": 28680 + }, + { + "epoch": 59.63617463617464, + "grad_norm": 4.154805719736032e-05, + "learning_rate": 0.055443332922317505, + "loss": 0.2673, + "num_input_tokens_seen": 21876848, + "step": 28685 + }, + { + "epoch": 59.646569646569645, + "grad_norm": 0.0001509753637947142, + "learning_rate": 0.055397613046537876, + "loss": 0.2692, + "num_input_tokens_seen": 21880624, + "step": 28690 + }, + { + "epoch": 59.656964656964654, + "grad_norm": 0.00053272774675861, + "learning_rate": 0.055351907759635145, + "loss": 0.2567, + "num_input_tokens_seen": 21884464, + "step": 28695 + }, + { + "epoch": 59.66735966735967, + "grad_norm": 0.00011537809041328728, + "learning_rate": 0.05530621706865772, + "loss": 0.2789, + "num_input_tokens_seen": 21888336, + "step": 28700 + }, + { + "epoch": 59.67775467775468, + "grad_norm": 6.822626892244443e-05, + "learning_rate": 0.055260540980651564, + "loss": 0.2539, + "num_input_tokens_seen": 21892304, + "step": 28705 + }, + { + "epoch": 59.688149688149686, + "grad_norm": 0.00023755435540806502, + "learning_rate": 0.05521487950266062, + "loss": 0.2614, + "num_input_tokens_seen": 21896176, + "step": 28710 + }, + { + "epoch": 59.6985446985447, + "grad_norm": 0.00015982553304638714, + "learning_rate": 0.055169232641726344, + "loss": 0.2797, + "num_input_tokens_seen": 21900016, + "step": 28715 + }, + { + "epoch": 59.70893970893971, + "grad_norm": 0.00027029550983570516, + "learning_rate": 0.055123600404888166, + "loss": 0.2566, + "num_input_tokens_seen": 21903856, + "step": 28720 + }, + { + "epoch": 59.71933471933472, + "grad_norm": 0.0006020541186444461, + "learning_rate": 0.05507798279918309, + "loss": 0.2547, + "num_input_tokens_seen": 21907632, + "step": 28725 + }, + { + "epoch": 59.729729729729726, + "grad_norm": 0.00022213983174879104, + "learning_rate": 0.0550323798316459, + "loss": 0.2538, + "num_input_tokens_seen": 21911472, + "step": 28730 + }, + { + "epoch": 59.74012474012474, + "grad_norm": 0.0003540258330758661, + "learning_rate": 0.05498679150930916, + "loss": 0.243, + "num_input_tokens_seen": 21915088, + "step": 28735 + }, + { + "epoch": 59.75051975051975, + "grad_norm": 3.7383029848570004e-05, + "learning_rate": 0.05494121783920323, + "loss": 0.2648, + "num_input_tokens_seen": 21918800, + "step": 28740 + }, + { + "epoch": 59.76091476091476, + "grad_norm": 9.713468898553401e-05, + "learning_rate": 0.05489565882835605, + "loss": 0.2845, + "num_input_tokens_seen": 21922576, + "step": 28745 + }, + { + "epoch": 59.771309771309774, + "grad_norm": 0.00011383960372768342, + "learning_rate": 0.05485011448379348, + "loss": 0.2829, + "num_input_tokens_seen": 21926448, + "step": 28750 + }, + { + "epoch": 59.78170478170478, + "grad_norm": 0.00040625041583552957, + "learning_rate": 0.05480458481253893, + "loss": 0.2548, + "num_input_tokens_seen": 21930256, + "step": 28755 + }, + { + "epoch": 59.79209979209979, + "grad_norm": 0.00011704157805070281, + "learning_rate": 0.054759069821613715, + "loss": 0.2737, + "num_input_tokens_seen": 21934224, + "step": 28760 + }, + { + "epoch": 59.802494802494806, + "grad_norm": 0.0003836415708065033, + "learning_rate": 0.05471356951803683, + "loss": 0.2801, + "num_input_tokens_seen": 21937968, + "step": 28765 + }, + { + "epoch": 59.812889812889814, + "grad_norm": 7.47228114050813e-05, + "learning_rate": 0.054668083908824945, + "loss": 0.2603, + "num_input_tokens_seen": 21941872, + "step": 28770 + }, + { + "epoch": 59.82328482328482, + "grad_norm": 0.00015790094039402902, + "learning_rate": 0.054622613000992526, + "loss": 0.2675, + "num_input_tokens_seen": 21945584, + "step": 28775 + }, + { + "epoch": 59.83367983367983, + "grad_norm": 0.0001254020753549412, + "learning_rate": 0.05457715680155182, + "loss": 0.2751, + "num_input_tokens_seen": 21949360, + "step": 28780 + }, + { + "epoch": 59.84407484407485, + "grad_norm": 0.00024265359388664365, + "learning_rate": 0.05453171531751265, + "loss": 0.2487, + "num_input_tokens_seen": 21953296, + "step": 28785 + }, + { + "epoch": 59.854469854469855, + "grad_norm": 0.0003725750430021435, + "learning_rate": 0.05448628855588276, + "loss": 0.2504, + "num_input_tokens_seen": 21956912, + "step": 28790 + }, + { + "epoch": 59.86486486486486, + "grad_norm": 0.00011277251178398728, + "learning_rate": 0.05444087652366746, + "loss": 0.2559, + "num_input_tokens_seen": 21960656, + "step": 28795 + }, + { + "epoch": 59.87525987525988, + "grad_norm": 2.916383346018847e-05, + "learning_rate": 0.05439547922786984, + "loss": 0.2765, + "num_input_tokens_seen": 21964272, + "step": 28800 + }, + { + "epoch": 59.87525987525988, + "eval_loss": 0.2477722465991974, + "eval_runtime": 13.4029, + "eval_samples_per_second": 63.867, + "eval_steps_per_second": 15.967, + "num_input_tokens_seen": 21964272, + "step": 28800 + }, + { + "epoch": 59.88565488565489, + "grad_norm": 0.00024082105665002018, + "learning_rate": 0.0543500966754908, + "loss": 0.271, + "num_input_tokens_seen": 21967952, + "step": 28805 + }, + { + "epoch": 59.896049896049895, + "grad_norm": 0.00011456009087851271, + "learning_rate": 0.05430472887352882, + "loss": 0.2321, + "num_input_tokens_seen": 21971792, + "step": 28810 + }, + { + "epoch": 59.906444906444904, + "grad_norm": 0.00023738724121358246, + "learning_rate": 0.05425937582898023, + "loss": 0.2795, + "num_input_tokens_seen": 21975440, + "step": 28815 + }, + { + "epoch": 59.91683991683992, + "grad_norm": 6.13569063716568e-05, + "learning_rate": 0.054214037548839085, + "loss": 0.2535, + "num_input_tokens_seen": 21979280, + "step": 28820 + }, + { + "epoch": 59.92723492723493, + "grad_norm": 0.0003845040046144277, + "learning_rate": 0.05416871404009703, + "loss": 0.2976, + "num_input_tokens_seen": 21983088, + "step": 28825 + }, + { + "epoch": 59.937629937629936, + "grad_norm": 0.00019938072364311665, + "learning_rate": 0.054123405309743605, + "loss": 0.2625, + "num_input_tokens_seen": 21986896, + "step": 28830 + }, + { + "epoch": 59.94802494802495, + "grad_norm": 0.00038890400901436806, + "learning_rate": 0.0540781113647659, + "loss": 0.2778, + "num_input_tokens_seen": 21990736, + "step": 28835 + }, + { + "epoch": 59.95841995841996, + "grad_norm": 0.00011236857972107828, + "learning_rate": 0.054032832212148836, + "loss": 0.2729, + "num_input_tokens_seen": 21994544, + "step": 28840 + }, + { + "epoch": 59.96881496881497, + "grad_norm": 0.00034014161792583764, + "learning_rate": 0.0539875678588751, + "loss": 0.2552, + "num_input_tokens_seen": 21998288, + "step": 28845 + }, + { + "epoch": 59.979209979209976, + "grad_norm": 0.0002531029167585075, + "learning_rate": 0.05394231831192492, + "loss": 0.2591, + "num_input_tokens_seen": 22002224, + "step": 28850 + }, + { + "epoch": 59.98960498960499, + "grad_norm": 0.00017999463307205588, + "learning_rate": 0.05389708357827639, + "loss": 0.2547, + "num_input_tokens_seen": 22005904, + "step": 28855 + }, + { + "epoch": 60.0, + "grad_norm": 0.00018988997908309102, + "learning_rate": 0.05385186366490533, + "loss": 0.2652, + "num_input_tokens_seen": 22009672, + "step": 28860 + }, + { + "epoch": 60.01039501039501, + "grad_norm": 0.00035744041088037193, + "learning_rate": 0.053806658578785166, + "loss": 0.2619, + "num_input_tokens_seen": 22013448, + "step": 28865 + }, + { + "epoch": 60.020790020790024, + "grad_norm": 0.00014614209067076445, + "learning_rate": 0.05376146832688705, + "loss": 0.2467, + "num_input_tokens_seen": 22017096, + "step": 28870 + }, + { + "epoch": 60.03118503118503, + "grad_norm": 2.8996011678827927e-05, + "learning_rate": 0.053716292916179964, + "loss": 0.2277, + "num_input_tokens_seen": 22020872, + "step": 28875 + }, + { + "epoch": 60.04158004158004, + "grad_norm": 0.0001485464454162866, + "learning_rate": 0.05367113235363045, + "loss": 0.2563, + "num_input_tokens_seen": 22024840, + "step": 28880 + }, + { + "epoch": 60.05197505197505, + "grad_norm": 0.0001977020438062027, + "learning_rate": 0.05362598664620289, + "loss": 0.2673, + "num_input_tokens_seen": 22028744, + "step": 28885 + }, + { + "epoch": 60.062370062370064, + "grad_norm": 0.00017309439135715365, + "learning_rate": 0.053580855800859285, + "loss": 0.2624, + "num_input_tokens_seen": 22032584, + "step": 28890 + }, + { + "epoch": 60.07276507276507, + "grad_norm": 0.00010227455641143024, + "learning_rate": 0.05353573982455938, + "loss": 0.2797, + "num_input_tokens_seen": 22036360, + "step": 28895 + }, + { + "epoch": 60.08316008316008, + "grad_norm": 0.00014359166380017996, + "learning_rate": 0.053490638724260686, + "loss": 0.2649, + "num_input_tokens_seen": 22040104, + "step": 28900 + }, + { + "epoch": 60.093555093555096, + "grad_norm": 2.7528898499440402e-05, + "learning_rate": 0.05344555250691827, + "loss": 0.2602, + "num_input_tokens_seen": 22043912, + "step": 28905 + }, + { + "epoch": 60.103950103950105, + "grad_norm": 0.0004923850647173822, + "learning_rate": 0.053400481179485086, + "loss": 0.2711, + "num_input_tokens_seen": 22047912, + "step": 28910 + }, + { + "epoch": 60.11434511434511, + "grad_norm": 0.0004644485015887767, + "learning_rate": 0.05335542474891159, + "loss": 0.292, + "num_input_tokens_seen": 22051720, + "step": 28915 + }, + { + "epoch": 60.12474012474012, + "grad_norm": 8.596223779022694e-05, + "learning_rate": 0.053310383222146124, + "loss": 0.2754, + "num_input_tokens_seen": 22055560, + "step": 28920 + }, + { + "epoch": 60.13513513513514, + "grad_norm": 0.000365704414434731, + "learning_rate": 0.053265356606134684, + "loss": 0.2555, + "num_input_tokens_seen": 22059464, + "step": 28925 + }, + { + "epoch": 60.145530145530145, + "grad_norm": 0.00037110046832822263, + "learning_rate": 0.053220344907820856, + "loss": 0.2631, + "num_input_tokens_seen": 22063272, + "step": 28930 + }, + { + "epoch": 60.15592515592515, + "grad_norm": 0.00031425326596945524, + "learning_rate": 0.05317534813414608, + "loss": 0.2703, + "num_input_tokens_seen": 22066920, + "step": 28935 + }, + { + "epoch": 60.16632016632017, + "grad_norm": 0.000520041212439537, + "learning_rate": 0.05313036629204942, + "loss": 0.2535, + "num_input_tokens_seen": 22070792, + "step": 28940 + }, + { + "epoch": 60.17671517671518, + "grad_norm": 0.00011217831342946738, + "learning_rate": 0.05308539938846756, + "loss": 0.2654, + "num_input_tokens_seen": 22074632, + "step": 28945 + }, + { + "epoch": 60.187110187110186, + "grad_norm": 9.818014950724319e-05, + "learning_rate": 0.05304044743033507, + "loss": 0.2453, + "num_input_tokens_seen": 22078440, + "step": 28950 + }, + { + "epoch": 60.197505197505194, + "grad_norm": 0.0002047102025244385, + "learning_rate": 0.05299551042458401, + "loss": 0.2686, + "num_input_tokens_seen": 22082216, + "step": 28955 + }, + { + "epoch": 60.20790020790021, + "grad_norm": 0.0004223485302645713, + "learning_rate": 0.052950588378144266, + "loss": 0.2625, + "num_input_tokens_seen": 22086024, + "step": 28960 + }, + { + "epoch": 60.21829521829522, + "grad_norm": 0.0001951005106093362, + "learning_rate": 0.052905681297943465, + "loss": 0.2812, + "num_input_tokens_seen": 22089992, + "step": 28965 + }, + { + "epoch": 60.228690228690226, + "grad_norm": 5.571204383159056e-05, + "learning_rate": 0.0528607891909067, + "loss": 0.2885, + "num_input_tokens_seen": 22093736, + "step": 28970 + }, + { + "epoch": 60.23908523908524, + "grad_norm": 0.00020731370022986084, + "learning_rate": 0.05281591206395697, + "loss": 0.2638, + "num_input_tokens_seen": 22097576, + "step": 28975 + }, + { + "epoch": 60.24948024948025, + "grad_norm": 0.0002497128152754158, + "learning_rate": 0.05277104992401496, + "loss": 0.2743, + "num_input_tokens_seen": 22101416, + "step": 28980 + }, + { + "epoch": 60.25987525987526, + "grad_norm": 0.0001987690629903227, + "learning_rate": 0.05272620277799884, + "loss": 0.2631, + "num_input_tokens_seen": 22105256, + "step": 28985 + }, + { + "epoch": 60.270270270270274, + "grad_norm": 0.0004770920204464346, + "learning_rate": 0.05268137063282473, + "loss": 0.2732, + "num_input_tokens_seen": 22109064, + "step": 28990 + }, + { + "epoch": 60.28066528066528, + "grad_norm": 0.00012539415911305696, + "learning_rate": 0.0526365534954062, + "loss": 0.2685, + "num_input_tokens_seen": 22112808, + "step": 28995 + }, + { + "epoch": 60.29106029106029, + "grad_norm": 0.00015280838124454021, + "learning_rate": 0.052591751372654656, + "loss": 0.2601, + "num_input_tokens_seen": 22116648, + "step": 29000 + }, + { + "epoch": 60.29106029106029, + "eval_loss": 0.25114792585372925, + "eval_runtime": 13.3985, + "eval_samples_per_second": 63.888, + "eval_steps_per_second": 15.972, + "num_input_tokens_seen": 22116648, + "step": 29000 + }, + { + "epoch": 60.3014553014553, + "grad_norm": 0.00017448792641516775, + "learning_rate": 0.05254696427147921, + "loss": 0.2777, + "num_input_tokens_seen": 22120424, + "step": 29005 + }, + { + "epoch": 60.311850311850314, + "grad_norm": 0.0002914345823228359, + "learning_rate": 0.052502192198786546, + "loss": 0.2475, + "num_input_tokens_seen": 22124232, + "step": 29010 + }, + { + "epoch": 60.32224532224532, + "grad_norm": 0.0006505593191832304, + "learning_rate": 0.05245743516148103, + "loss": 0.226, + "num_input_tokens_seen": 22128040, + "step": 29015 + }, + { + "epoch": 60.33264033264033, + "grad_norm": 0.00016514386516064405, + "learning_rate": 0.05241269316646486, + "loss": 0.275, + "num_input_tokens_seen": 22131848, + "step": 29020 + }, + { + "epoch": 60.343035343035346, + "grad_norm": 0.0004972768947482109, + "learning_rate": 0.052367966220637725, + "loss": 0.2525, + "num_input_tokens_seen": 22135400, + "step": 29025 + }, + { + "epoch": 60.353430353430355, + "grad_norm": 0.00012201243225717917, + "learning_rate": 0.05232325433089716, + "loss": 0.2541, + "num_input_tokens_seen": 22139144, + "step": 29030 + }, + { + "epoch": 60.36382536382536, + "grad_norm": 0.00014078826643526554, + "learning_rate": 0.052278557504138214, + "loss": 0.2643, + "num_input_tokens_seen": 22142920, + "step": 29035 + }, + { + "epoch": 60.37422037422037, + "grad_norm": 0.0003011383814737201, + "learning_rate": 0.05223387574725372, + "loss": 0.2563, + "num_input_tokens_seen": 22146792, + "step": 29040 + }, + { + "epoch": 60.38461538461539, + "grad_norm": 5.9105397667735815e-05, + "learning_rate": 0.05218920906713428, + "loss": 0.2528, + "num_input_tokens_seen": 22150632, + "step": 29045 + }, + { + "epoch": 60.395010395010395, + "grad_norm": 0.00043190570431761444, + "learning_rate": 0.05214455747066789, + "loss": 0.2856, + "num_input_tokens_seen": 22154568, + "step": 29050 + }, + { + "epoch": 60.4054054054054, + "grad_norm": 0.00024825692526064813, + "learning_rate": 0.05209992096474048, + "loss": 0.2477, + "num_input_tokens_seen": 22158344, + "step": 29055 + }, + { + "epoch": 60.41580041580042, + "grad_norm": 0.00041552301263436675, + "learning_rate": 0.05205529955623559, + "loss": 0.2677, + "num_input_tokens_seen": 22162088, + "step": 29060 + }, + { + "epoch": 60.42619542619543, + "grad_norm": 4.142362013226375e-05, + "learning_rate": 0.052010693252034314, + "loss": 0.2763, + "num_input_tokens_seen": 22165800, + "step": 29065 + }, + { + "epoch": 60.436590436590436, + "grad_norm": 0.00044333390542306006, + "learning_rate": 0.0519661020590156, + "loss": 0.2383, + "num_input_tokens_seen": 22169672, + "step": 29070 + }, + { + "epoch": 60.446985446985444, + "grad_norm": 0.00011101623385911807, + "learning_rate": 0.05192152598405586, + "loss": 0.2781, + "num_input_tokens_seen": 22173448, + "step": 29075 + }, + { + "epoch": 60.45738045738046, + "grad_norm": 0.0001243502920260653, + "learning_rate": 0.05187696503402941, + "loss": 0.2607, + "num_input_tokens_seen": 22177320, + "step": 29080 + }, + { + "epoch": 60.46777546777547, + "grad_norm": 0.0005695197614841163, + "learning_rate": 0.05183241921580798, + "loss": 0.2731, + "num_input_tokens_seen": 22181064, + "step": 29085 + }, + { + "epoch": 60.478170478170476, + "grad_norm": 0.00020992530335206538, + "learning_rate": 0.051787888536261206, + "loss": 0.2679, + "num_input_tokens_seen": 22184936, + "step": 29090 + }, + { + "epoch": 60.48856548856549, + "grad_norm": 0.0002746711252257228, + "learning_rate": 0.051743373002256184, + "loss": 0.2292, + "num_input_tokens_seen": 22188712, + "step": 29095 + }, + { + "epoch": 60.4989604989605, + "grad_norm": 0.00014733635180164129, + "learning_rate": 0.05169887262065787, + "loss": 0.2229, + "num_input_tokens_seen": 22192456, + "step": 29100 + }, + { + "epoch": 60.50935550935551, + "grad_norm": 0.00013522559311240911, + "learning_rate": 0.051654387398328665, + "loss": 0.2788, + "num_input_tokens_seen": 22196264, + "step": 29105 + }, + { + "epoch": 60.51975051975052, + "grad_norm": 6.926549394847825e-05, + "learning_rate": 0.05160991734212888, + "loss": 0.2841, + "num_input_tokens_seen": 22200040, + "step": 29110 + }, + { + "epoch": 60.53014553014553, + "grad_norm": 0.00014191550144460052, + "learning_rate": 0.051565462458916224, + "loss": 0.2492, + "num_input_tokens_seen": 22203848, + "step": 29115 + }, + { + "epoch": 60.54054054054054, + "grad_norm": 0.00022297805116977543, + "learning_rate": 0.05152102275554627, + "loss": 0.2489, + "num_input_tokens_seen": 22207752, + "step": 29120 + }, + { + "epoch": 60.55093555093555, + "grad_norm": 0.00021488354832399637, + "learning_rate": 0.05147659823887222, + "loss": 0.2953, + "num_input_tokens_seen": 22211592, + "step": 29125 + }, + { + "epoch": 60.561330561330564, + "grad_norm": 0.00020440088701434433, + "learning_rate": 0.05143218891574479, + "loss": 0.2447, + "num_input_tokens_seen": 22215464, + "step": 29130 + }, + { + "epoch": 60.57172557172557, + "grad_norm": 0.00023655552649870515, + "learning_rate": 0.0513877947930125, + "loss": 0.2432, + "num_input_tokens_seen": 22219368, + "step": 29135 + }, + { + "epoch": 60.58212058212058, + "grad_norm": 0.0003324352437630296, + "learning_rate": 0.051343415877521566, + "loss": 0.2377, + "num_input_tokens_seen": 22223080, + "step": 29140 + }, + { + "epoch": 60.59251559251559, + "grad_norm": 6.227132689673454e-05, + "learning_rate": 0.051299052176115634, + "loss": 0.2612, + "num_input_tokens_seen": 22226952, + "step": 29145 + }, + { + "epoch": 60.602910602910605, + "grad_norm": 0.0003988379903603345, + "learning_rate": 0.051254703695636256, + "loss": 0.2714, + "num_input_tokens_seen": 22230824, + "step": 29150 + }, + { + "epoch": 60.61330561330561, + "grad_norm": 0.0001947104901773855, + "learning_rate": 0.05121037044292249, + "loss": 0.2706, + "num_input_tokens_seen": 22234632, + "step": 29155 + }, + { + "epoch": 60.62370062370062, + "grad_norm": 0.0004076639888808131, + "learning_rate": 0.05116605242481101, + "loss": 0.2605, + "num_input_tokens_seen": 22238504, + "step": 29160 + }, + { + "epoch": 60.63409563409564, + "grad_norm": 0.0001246178726432845, + "learning_rate": 0.05112174964813634, + "loss": 0.2548, + "num_input_tokens_seen": 22242376, + "step": 29165 + }, + { + "epoch": 60.644490644490645, + "grad_norm": 7.189393363660201e-05, + "learning_rate": 0.05107746211973038, + "loss": 0.2713, + "num_input_tokens_seen": 22246120, + "step": 29170 + }, + { + "epoch": 60.65488565488565, + "grad_norm": 0.00023302558111026883, + "learning_rate": 0.05103318984642291, + "loss": 0.253, + "num_input_tokens_seen": 22249832, + "step": 29175 + }, + { + "epoch": 60.66528066528066, + "grad_norm": 0.00040809568599797785, + "learning_rate": 0.05098893283504131, + "loss": 0.2791, + "num_input_tokens_seen": 22253576, + "step": 29180 + }, + { + "epoch": 60.67567567567568, + "grad_norm": 0.00010664543515304103, + "learning_rate": 0.050944691092410475, + "loss": 0.2544, + "num_input_tokens_seen": 22257288, + "step": 29185 + }, + { + "epoch": 60.686070686070686, + "grad_norm": 0.00043277291115373373, + "learning_rate": 0.05090046462535313, + "loss": 0.2703, + "num_input_tokens_seen": 22261192, + "step": 29190 + }, + { + "epoch": 60.696465696465694, + "grad_norm": 0.0006072860560379922, + "learning_rate": 0.050856253440689454, + "loss": 0.2733, + "num_input_tokens_seen": 22265160, + "step": 29195 + }, + { + "epoch": 60.70686070686071, + "grad_norm": 0.00015070570225361735, + "learning_rate": 0.050812057545237405, + "loss": 0.2744, + "num_input_tokens_seen": 22269032, + "step": 29200 + }, + { + "epoch": 60.70686070686071, + "eval_loss": 0.2513991594314575, + "eval_runtime": 13.386, + "eval_samples_per_second": 63.948, + "eval_steps_per_second": 15.987, + "num_input_tokens_seen": 22269032, + "step": 29200 + }, + { + "epoch": 60.71725571725572, + "grad_norm": 0.00026994626387022436, + "learning_rate": 0.0507678769458126, + "loss": 0.2702, + "num_input_tokens_seen": 22272808, + "step": 29205 + }, + { + "epoch": 60.727650727650726, + "grad_norm": 0.0002789971185848117, + "learning_rate": 0.050723711649228155, + "loss": 0.2669, + "num_input_tokens_seen": 22276648, + "step": 29210 + }, + { + "epoch": 60.73804573804574, + "grad_norm": 0.0003134275902993977, + "learning_rate": 0.05067956166229496, + "loss": 0.2711, + "num_input_tokens_seen": 22280392, + "step": 29215 + }, + { + "epoch": 60.74844074844075, + "grad_norm": 3.340885086799972e-05, + "learning_rate": 0.05063542699182155, + "loss": 0.2607, + "num_input_tokens_seen": 22284104, + "step": 29220 + }, + { + "epoch": 60.75883575883576, + "grad_norm": 0.0002111646463163197, + "learning_rate": 0.050591307644613996, + "loss": 0.2674, + "num_input_tokens_seen": 22288168, + "step": 29225 + }, + { + "epoch": 60.76923076923077, + "grad_norm": 0.00032830823329277337, + "learning_rate": 0.05054720362747599, + "loss": 0.2827, + "num_input_tokens_seen": 22291976, + "step": 29230 + }, + { + "epoch": 60.77962577962578, + "grad_norm": 0.0002581693115644157, + "learning_rate": 0.050503114947209035, + "loss": 0.2701, + "num_input_tokens_seen": 22295816, + "step": 29235 + }, + { + "epoch": 60.79002079002079, + "grad_norm": 0.00022006980725564063, + "learning_rate": 0.05045904161061207, + "loss": 0.2655, + "num_input_tokens_seen": 22299752, + "step": 29240 + }, + { + "epoch": 60.8004158004158, + "grad_norm": 0.00016812840476632118, + "learning_rate": 0.05041498362448185, + "loss": 0.2621, + "num_input_tokens_seen": 22303688, + "step": 29245 + }, + { + "epoch": 60.810810810810814, + "grad_norm": 0.0008603687747381628, + "learning_rate": 0.05037094099561256, + "loss": 0.2753, + "num_input_tokens_seen": 22307560, + "step": 29250 + }, + { + "epoch": 60.82120582120582, + "grad_norm": 0.00018224988889414817, + "learning_rate": 0.05032691373079624, + "loss": 0.283, + "num_input_tokens_seen": 22311464, + "step": 29255 + }, + { + "epoch": 60.83160083160083, + "grad_norm": 0.00019357199198566377, + "learning_rate": 0.05028290183682234, + "loss": 0.2703, + "num_input_tokens_seen": 22315304, + "step": 29260 + }, + { + "epoch": 60.84199584199584, + "grad_norm": 0.0003141420893371105, + "learning_rate": 0.050238905320478096, + "loss": 0.2632, + "num_input_tokens_seen": 22319112, + "step": 29265 + }, + { + "epoch": 60.852390852390855, + "grad_norm": 0.0001770677336025983, + "learning_rate": 0.05019492418854838, + "loss": 0.279, + "num_input_tokens_seen": 22322920, + "step": 29270 + }, + { + "epoch": 60.86278586278586, + "grad_norm": 0.0004777581780217588, + "learning_rate": 0.05015095844781554, + "loss": 0.277, + "num_input_tokens_seen": 22326824, + "step": 29275 + }, + { + "epoch": 60.87318087318087, + "grad_norm": 0.000332927011186257, + "learning_rate": 0.05010700810505968, + "loss": 0.285, + "num_input_tokens_seen": 22330664, + "step": 29280 + }, + { + "epoch": 60.88357588357589, + "grad_norm": 6.426603067666292e-05, + "learning_rate": 0.05006307316705856, + "loss": 0.2695, + "num_input_tokens_seen": 22334376, + "step": 29285 + }, + { + "epoch": 60.893970893970895, + "grad_norm": 0.000273017561994493, + "learning_rate": 0.0500191536405874, + "loss": 0.2716, + "num_input_tokens_seen": 22338280, + "step": 29290 + }, + { + "epoch": 60.9043659043659, + "grad_norm": 0.00046285157441161573, + "learning_rate": 0.04997524953241922, + "loss": 0.2684, + "num_input_tokens_seen": 22342344, + "step": 29295 + }, + { + "epoch": 60.91476091476091, + "grad_norm": 0.0001919929782161489, + "learning_rate": 0.049931360849324556, + "loss": 0.2505, + "num_input_tokens_seen": 22346088, + "step": 29300 + }, + { + "epoch": 60.92515592515593, + "grad_norm": 5.1948973123217e-05, + "learning_rate": 0.04988748759807155, + "loss": 0.2652, + "num_input_tokens_seen": 22349736, + "step": 29305 + }, + { + "epoch": 60.935550935550935, + "grad_norm": 6.86116109136492e-05, + "learning_rate": 0.0498436297854261, + "loss": 0.2388, + "num_input_tokens_seen": 22353480, + "step": 29310 + }, + { + "epoch": 60.945945945945944, + "grad_norm": 0.00023443944519385695, + "learning_rate": 0.04979978741815152, + "loss": 0.2489, + "num_input_tokens_seen": 22357096, + "step": 29315 + }, + { + "epoch": 60.95634095634096, + "grad_norm": 4.803876072401181e-05, + "learning_rate": 0.04975596050300891, + "loss": 0.2546, + "num_input_tokens_seen": 22360840, + "step": 29320 + }, + { + "epoch": 60.96673596673597, + "grad_norm": 0.0002800797810778022, + "learning_rate": 0.049712149046757005, + "loss": 0.2714, + "num_input_tokens_seen": 22364872, + "step": 29325 + }, + { + "epoch": 60.977130977130976, + "grad_norm": 6.997302261879668e-05, + "learning_rate": 0.04966835305615194, + "loss": 0.2531, + "num_input_tokens_seen": 22368680, + "step": 29330 + }, + { + "epoch": 60.987525987525984, + "grad_norm": 0.00043051608372479677, + "learning_rate": 0.049624572537947755, + "loss": 0.2848, + "num_input_tokens_seen": 22372520, + "step": 29335 + }, + { + "epoch": 60.997920997921, + "grad_norm": 0.00014155283861327916, + "learning_rate": 0.04958080749889582, + "loss": 0.264, + "num_input_tokens_seen": 22376264, + "step": 29340 + }, + { + "epoch": 61.00831600831601, + "grad_norm": 0.00045308491098694503, + "learning_rate": 0.049537057945745304, + "loss": 0.2402, + "num_input_tokens_seen": 22380056, + "step": 29345 + }, + { + "epoch": 61.018711018711016, + "grad_norm": 8.39843341964297e-05, + "learning_rate": 0.049493323885243, + "loss": 0.2631, + "num_input_tokens_seen": 22383960, + "step": 29350 + }, + { + "epoch": 61.02910602910603, + "grad_norm": 6.288572330959141e-05, + "learning_rate": 0.04944960532413318, + "loss": 0.2887, + "num_input_tokens_seen": 22387768, + "step": 29355 + }, + { + "epoch": 61.03950103950104, + "grad_norm": 0.00014656568237114698, + "learning_rate": 0.049405902269157774, + "loss": 0.2599, + "num_input_tokens_seen": 22391576, + "step": 29360 + }, + { + "epoch": 61.04989604989605, + "grad_norm": 0.00013488282274920493, + "learning_rate": 0.04936221472705646, + "loss": 0.2734, + "num_input_tokens_seen": 22395320, + "step": 29365 + }, + { + "epoch": 61.06029106029106, + "grad_norm": 4.0780701965559274e-05, + "learning_rate": 0.04931854270456632, + "loss": 0.2569, + "num_input_tokens_seen": 22399192, + "step": 29370 + }, + { + "epoch": 61.07068607068607, + "grad_norm": 0.00010651641787262633, + "learning_rate": 0.049274886208422075, + "loss": 0.2714, + "num_input_tokens_seen": 22402968, + "step": 29375 + }, + { + "epoch": 61.08108108108108, + "grad_norm": 0.00022704314324073493, + "learning_rate": 0.049231245245356235, + "loss": 0.2363, + "num_input_tokens_seen": 22406712, + "step": 29380 + }, + { + "epoch": 61.09147609147609, + "grad_norm": 0.00042226497316733, + "learning_rate": 0.049187619822098655, + "loss": 0.2573, + "num_input_tokens_seen": 22410616, + "step": 29385 + }, + { + "epoch": 61.101871101871104, + "grad_norm": 0.00012956962746102363, + "learning_rate": 0.04914400994537705, + "loss": 0.2667, + "num_input_tokens_seen": 22414232, + "step": 29390 + }, + { + "epoch": 61.11226611226611, + "grad_norm": 0.00024197938910219818, + "learning_rate": 0.049100415621916485, + "loss": 0.2783, + "num_input_tokens_seen": 22418104, + "step": 29395 + }, + { + "epoch": 61.12266112266112, + "grad_norm": 0.00019310711650177836, + "learning_rate": 0.04905683685843981, + "loss": 0.2519, + "num_input_tokens_seen": 22421944, + "step": 29400 + }, + { + "epoch": 61.12266112266112, + "eval_loss": 0.24778404831886292, + "eval_runtime": 13.4094, + "eval_samples_per_second": 63.836, + "eval_steps_per_second": 15.959, + "num_input_tokens_seen": 22421944, + "step": 29400 + }, + { + "epoch": 61.13305613305613, + "grad_norm": 0.0006019994616508484, + "learning_rate": 0.049013273661667495, + "loss": 0.255, + "num_input_tokens_seen": 22425688, + "step": 29405 + }, + { + "epoch": 61.143451143451145, + "grad_norm": 0.0001796743308659643, + "learning_rate": 0.048969726038317396, + "loss": 0.2627, + "num_input_tokens_seen": 22429592, + "step": 29410 + }, + { + "epoch": 61.15384615384615, + "grad_norm": 7.682923751417547e-05, + "learning_rate": 0.048926193995105206, + "loss": 0.2624, + "num_input_tokens_seen": 22433432, + "step": 29415 + }, + { + "epoch": 61.16424116424116, + "grad_norm": 0.00025951353018172085, + "learning_rate": 0.048882677538744035, + "loss": 0.2767, + "num_input_tokens_seen": 22437176, + "step": 29420 + }, + { + "epoch": 61.17463617463618, + "grad_norm": 0.00020866339036729187, + "learning_rate": 0.048839176675944715, + "loss": 0.2593, + "num_input_tokens_seen": 22441048, + "step": 29425 + }, + { + "epoch": 61.185031185031185, + "grad_norm": 0.00026405375683680177, + "learning_rate": 0.04879569141341566, + "loss": 0.2817, + "num_input_tokens_seen": 22444856, + "step": 29430 + }, + { + "epoch": 61.195426195426194, + "grad_norm": 6.615067832171917e-05, + "learning_rate": 0.04875222175786274, + "loss": 0.2646, + "num_input_tokens_seen": 22448632, + "step": 29435 + }, + { + "epoch": 61.20582120582121, + "grad_norm": 0.00015455170068889856, + "learning_rate": 0.04870876771598966, + "loss": 0.2503, + "num_input_tokens_seen": 22452440, + "step": 29440 + }, + { + "epoch": 61.21621621621622, + "grad_norm": 0.0003888489445671439, + "learning_rate": 0.04866532929449744, + "loss": 0.2533, + "num_input_tokens_seen": 22456248, + "step": 29445 + }, + { + "epoch": 61.226611226611226, + "grad_norm": 0.0006923777400515974, + "learning_rate": 0.048621906500084945, + "loss": 0.2772, + "num_input_tokens_seen": 22460024, + "step": 29450 + }, + { + "epoch": 61.237006237006234, + "grad_norm": 0.0002870278840418905, + "learning_rate": 0.04857849933944845, + "loss": 0.2493, + "num_input_tokens_seen": 22463928, + "step": 29455 + }, + { + "epoch": 61.24740124740125, + "grad_norm": 0.00015916672418825328, + "learning_rate": 0.048535107819281866, + "loss": 0.2473, + "num_input_tokens_seen": 22467800, + "step": 29460 + }, + { + "epoch": 61.25779625779626, + "grad_norm": 0.00030887455795891583, + "learning_rate": 0.04849173194627675, + "loss": 0.2352, + "num_input_tokens_seen": 22471672, + "step": 29465 + }, + { + "epoch": 61.268191268191266, + "grad_norm": 0.00023358812904916704, + "learning_rate": 0.04844837172712223, + "loss": 0.2813, + "num_input_tokens_seen": 22475544, + "step": 29470 + }, + { + "epoch": 61.27858627858628, + "grad_norm": 0.0003031744563486427, + "learning_rate": 0.04840502716850494, + "loss": 0.2604, + "num_input_tokens_seen": 22479576, + "step": 29475 + }, + { + "epoch": 61.28898128898129, + "grad_norm": 0.0006664341781288385, + "learning_rate": 0.04836169827710916, + "loss": 0.2707, + "num_input_tokens_seen": 22483480, + "step": 29480 + }, + { + "epoch": 61.2993762993763, + "grad_norm": 0.0001676773972576484, + "learning_rate": 0.04831838505961684, + "loss": 0.2679, + "num_input_tokens_seen": 22487320, + "step": 29485 + }, + { + "epoch": 61.30977130977131, + "grad_norm": 0.0006817649118602276, + "learning_rate": 0.048275087522707295, + "loss": 0.2381, + "num_input_tokens_seen": 22491032, + "step": 29490 + }, + { + "epoch": 61.32016632016632, + "grad_norm": 0.0005874313064850867, + "learning_rate": 0.04823180567305766, + "loss": 0.2774, + "num_input_tokens_seen": 22495032, + "step": 29495 + }, + { + "epoch": 61.33056133056133, + "grad_norm": 0.0006164637161418796, + "learning_rate": 0.04818853951734244, + "loss": 0.2798, + "num_input_tokens_seen": 22498808, + "step": 29500 + }, + { + "epoch": 61.34095634095634, + "grad_norm": 0.00037387784686870873, + "learning_rate": 0.04814528906223387, + "loss": 0.2458, + "num_input_tokens_seen": 22502552, + "step": 29505 + }, + { + "epoch": 61.351351351351354, + "grad_norm": 8.99830847629346e-05, + "learning_rate": 0.04810205431440177, + "loss": 0.2581, + "num_input_tokens_seen": 22506392, + "step": 29510 + }, + { + "epoch": 61.36174636174636, + "grad_norm": 0.00037280903779901564, + "learning_rate": 0.04805883528051341, + "loss": 0.2564, + "num_input_tokens_seen": 22510200, + "step": 29515 + }, + { + "epoch": 61.37214137214137, + "grad_norm": 0.00019036447338294238, + "learning_rate": 0.048015631967233685, + "loss": 0.2753, + "num_input_tokens_seen": 22514136, + "step": 29520 + }, + { + "epoch": 61.38253638253638, + "grad_norm": 0.00044575455831363797, + "learning_rate": 0.04797244438122517, + "loss": 0.2634, + "num_input_tokens_seen": 22517816, + "step": 29525 + }, + { + "epoch": 61.392931392931395, + "grad_norm": 0.00013188677257858217, + "learning_rate": 0.04792927252914784, + "loss": 0.2696, + "num_input_tokens_seen": 22521816, + "step": 29530 + }, + { + "epoch": 61.4033264033264, + "grad_norm": 8.169235661625862e-05, + "learning_rate": 0.04788611641765944, + "loss": 0.2618, + "num_input_tokens_seen": 22525464, + "step": 29535 + }, + { + "epoch": 61.41372141372141, + "grad_norm": 0.00018782488768920302, + "learning_rate": 0.04784297605341508, + "loss": 0.2537, + "num_input_tokens_seen": 22529144, + "step": 29540 + }, + { + "epoch": 61.42411642411643, + "grad_norm": 0.00044537067878991365, + "learning_rate": 0.04779985144306761, + "loss": 0.2722, + "num_input_tokens_seen": 22533016, + "step": 29545 + }, + { + "epoch": 61.434511434511435, + "grad_norm": 0.00034549308475106955, + "learning_rate": 0.047756742593267405, + "loss": 0.264, + "num_input_tokens_seen": 22536792, + "step": 29550 + }, + { + "epoch": 61.444906444906444, + "grad_norm": 0.0002115590323228389, + "learning_rate": 0.047713649510662315, + "loss": 0.2627, + "num_input_tokens_seen": 22540696, + "step": 29555 + }, + { + "epoch": 61.45530145530145, + "grad_norm": 0.00027155931456945837, + "learning_rate": 0.04767057220189789, + "loss": 0.2658, + "num_input_tokens_seen": 22544504, + "step": 29560 + }, + { + "epoch": 61.46569646569647, + "grad_norm": 7.145785639295354e-05, + "learning_rate": 0.04762751067361722, + "loss": 0.2651, + "num_input_tokens_seen": 22548312, + "step": 29565 + }, + { + "epoch": 61.476091476091476, + "grad_norm": 0.00030012940987944603, + "learning_rate": 0.04758446493246086, + "loss": 0.2729, + "num_input_tokens_seen": 22552056, + "step": 29570 + }, + { + "epoch": 61.486486486486484, + "grad_norm": 0.00046009692596271634, + "learning_rate": 0.047541434985067084, + "loss": 0.2666, + "num_input_tokens_seen": 22555896, + "step": 29575 + }, + { + "epoch": 61.4968814968815, + "grad_norm": 4.3458341679070145e-05, + "learning_rate": 0.047498420838071556, + "loss": 0.2759, + "num_input_tokens_seen": 22559896, + "step": 29580 + }, + { + "epoch": 61.50727650727651, + "grad_norm": 7.798478327458724e-05, + "learning_rate": 0.04745542249810772, + "loss": 0.271, + "num_input_tokens_seen": 22563704, + "step": 29585 + }, + { + "epoch": 61.517671517671516, + "grad_norm": 6.683883111691102e-05, + "learning_rate": 0.047412439971806324, + "loss": 0.2674, + "num_input_tokens_seen": 22567512, + "step": 29590 + }, + { + "epoch": 61.528066528066525, + "grad_norm": 0.00041826171218417585, + "learning_rate": 0.04736947326579592, + "loss": 0.2721, + "num_input_tokens_seen": 22571352, + "step": 29595 + }, + { + "epoch": 61.53846153846154, + "grad_norm": 0.0002417449577478692, + "learning_rate": 0.04732652238670245, + "loss": 0.2741, + "num_input_tokens_seen": 22574936, + "step": 29600 + }, + { + "epoch": 61.53846153846154, + "eval_loss": 0.25261032581329346, + "eval_runtime": 13.3994, + "eval_samples_per_second": 63.884, + "eval_steps_per_second": 15.971, + "num_input_tokens_seen": 22574936, + "step": 29600 + }, + { + "epoch": 61.54885654885655, + "grad_norm": 5.4587908380199224e-05, + "learning_rate": 0.04728358734114952, + "loss": 0.2627, + "num_input_tokens_seen": 22578552, + "step": 29605 + }, + { + "epoch": 61.55925155925156, + "grad_norm": 0.0004376430588308722, + "learning_rate": 0.04724066813575821, + "loss": 0.288, + "num_input_tokens_seen": 22582360, + "step": 29610 + }, + { + "epoch": 61.56964656964657, + "grad_norm": 0.00014969435869716108, + "learning_rate": 0.04719776477714729, + "loss": 0.27, + "num_input_tokens_seen": 22586104, + "step": 29615 + }, + { + "epoch": 61.58004158004158, + "grad_norm": 0.00023263019102159888, + "learning_rate": 0.047154877271932856, + "loss": 0.2659, + "num_input_tokens_seen": 22589944, + "step": 29620 + }, + { + "epoch": 61.59043659043659, + "grad_norm": 0.0001864308287622407, + "learning_rate": 0.0471120056267288, + "loss": 0.2723, + "num_input_tokens_seen": 22593720, + "step": 29625 + }, + { + "epoch": 61.6008316008316, + "grad_norm": 0.00021168560488149524, + "learning_rate": 0.047069149848146495, + "loss": 0.2719, + "num_input_tokens_seen": 22597496, + "step": 29630 + }, + { + "epoch": 61.61122661122661, + "grad_norm": 0.00044054165482521057, + "learning_rate": 0.04702630994279473, + "loss": 0.2655, + "num_input_tokens_seen": 22601336, + "step": 29635 + }, + { + "epoch": 61.62162162162162, + "grad_norm": 0.0003041196323465556, + "learning_rate": 0.046983485917280035, + "loss": 0.2588, + "num_input_tokens_seen": 22605176, + "step": 29640 + }, + { + "epoch": 61.63201663201663, + "grad_norm": 0.00019611175230238587, + "learning_rate": 0.04694067777820644, + "loss": 0.2717, + "num_input_tokens_seen": 22608984, + "step": 29645 + }, + { + "epoch": 61.642411642411645, + "grad_norm": 0.0003394393716007471, + "learning_rate": 0.046897885532175415, + "loss": 0.2557, + "num_input_tokens_seen": 22612728, + "step": 29650 + }, + { + "epoch": 61.65280665280665, + "grad_norm": 0.0002699804899748415, + "learning_rate": 0.04685510918578613, + "loss": 0.2739, + "num_input_tokens_seen": 22616504, + "step": 29655 + }, + { + "epoch": 61.66320166320166, + "grad_norm": 0.0001857308525359258, + "learning_rate": 0.04681234874563519, + "loss": 0.2469, + "num_input_tokens_seen": 22620312, + "step": 29660 + }, + { + "epoch": 61.67359667359668, + "grad_norm": 0.0004041350621264428, + "learning_rate": 0.046769604218316836, + "loss": 0.264, + "num_input_tokens_seen": 22624216, + "step": 29665 + }, + { + "epoch": 61.683991683991685, + "grad_norm": 0.0004000420158263296, + "learning_rate": 0.04672687561042279, + "loss": 0.2675, + "num_input_tokens_seen": 22627896, + "step": 29670 + }, + { + "epoch": 61.694386694386694, + "grad_norm": 0.0004125539562664926, + "learning_rate": 0.046684162928542286, + "loss": 0.251, + "num_input_tokens_seen": 22631608, + "step": 29675 + }, + { + "epoch": 61.7047817047817, + "grad_norm": 0.00015640696801710874, + "learning_rate": 0.04664146617926222, + "loss": 0.2665, + "num_input_tokens_seen": 22635448, + "step": 29680 + }, + { + "epoch": 61.71517671517672, + "grad_norm": 0.00010408647358417511, + "learning_rate": 0.046598785369167, + "loss": 0.2477, + "num_input_tokens_seen": 22639224, + "step": 29685 + }, + { + "epoch": 61.725571725571726, + "grad_norm": 0.00019091191643383354, + "learning_rate": 0.046556120504838434, + "loss": 0.2802, + "num_input_tokens_seen": 22642968, + "step": 29690 + }, + { + "epoch": 61.735966735966734, + "grad_norm": 0.0005145528120920062, + "learning_rate": 0.04651347159285609, + "loss": 0.2584, + "num_input_tokens_seen": 22646712, + "step": 29695 + }, + { + "epoch": 61.74636174636175, + "grad_norm": 0.0005875641363672912, + "learning_rate": 0.04647083863979688, + "loss": 0.2512, + "num_input_tokens_seen": 22650520, + "step": 29700 + }, + { + "epoch": 61.75675675675676, + "grad_norm": 9.322630648966879e-05, + "learning_rate": 0.04642822165223538, + "loss": 0.2665, + "num_input_tokens_seen": 22654456, + "step": 29705 + }, + { + "epoch": 61.767151767151766, + "grad_norm": 0.00034378416603431106, + "learning_rate": 0.046385620636743716, + "loss": 0.277, + "num_input_tokens_seen": 22658328, + "step": 29710 + }, + { + "epoch": 61.777546777546775, + "grad_norm": 0.00016907391545828432, + "learning_rate": 0.04634303559989141, + "loss": 0.2636, + "num_input_tokens_seen": 22661976, + "step": 29715 + }, + { + "epoch": 61.78794178794179, + "grad_norm": 7.173990161390975e-05, + "learning_rate": 0.046300466548245635, + "loss": 0.2509, + "num_input_tokens_seen": 22665784, + "step": 29720 + }, + { + "epoch": 61.7983367983368, + "grad_norm": 0.0004067953850608319, + "learning_rate": 0.04625791348837114, + "loss": 0.2654, + "num_input_tokens_seen": 22669720, + "step": 29725 + }, + { + "epoch": 61.80873180873181, + "grad_norm": 0.00038648044574074447, + "learning_rate": 0.046215376426830095, + "loss": 0.2493, + "num_input_tokens_seen": 22673720, + "step": 29730 + }, + { + "epoch": 61.81912681912682, + "grad_norm": 0.000145882906508632, + "learning_rate": 0.04617285537018219, + "loss": 0.2703, + "num_input_tokens_seen": 22677464, + "step": 29735 + }, + { + "epoch": 61.82952182952183, + "grad_norm": 0.00019403253099881113, + "learning_rate": 0.046130350324984803, + "loss": 0.2508, + "num_input_tokens_seen": 22681336, + "step": 29740 + }, + { + "epoch": 61.83991683991684, + "grad_norm": 7.360886229434982e-05, + "learning_rate": 0.046087861297792666, + "loss": 0.2627, + "num_input_tokens_seen": 22685016, + "step": 29745 + }, + { + "epoch": 61.85031185031185, + "grad_norm": 0.0003178084152750671, + "learning_rate": 0.0460453882951582, + "loss": 0.2592, + "num_input_tokens_seen": 22688952, + "step": 29750 + }, + { + "epoch": 61.86070686070686, + "grad_norm": 0.00014706020010635257, + "learning_rate": 0.04600293132363119, + "loss": 0.2763, + "num_input_tokens_seen": 22692632, + "step": 29755 + }, + { + "epoch": 61.87110187110187, + "grad_norm": 0.00011950283078476787, + "learning_rate": 0.045960490389759086, + "loss": 0.2807, + "num_input_tokens_seen": 22696504, + "step": 29760 + }, + { + "epoch": 61.88149688149688, + "grad_norm": 0.0005953563959337771, + "learning_rate": 0.04591806550008685, + "loss": 0.2666, + "num_input_tokens_seen": 22700312, + "step": 29765 + }, + { + "epoch": 61.891891891891895, + "grad_norm": 0.0002817870117723942, + "learning_rate": 0.045875656661156825, + "loss": 0.2598, + "num_input_tokens_seen": 22703928, + "step": 29770 + }, + { + "epoch": 61.9022869022869, + "grad_norm": 5.4918957175686955e-05, + "learning_rate": 0.04583326387950911, + "loss": 0.2521, + "num_input_tokens_seen": 22707672, + "step": 29775 + }, + { + "epoch": 61.91268191268191, + "grad_norm": 8.617278945166618e-05, + "learning_rate": 0.0457908871616811, + "loss": 0.2783, + "num_input_tokens_seen": 22711480, + "step": 29780 + }, + { + "epoch": 61.92307692307692, + "grad_norm": 0.0005102165159769356, + "learning_rate": 0.04574852651420786, + "loss": 0.245, + "num_input_tokens_seen": 22715192, + "step": 29785 + }, + { + "epoch": 61.933471933471935, + "grad_norm": 0.00015870384231675416, + "learning_rate": 0.045706181943621985, + "loss": 0.2697, + "num_input_tokens_seen": 22719096, + "step": 29790 + }, + { + "epoch": 61.943866943866944, + "grad_norm": 0.00010720246064011008, + "learning_rate": 0.04566385345645344, + "loss": 0.2531, + "num_input_tokens_seen": 22723032, + "step": 29795 + }, + { + "epoch": 61.95426195426195, + "grad_norm": 0.00012251581938471645, + "learning_rate": 0.04562154105922993, + "loss": 0.2615, + "num_input_tokens_seen": 22727064, + "step": 29800 + }, + { + "epoch": 61.95426195426195, + "eval_loss": 0.24720001220703125, + "eval_runtime": 13.4156, + "eval_samples_per_second": 63.807, + "eval_steps_per_second": 15.952, + "num_input_tokens_seen": 22727064, + "step": 29800 + }, + { + "epoch": 61.96465696465697, + "grad_norm": 0.00013783949543721974, + "learning_rate": 0.04557924475847642, + "loss": 0.2804, + "num_input_tokens_seen": 22730904, + "step": 29805 + }, + { + "epoch": 61.975051975051976, + "grad_norm": 0.00031125059467740357, + "learning_rate": 0.04553696456071567, + "loss": 0.2474, + "num_input_tokens_seen": 22734872, + "step": 29810 + }, + { + "epoch": 61.985446985446984, + "grad_norm": 0.0003392488870304078, + "learning_rate": 0.045494700472467724, + "loss": 0.285, + "num_input_tokens_seen": 22738584, + "step": 29815 + }, + { + "epoch": 61.99584199584199, + "grad_norm": 8.49536299938336e-05, + "learning_rate": 0.04545245250025024, + "loss": 0.2519, + "num_input_tokens_seen": 22742488, + "step": 29820 + }, + { + "epoch": 62.00623700623701, + "grad_norm": 0.00032907864078879356, + "learning_rate": 0.045410220650578384, + "loss": 0.2576, + "num_input_tokens_seen": 22746400, + "step": 29825 + }, + { + "epoch": 62.016632016632016, + "grad_norm": 0.0005208990187384188, + "learning_rate": 0.04536800492996492, + "loss": 0.2777, + "num_input_tokens_seen": 22750272, + "step": 29830 + }, + { + "epoch": 62.027027027027025, + "grad_norm": 0.0005330211133696139, + "learning_rate": 0.04532580534491994, + "loss": 0.2477, + "num_input_tokens_seen": 22754048, + "step": 29835 + }, + { + "epoch": 62.03742203742204, + "grad_norm": 0.0002972113434225321, + "learning_rate": 0.045283621901951183, + "loss": 0.2843, + "num_input_tokens_seen": 22758112, + "step": 29840 + }, + { + "epoch": 62.04781704781705, + "grad_norm": 0.00015745799464639276, + "learning_rate": 0.04524145460756393, + "loss": 0.2756, + "num_input_tokens_seen": 22761952, + "step": 29845 + }, + { + "epoch": 62.05821205821206, + "grad_norm": 0.00020698031585197896, + "learning_rate": 0.045199303468260794, + "loss": 0.2539, + "num_input_tokens_seen": 22765728, + "step": 29850 + }, + { + "epoch": 62.06860706860707, + "grad_norm": 0.00015088576765265316, + "learning_rate": 0.04515716849054214, + "loss": 0.2656, + "num_input_tokens_seen": 22769536, + "step": 29855 + }, + { + "epoch": 62.07900207900208, + "grad_norm": 0.0002759074850473553, + "learning_rate": 0.04511504968090558, + "loss": 0.2675, + "num_input_tokens_seen": 22773344, + "step": 29860 + }, + { + "epoch": 62.08939708939709, + "grad_norm": 0.0008666822104714811, + "learning_rate": 0.04507294704584644, + "loss": 0.2562, + "num_input_tokens_seen": 22777184, + "step": 29865 + }, + { + "epoch": 62.0997920997921, + "grad_norm": 0.0007040095515549183, + "learning_rate": 0.04503086059185749, + "loss": 0.2682, + "num_input_tokens_seen": 22781024, + "step": 29870 + }, + { + "epoch": 62.11018711018711, + "grad_norm": 0.00021752313477918506, + "learning_rate": 0.04498879032542893, + "loss": 0.2478, + "num_input_tokens_seen": 22784864, + "step": 29875 + }, + { + "epoch": 62.12058212058212, + "grad_norm": 0.0005560569115914404, + "learning_rate": 0.0449467362530486, + "loss": 0.2713, + "num_input_tokens_seen": 22788704, + "step": 29880 + }, + { + "epoch": 62.13097713097713, + "grad_norm": 0.0003077409928664565, + "learning_rate": 0.04490469838120171, + "loss": 0.2909, + "num_input_tokens_seen": 22792480, + "step": 29885 + }, + { + "epoch": 62.141372141372145, + "grad_norm": 0.000513959676027298, + "learning_rate": 0.04486267671637101, + "loss": 0.2784, + "num_input_tokens_seen": 22796256, + "step": 29890 + }, + { + "epoch": 62.15176715176715, + "grad_norm": 0.0005690513062290847, + "learning_rate": 0.04482067126503683, + "loss": 0.2727, + "num_input_tokens_seen": 22800128, + "step": 29895 + }, + { + "epoch": 62.16216216216216, + "grad_norm": 0.00024337589275091887, + "learning_rate": 0.04477868203367687, + "loss": 0.2658, + "num_input_tokens_seen": 22803776, + "step": 29900 + }, + { + "epoch": 62.17255717255717, + "grad_norm": 0.00020775568555109203, + "learning_rate": 0.044736709028766426, + "loss": 0.2672, + "num_input_tokens_seen": 22807680, + "step": 29905 + }, + { + "epoch": 62.182952182952185, + "grad_norm": 0.0008607160416431725, + "learning_rate": 0.04469475225677832, + "loss": 0.2696, + "num_input_tokens_seen": 22811456, + "step": 29910 + }, + { + "epoch": 62.19334719334719, + "grad_norm": 0.00029529407038353384, + "learning_rate": 0.04465281172418273, + "loss": 0.2648, + "num_input_tokens_seen": 22815296, + "step": 29915 + }, + { + "epoch": 62.2037422037422, + "grad_norm": 0.0005173709359951317, + "learning_rate": 0.044610887437447476, + "loss": 0.2441, + "num_input_tokens_seen": 22819168, + "step": 29920 + }, + { + "epoch": 62.21413721413722, + "grad_norm": 0.00030752719612792134, + "learning_rate": 0.044568979403037744, + "loss": 0.2571, + "num_input_tokens_seen": 22823136, + "step": 29925 + }, + { + "epoch": 62.224532224532226, + "grad_norm": 0.0001653421059018001, + "learning_rate": 0.04452708762741631, + "loss": 0.2562, + "num_input_tokens_seen": 22826784, + "step": 29930 + }, + { + "epoch": 62.234927234927234, + "grad_norm": 0.0002511379134375602, + "learning_rate": 0.044485212117043475, + "loss": 0.2507, + "num_input_tokens_seen": 22830592, + "step": 29935 + }, + { + "epoch": 62.24532224532224, + "grad_norm": 0.00027206551749259233, + "learning_rate": 0.04444335287837687, + "loss": 0.2745, + "num_input_tokens_seen": 22834464, + "step": 29940 + }, + { + "epoch": 62.25571725571726, + "grad_norm": 0.0003265838313382119, + "learning_rate": 0.04440150991787179, + "loss": 0.2732, + "num_input_tokens_seen": 22838048, + "step": 29945 + }, + { + "epoch": 62.266112266112266, + "grad_norm": 0.0002748042461462319, + "learning_rate": 0.04435968324198088, + "loss": 0.2779, + "num_input_tokens_seen": 22842016, + "step": 29950 + }, + { + "epoch": 62.276507276507274, + "grad_norm": 5.0398575694998726e-05, + "learning_rate": 0.04431787285715442, + "loss": 0.262, + "num_input_tokens_seen": 22845920, + "step": 29955 + }, + { + "epoch": 62.28690228690229, + "grad_norm": 0.0003121078771073371, + "learning_rate": 0.04427607876984004, + "loss": 0.2567, + "num_input_tokens_seen": 22849728, + "step": 29960 + }, + { + "epoch": 62.2972972972973, + "grad_norm": 0.00014542232383973897, + "learning_rate": 0.044234300986482886, + "loss": 0.2603, + "num_input_tokens_seen": 22853568, + "step": 29965 + }, + { + "epoch": 62.30769230769231, + "grad_norm": 0.00026760200853459537, + "learning_rate": 0.04419253951352566, + "loss": 0.2539, + "num_input_tokens_seen": 22857280, + "step": 29970 + }, + { + "epoch": 62.318087318087315, + "grad_norm": 0.0003560234617907554, + "learning_rate": 0.044150794357408533, + "loss": 0.2356, + "num_input_tokens_seen": 22861088, + "step": 29975 + }, + { + "epoch": 62.32848232848233, + "grad_norm": 0.00013567270070780069, + "learning_rate": 0.044109065524569065, + "loss": 0.2568, + "num_input_tokens_seen": 22864928, + "step": 29980 + }, + { + "epoch": 62.33887733887734, + "grad_norm": 5.2321087423479185e-05, + "learning_rate": 0.0440673530214424, + "loss": 0.2464, + "num_input_tokens_seen": 22868704, + "step": 29985 + }, + { + "epoch": 62.34927234927235, + "grad_norm": 0.00024049948842730373, + "learning_rate": 0.04402565685446117, + "loss": 0.2572, + "num_input_tokens_seen": 22872608, + "step": 29990 + }, + { + "epoch": 62.35966735966736, + "grad_norm": 0.00029498376534320414, + "learning_rate": 0.04398397703005536, + "loss": 0.2626, + "num_input_tokens_seen": 22876448, + "step": 29995 + }, + { + "epoch": 62.37006237006237, + "grad_norm": 0.0004322648164816201, + "learning_rate": 0.043942313554652626, + "loss": 0.2514, + "num_input_tokens_seen": 22880256, + "step": 30000 + }, + { + "epoch": 62.37006237006237, + "eval_loss": 0.24741502106189728, + "eval_runtime": 13.4102, + "eval_samples_per_second": 63.832, + "eval_steps_per_second": 15.958, + "num_input_tokens_seen": 22880256, + "step": 30000 + }, + { + "epoch": 62.38045738045738, + "grad_norm": 0.00032661177101545036, + "learning_rate": 0.0439006664346779, + "loss": 0.2654, + "num_input_tokens_seen": 22883936, + "step": 30005 + }, + { + "epoch": 62.39085239085239, + "grad_norm": 0.00023311520635616034, + "learning_rate": 0.043859035676553755, + "loss": 0.2537, + "num_input_tokens_seen": 22887808, + "step": 30010 + }, + { + "epoch": 62.4012474012474, + "grad_norm": 0.0003312415792606771, + "learning_rate": 0.043817421286700194, + "loss": 0.2627, + "num_input_tokens_seen": 22891680, + "step": 30015 + }, + { + "epoch": 62.41164241164241, + "grad_norm": 9.473994578002021e-05, + "learning_rate": 0.043775823271534585, + "loss": 0.2591, + "num_input_tokens_seen": 22895552, + "step": 30020 + }, + { + "epoch": 62.42203742203742, + "grad_norm": 0.00018004674348048866, + "learning_rate": 0.04373424163747197, + "loss": 0.2487, + "num_input_tokens_seen": 22899360, + "step": 30025 + }, + { + "epoch": 62.432432432432435, + "grad_norm": 0.0005645894561894238, + "learning_rate": 0.04369267639092473, + "loss": 0.2394, + "num_input_tokens_seen": 22903136, + "step": 30030 + }, + { + "epoch": 62.44282744282744, + "grad_norm": 0.00014490137982647866, + "learning_rate": 0.04365112753830268, + "loss": 0.2595, + "num_input_tokens_seen": 22906752, + "step": 30035 + }, + { + "epoch": 62.45322245322245, + "grad_norm": 0.0002702472556848079, + "learning_rate": 0.04360959508601327, + "loss": 0.2581, + "num_input_tokens_seen": 22910688, + "step": 30040 + }, + { + "epoch": 62.46361746361746, + "grad_norm": 0.0002903661224991083, + "learning_rate": 0.04356807904046123, + "loss": 0.2449, + "num_input_tokens_seen": 22914368, + "step": 30045 + }, + { + "epoch": 62.474012474012476, + "grad_norm": 0.000628048786893487, + "learning_rate": 0.04352657940804892, + "loss": 0.2634, + "num_input_tokens_seen": 22918144, + "step": 30050 + }, + { + "epoch": 62.484407484407484, + "grad_norm": 0.0006930386298336089, + "learning_rate": 0.04348509619517613, + "loss": 0.2279, + "num_input_tokens_seen": 22921952, + "step": 30055 + }, + { + "epoch": 62.49480249480249, + "grad_norm": 3.640597788034938e-05, + "learning_rate": 0.04344362940824002, + "loss": 0.2583, + "num_input_tokens_seen": 22925824, + "step": 30060 + }, + { + "epoch": 62.50519750519751, + "grad_norm": 0.00035645178286358714, + "learning_rate": 0.04340217905363533, + "loss": 0.2671, + "num_input_tokens_seen": 22929504, + "step": 30065 + }, + { + "epoch": 62.515592515592516, + "grad_norm": 0.000650412286631763, + "learning_rate": 0.04336074513775425, + "loss": 0.2441, + "num_input_tokens_seen": 22933280, + "step": 30070 + }, + { + "epoch": 62.525987525987524, + "grad_norm": 0.000225075549678877, + "learning_rate": 0.04331932766698636, + "loss": 0.2792, + "num_input_tokens_seen": 22936928, + "step": 30075 + }, + { + "epoch": 62.53638253638254, + "grad_norm": 0.00014629241195507348, + "learning_rate": 0.0432779266477188, + "loss": 0.2684, + "num_input_tokens_seen": 22940864, + "step": 30080 + }, + { + "epoch": 62.54677754677755, + "grad_norm": 6.334840873023495e-05, + "learning_rate": 0.04323654208633607, + "loss": 0.2367, + "num_input_tokens_seen": 22944416, + "step": 30085 + }, + { + "epoch": 62.55717255717256, + "grad_norm": 0.00044576849904842675, + "learning_rate": 0.04319517398922024, + "loss": 0.2719, + "num_input_tokens_seen": 22948160, + "step": 30090 + }, + { + "epoch": 62.567567567567565, + "grad_norm": 0.0011152951046824455, + "learning_rate": 0.04315382236275079, + "loss": 0.2533, + "num_input_tokens_seen": 22951904, + "step": 30095 + }, + { + "epoch": 62.57796257796258, + "grad_norm": 0.0003502096515148878, + "learning_rate": 0.043112487213304664, + "loss": 0.2833, + "num_input_tokens_seen": 22955744, + "step": 30100 + }, + { + "epoch": 62.58835758835759, + "grad_norm": 4.0146405808627605e-05, + "learning_rate": 0.04307116854725618, + "loss": 0.2655, + "num_input_tokens_seen": 22959616, + "step": 30105 + }, + { + "epoch": 62.5987525987526, + "grad_norm": 0.0002620772283989936, + "learning_rate": 0.043029866370977325, + "loss": 0.2915, + "num_input_tokens_seen": 22963520, + "step": 30110 + }, + { + "epoch": 62.60914760914761, + "grad_norm": 0.000122947632917203, + "learning_rate": 0.04298858069083728, + "loss": 0.264, + "num_input_tokens_seen": 22967392, + "step": 30115 + }, + { + "epoch": 62.61954261954262, + "grad_norm": 0.00012421427527442575, + "learning_rate": 0.04294731151320295, + "loss": 0.279, + "num_input_tokens_seen": 22971168, + "step": 30120 + }, + { + "epoch": 62.62993762993763, + "grad_norm": 0.0001404361828463152, + "learning_rate": 0.04290605884443841, + "loss": 0.2765, + "num_input_tokens_seen": 22974944, + "step": 30125 + }, + { + "epoch": 62.64033264033264, + "grad_norm": 0.00047713748062960804, + "learning_rate": 0.04286482269090545, + "loss": 0.2605, + "num_input_tokens_seen": 22978624, + "step": 30130 + }, + { + "epoch": 62.65072765072765, + "grad_norm": 0.000299125473247841, + "learning_rate": 0.04282360305896323, + "loss": 0.2679, + "num_input_tokens_seen": 22982464, + "step": 30135 + }, + { + "epoch": 62.66112266112266, + "grad_norm": 0.00031978421611711383, + "learning_rate": 0.04278239995496822, + "loss": 0.2682, + "num_input_tokens_seen": 22986176, + "step": 30140 + }, + { + "epoch": 62.67151767151767, + "grad_norm": 0.0005218606092967093, + "learning_rate": 0.042741213385274514, + "loss": 0.25, + "num_input_tokens_seen": 22990016, + "step": 30145 + }, + { + "epoch": 62.681912681912685, + "grad_norm": 0.00010784743790281937, + "learning_rate": 0.04270004335623366, + "loss": 0.258, + "num_input_tokens_seen": 22993792, + "step": 30150 + }, + { + "epoch": 62.69230769230769, + "grad_norm": 0.00027760062948800623, + "learning_rate": 0.04265888987419448, + "loss": 0.2438, + "num_input_tokens_seen": 22997568, + "step": 30155 + }, + { + "epoch": 62.7027027027027, + "grad_norm": 0.0004221794370096177, + "learning_rate": 0.04261775294550346, + "loss": 0.2593, + "num_input_tokens_seen": 23001408, + "step": 30160 + }, + { + "epoch": 62.71309771309771, + "grad_norm": 4.818370143766515e-05, + "learning_rate": 0.042576632576504354, + "loss": 0.2727, + "num_input_tokens_seen": 23005312, + "step": 30165 + }, + { + "epoch": 62.723492723492726, + "grad_norm": 0.0001188266251119785, + "learning_rate": 0.0425355287735385, + "loss": 0.2882, + "num_input_tokens_seen": 23009312, + "step": 30170 + }, + { + "epoch": 62.733887733887734, + "grad_norm": 0.000365598127245903, + "learning_rate": 0.0424944415429446, + "loss": 0.2876, + "num_input_tokens_seen": 23013184, + "step": 30175 + }, + { + "epoch": 62.74428274428274, + "grad_norm": 0.0004075798497069627, + "learning_rate": 0.04245337089105877, + "loss": 0.2704, + "num_input_tokens_seen": 23017120, + "step": 30180 + }, + { + "epoch": 62.75467775467776, + "grad_norm": 0.0002494051877874881, + "learning_rate": 0.04241231682421467, + "loss": 0.2642, + "num_input_tokens_seen": 23021088, + "step": 30185 + }, + { + "epoch": 62.765072765072766, + "grad_norm": 0.00021104175539221615, + "learning_rate": 0.04237127934874337, + "loss": 0.2617, + "num_input_tokens_seen": 23025216, + "step": 30190 + }, + { + "epoch": 62.775467775467774, + "grad_norm": 0.0001161335312644951, + "learning_rate": 0.042330258470973305, + "loss": 0.2754, + "num_input_tokens_seen": 23028960, + "step": 30195 + }, + { + "epoch": 62.78586278586278, + "grad_norm": 0.0003201899817213416, + "learning_rate": 0.042289254197230515, + "loss": 0.2572, + "num_input_tokens_seen": 23032800, + "step": 30200 + }, + { + "epoch": 62.78586278586278, + "eval_loss": 0.2501713037490845, + "eval_runtime": 13.4296, + "eval_samples_per_second": 63.74, + "eval_steps_per_second": 15.935, + "num_input_tokens_seen": 23032800, + "step": 30200 + }, + { + "epoch": 62.7962577962578, + "grad_norm": 0.00010216915688943118, + "learning_rate": 0.04224826653383823, + "loss": 0.255, + "num_input_tokens_seen": 23036704, + "step": 30205 + }, + { + "epoch": 62.80665280665281, + "grad_norm": 8.968092879513279e-05, + "learning_rate": 0.04220729548711735, + "loss": 0.2734, + "num_input_tokens_seen": 23040704, + "step": 30210 + }, + { + "epoch": 62.817047817047815, + "grad_norm": 0.0002949389163404703, + "learning_rate": 0.04216634106338616, + "loss": 0.2539, + "num_input_tokens_seen": 23044544, + "step": 30215 + }, + { + "epoch": 62.82744282744283, + "grad_norm": 0.0002409187873126939, + "learning_rate": 0.04212540326896025, + "loss": 0.2474, + "num_input_tokens_seen": 23048448, + "step": 30220 + }, + { + "epoch": 62.83783783783784, + "grad_norm": 0.0002916075463872403, + "learning_rate": 0.0420844821101528, + "loss": 0.2975, + "num_input_tokens_seen": 23052320, + "step": 30225 + }, + { + "epoch": 62.84823284823285, + "grad_norm": 0.0003147415118291974, + "learning_rate": 0.04204357759327441, + "loss": 0.2718, + "num_input_tokens_seen": 23056000, + "step": 30230 + }, + { + "epoch": 62.858627858627855, + "grad_norm": 0.0001354873675154522, + "learning_rate": 0.042002689724632954, + "loss": 0.2607, + "num_input_tokens_seen": 23059776, + "step": 30235 + }, + { + "epoch": 62.86902286902287, + "grad_norm": 0.0003973265702370554, + "learning_rate": 0.04196181851053398, + "loss": 0.2799, + "num_input_tokens_seen": 23063520, + "step": 30240 + }, + { + "epoch": 62.87941787941788, + "grad_norm": 0.0003862975863739848, + "learning_rate": 0.041920963957280295, + "loss": 0.2807, + "num_input_tokens_seen": 23067200, + "step": 30245 + }, + { + "epoch": 62.88981288981289, + "grad_norm": 0.000357545621227473, + "learning_rate": 0.04188012607117212, + "loss": 0.2716, + "num_input_tokens_seen": 23071008, + "step": 30250 + }, + { + "epoch": 62.9002079002079, + "grad_norm": 0.0001907086552819237, + "learning_rate": 0.04183930485850725, + "loss": 0.2727, + "num_input_tokens_seen": 23074688, + "step": 30255 + }, + { + "epoch": 62.91060291060291, + "grad_norm": 0.00019051968411076814, + "learning_rate": 0.04179850032558078, + "loss": 0.2731, + "num_input_tokens_seen": 23078560, + "step": 30260 + }, + { + "epoch": 62.92099792099792, + "grad_norm": 0.000594455748796463, + "learning_rate": 0.041757712478685295, + "loss": 0.2522, + "num_input_tokens_seen": 23082304, + "step": 30265 + }, + { + "epoch": 62.931392931392935, + "grad_norm": 0.00017930276226252317, + "learning_rate": 0.04171694132411085, + "loss": 0.274, + "num_input_tokens_seen": 23086112, + "step": 30270 + }, + { + "epoch": 62.94178794178794, + "grad_norm": 6.014187238179147e-05, + "learning_rate": 0.04167618686814479, + "loss": 0.2598, + "num_input_tokens_seen": 23089920, + "step": 30275 + }, + { + "epoch": 62.95218295218295, + "grad_norm": 0.0003506779030431062, + "learning_rate": 0.041635449117072024, + "loss": 0.2483, + "num_input_tokens_seen": 23093600, + "step": 30280 + }, + { + "epoch": 62.96257796257796, + "grad_norm": 0.00033882210846059024, + "learning_rate": 0.04159472807717477, + "loss": 0.266, + "num_input_tokens_seen": 23097504, + "step": 30285 + }, + { + "epoch": 62.972972972972975, + "grad_norm": 0.00043650579755194485, + "learning_rate": 0.041554023754732744, + "loss": 0.3049, + "num_input_tokens_seen": 23101344, + "step": 30290 + }, + { + "epoch": 62.983367983367984, + "grad_norm": 0.00032060674857348204, + "learning_rate": 0.04151333615602311, + "loss": 0.2749, + "num_input_tokens_seen": 23105120, + "step": 30295 + }, + { + "epoch": 62.99376299376299, + "grad_norm": 8.948193135438487e-05, + "learning_rate": 0.04147266528732034, + "loss": 0.2532, + "num_input_tokens_seen": 23108896, + "step": 30300 + }, + { + "epoch": 63.00415800415801, + "grad_norm": 0.00022815649572294205, + "learning_rate": 0.0414320111548964, + "loss": 0.2616, + "num_input_tokens_seen": 23112808, + "step": 30305 + }, + { + "epoch": 63.014553014553016, + "grad_norm": 4.847226227866486e-05, + "learning_rate": 0.04139137376502076, + "loss": 0.2519, + "num_input_tokens_seen": 23116488, + "step": 30310 + }, + { + "epoch": 63.024948024948024, + "grad_norm": 0.000665325322188437, + "learning_rate": 0.04135075312396014, + "loss": 0.2778, + "num_input_tokens_seen": 23120360, + "step": 30315 + }, + { + "epoch": 63.03534303534303, + "grad_norm": 0.00036534195533022285, + "learning_rate": 0.04131014923797875, + "loss": 0.2797, + "num_input_tokens_seen": 23124040, + "step": 30320 + }, + { + "epoch": 63.04573804573805, + "grad_norm": 0.0006709165754728019, + "learning_rate": 0.04126956211333819, + "loss": 0.2962, + "num_input_tokens_seen": 23127784, + "step": 30325 + }, + { + "epoch": 63.056133056133056, + "grad_norm": 0.0003930821258109063, + "learning_rate": 0.041228991756297545, + "loss": 0.2758, + "num_input_tokens_seen": 23131560, + "step": 30330 + }, + { + "epoch": 63.066528066528065, + "grad_norm": 0.00016460947517771274, + "learning_rate": 0.04118843817311332, + "loss": 0.263, + "num_input_tokens_seen": 23135336, + "step": 30335 + }, + { + "epoch": 63.07692307692308, + "grad_norm": 8.551117934985086e-05, + "learning_rate": 0.0411479013700393, + "loss": 0.2714, + "num_input_tokens_seen": 23139112, + "step": 30340 + }, + { + "epoch": 63.08731808731809, + "grad_norm": 0.0005092486971989274, + "learning_rate": 0.0411073813533268, + "loss": 0.2767, + "num_input_tokens_seen": 23143016, + "step": 30345 + }, + { + "epoch": 63.0977130977131, + "grad_norm": 0.0003225219261366874, + "learning_rate": 0.04106687812922456, + "loss": 0.2649, + "num_input_tokens_seen": 23146760, + "step": 30350 + }, + { + "epoch": 63.108108108108105, + "grad_norm": 0.00022369003272615373, + "learning_rate": 0.041026391703978635, + "loss": 0.2417, + "num_input_tokens_seen": 23150632, + "step": 30355 + }, + { + "epoch": 63.11850311850312, + "grad_norm": 0.0002041378611465916, + "learning_rate": 0.04098592208383259, + "loss": 0.2932, + "num_input_tokens_seen": 23154376, + "step": 30360 + }, + { + "epoch": 63.12889812889813, + "grad_norm": 0.0004909907002002001, + "learning_rate": 0.040945469275027256, + "loss": 0.2585, + "num_input_tokens_seen": 23158088, + "step": 30365 + }, + { + "epoch": 63.13929313929314, + "grad_norm": 0.00029031949816271663, + "learning_rate": 0.04090503328380104, + "loss": 0.279, + "num_input_tokens_seen": 23161864, + "step": 30370 + }, + { + "epoch": 63.14968814968815, + "grad_norm": 0.00034845375921577215, + "learning_rate": 0.04086461411638971, + "loss": 0.2656, + "num_input_tokens_seen": 23165640, + "step": 30375 + }, + { + "epoch": 63.16008316008316, + "grad_norm": 0.00021484777971636504, + "learning_rate": 0.04082421177902631, + "loss": 0.2727, + "num_input_tokens_seen": 23169480, + "step": 30380 + }, + { + "epoch": 63.17047817047817, + "grad_norm": 0.0002114875242114067, + "learning_rate": 0.04078382627794149, + "loss": 0.2623, + "num_input_tokens_seen": 23173256, + "step": 30385 + }, + { + "epoch": 63.18087318087318, + "grad_norm": 0.0002916967496275902, + "learning_rate": 0.04074345761936316, + "loss": 0.2571, + "num_input_tokens_seen": 23177000, + "step": 30390 + }, + { + "epoch": 63.19126819126819, + "grad_norm": 0.000641339342109859, + "learning_rate": 0.04070310580951663, + "loss": 0.2441, + "num_input_tokens_seen": 23180936, + "step": 30395 + }, + { + "epoch": 63.2016632016632, + "grad_norm": 5.407605567597784e-05, + "learning_rate": 0.040662770854624726, + "loss": 0.2698, + "num_input_tokens_seen": 23184744, + "step": 30400 + }, + { + "epoch": 63.2016632016632, + "eval_loss": 0.24731118977069855, + "eval_runtime": 13.3928, + "eval_samples_per_second": 63.915, + "eval_steps_per_second": 15.979, + "num_input_tokens_seen": 23184744, + "step": 30400 + }, + { + "epoch": 63.21205821205821, + "grad_norm": 0.0004855773877352476, + "learning_rate": 0.040622452760907535, + "loss": 0.2608, + "num_input_tokens_seen": 23188680, + "step": 30405 + }, + { + "epoch": 63.222453222453225, + "grad_norm": 0.00034704344579949975, + "learning_rate": 0.04058215153458265, + "loss": 0.2588, + "num_input_tokens_seen": 23192424, + "step": 30410 + }, + { + "epoch": 63.232848232848234, + "grad_norm": 0.0005789669812656939, + "learning_rate": 0.04054186718186507, + "loss": 0.2786, + "num_input_tokens_seen": 23196264, + "step": 30415 + }, + { + "epoch": 63.24324324324324, + "grad_norm": 0.0005630963132716715, + "learning_rate": 0.04050159970896708, + "loss": 0.2713, + "num_input_tokens_seen": 23200264, + "step": 30420 + }, + { + "epoch": 63.25363825363825, + "grad_norm": 0.00013894474250264466, + "learning_rate": 0.04046134912209843, + "loss": 0.2716, + "num_input_tokens_seen": 23204040, + "step": 30425 + }, + { + "epoch": 63.264033264033266, + "grad_norm": 0.0004776244459208101, + "learning_rate": 0.040421115427466354, + "loss": 0.2415, + "num_input_tokens_seen": 23207816, + "step": 30430 + }, + { + "epoch": 63.274428274428274, + "grad_norm": 0.0002857446379493922, + "learning_rate": 0.04038089863127529, + "loss": 0.2719, + "num_input_tokens_seen": 23211784, + "step": 30435 + }, + { + "epoch": 63.28482328482328, + "grad_norm": 0.00017533684149384499, + "learning_rate": 0.04034069873972727, + "loss": 0.251, + "num_input_tokens_seen": 23215592, + "step": 30440 + }, + { + "epoch": 63.2952182952183, + "grad_norm": 0.0001463294174754992, + "learning_rate": 0.040300515759021514, + "loss": 0.2394, + "num_input_tokens_seen": 23219400, + "step": 30445 + }, + { + "epoch": 63.305613305613306, + "grad_norm": 0.0003088580851908773, + "learning_rate": 0.04026034969535478, + "loss": 0.2636, + "num_input_tokens_seen": 23223208, + "step": 30450 + }, + { + "epoch": 63.316008316008315, + "grad_norm": 0.0003944539639633149, + "learning_rate": 0.040220200554921266, + "loss": 0.2533, + "num_input_tokens_seen": 23227144, + "step": 30455 + }, + { + "epoch": 63.32640332640332, + "grad_norm": 0.00015389127656817436, + "learning_rate": 0.0401800683439124, + "loss": 0.2632, + "num_input_tokens_seen": 23230888, + "step": 30460 + }, + { + "epoch": 63.33679833679834, + "grad_norm": 4.685593376052566e-05, + "learning_rate": 0.04013995306851704, + "loss": 0.2674, + "num_input_tokens_seen": 23234600, + "step": 30465 + }, + { + "epoch": 63.34719334719335, + "grad_norm": 0.0005774143501184881, + "learning_rate": 0.040099854734921545, + "loss": 0.2628, + "num_input_tokens_seen": 23238376, + "step": 30470 + }, + { + "epoch": 63.357588357588355, + "grad_norm": 0.0001762321189744398, + "learning_rate": 0.0400597733493095, + "loss": 0.2847, + "num_input_tokens_seen": 23242120, + "step": 30475 + }, + { + "epoch": 63.36798336798337, + "grad_norm": 0.00026515021454542875, + "learning_rate": 0.04001970891786203, + "loss": 0.2557, + "num_input_tokens_seen": 23245832, + "step": 30480 + }, + { + "epoch": 63.37837837837838, + "grad_norm": 0.0002766774268820882, + "learning_rate": 0.03997966144675752, + "loss": 0.2776, + "num_input_tokens_seen": 23249576, + "step": 30485 + }, + { + "epoch": 63.38877338877339, + "grad_norm": 0.0006170090055093169, + "learning_rate": 0.039939630942171796, + "loss": 0.2499, + "num_input_tokens_seen": 23253256, + "step": 30490 + }, + { + "epoch": 63.3991683991684, + "grad_norm": 7.146267307689413e-05, + "learning_rate": 0.03989961741027815, + "loss": 0.2632, + "num_input_tokens_seen": 23257000, + "step": 30495 + }, + { + "epoch": 63.40956340956341, + "grad_norm": 0.0002587572962511331, + "learning_rate": 0.03985962085724704, + "loss": 0.2648, + "num_input_tokens_seen": 23260904, + "step": 30500 + }, + { + "epoch": 63.41995841995842, + "grad_norm": 0.0002582099987193942, + "learning_rate": 0.03981964128924656, + "loss": 0.2585, + "num_input_tokens_seen": 23264680, + "step": 30505 + }, + { + "epoch": 63.43035343035343, + "grad_norm": 0.00040622326196171343, + "learning_rate": 0.03977967871244197, + "loss": 0.2457, + "num_input_tokens_seen": 23268584, + "step": 30510 + }, + { + "epoch": 63.44074844074844, + "grad_norm": 0.0006614467711187899, + "learning_rate": 0.03973973313299602, + "loss": 0.2454, + "num_input_tokens_seen": 23272392, + "step": 30515 + }, + { + "epoch": 63.45114345114345, + "grad_norm": 0.0002748437982518226, + "learning_rate": 0.0396998045570689, + "loss": 0.2839, + "num_input_tokens_seen": 23276168, + "step": 30520 + }, + { + "epoch": 63.46153846153846, + "grad_norm": 0.000268623378360644, + "learning_rate": 0.03965989299081798, + "loss": 0.262, + "num_input_tokens_seen": 23279880, + "step": 30525 + }, + { + "epoch": 63.471933471933475, + "grad_norm": 0.00028671478503383696, + "learning_rate": 0.039619998440398235, + "loss": 0.2697, + "num_input_tokens_seen": 23283656, + "step": 30530 + }, + { + "epoch": 63.482328482328484, + "grad_norm": 0.0003019487194251269, + "learning_rate": 0.03958012091196184, + "loss": 0.2601, + "num_input_tokens_seen": 23287656, + "step": 30535 + }, + { + "epoch": 63.49272349272349, + "grad_norm": 0.00023189772036857903, + "learning_rate": 0.039540260411658396, + "loss": 0.2404, + "num_input_tokens_seen": 23291528, + "step": 30540 + }, + { + "epoch": 63.5031185031185, + "grad_norm": 8.467365842079744e-05, + "learning_rate": 0.03950041694563496, + "loss": 0.2602, + "num_input_tokens_seen": 23295272, + "step": 30545 + }, + { + "epoch": 63.513513513513516, + "grad_norm": 8.710068505024537e-05, + "learning_rate": 0.0394605905200358, + "loss": 0.2761, + "num_input_tokens_seen": 23299112, + "step": 30550 + }, + { + "epoch": 63.523908523908524, + "grad_norm": 5.7084031141130254e-05, + "learning_rate": 0.03942078114100272, + "loss": 0.2811, + "num_input_tokens_seen": 23302952, + "step": 30555 + }, + { + "epoch": 63.53430353430353, + "grad_norm": 0.0003650786529760808, + "learning_rate": 0.03938098881467485, + "loss": 0.2568, + "num_input_tokens_seen": 23306792, + "step": 30560 + }, + { + "epoch": 63.54469854469855, + "grad_norm": 0.00011326957610435784, + "learning_rate": 0.039341213547188586, + "loss": 0.2575, + "num_input_tokens_seen": 23310504, + "step": 30565 + }, + { + "epoch": 63.555093555093556, + "grad_norm": 0.00041375195723958313, + "learning_rate": 0.03930145534467782, + "loss": 0.2562, + "num_input_tokens_seen": 23314344, + "step": 30570 + }, + { + "epoch": 63.565488565488565, + "grad_norm": 0.00056034279987216, + "learning_rate": 0.0392617142132738, + "loss": 0.2801, + "num_input_tokens_seen": 23318152, + "step": 30575 + }, + { + "epoch": 63.57588357588357, + "grad_norm": 0.0005134684033691883, + "learning_rate": 0.03922199015910504, + "loss": 0.2602, + "num_input_tokens_seen": 23321992, + "step": 30580 + }, + { + "epoch": 63.58627858627859, + "grad_norm": 0.0005481477710418403, + "learning_rate": 0.039182283188297556, + "loss": 0.2453, + "num_input_tokens_seen": 23325672, + "step": 30585 + }, + { + "epoch": 63.5966735966736, + "grad_norm": 0.0004044313100166619, + "learning_rate": 0.039142593306974595, + "loss": 0.2633, + "num_input_tokens_seen": 23329416, + "step": 30590 + }, + { + "epoch": 63.607068607068605, + "grad_norm": 0.0006246328121051192, + "learning_rate": 0.039102920521256856, + "loss": 0.2418, + "num_input_tokens_seen": 23333128, + "step": 30595 + }, + { + "epoch": 63.61746361746362, + "grad_norm": 0.0004957814235240221, + "learning_rate": 0.03906326483726243, + "loss": 0.2654, + "num_input_tokens_seen": 23336904, + "step": 30600 + }, + { + "epoch": 63.61746361746362, + "eval_loss": 0.24733735620975494, + "eval_runtime": 13.3842, + "eval_samples_per_second": 63.956, + "eval_steps_per_second": 15.989, + "num_input_tokens_seen": 23336904, + "step": 30600 + }, + { + "epoch": 63.62785862785863, + "grad_norm": 0.00015600465121679008, + "learning_rate": 0.039023626261106704, + "loss": 0.2626, + "num_input_tokens_seen": 23340712, + "step": 30605 + }, + { + "epoch": 63.63825363825364, + "grad_norm": 0.0005576710682362318, + "learning_rate": 0.03898400479890237, + "loss": 0.2677, + "num_input_tokens_seen": 23344456, + "step": 30610 + }, + { + "epoch": 63.648648648648646, + "grad_norm": 0.00016586252604611218, + "learning_rate": 0.038944400456759655, + "loss": 0.2841, + "num_input_tokens_seen": 23348296, + "step": 30615 + }, + { + "epoch": 63.65904365904366, + "grad_norm": 0.00016067402611952275, + "learning_rate": 0.038904813240785964, + "loss": 0.2699, + "num_input_tokens_seen": 23352008, + "step": 30620 + }, + { + "epoch": 63.66943866943867, + "grad_norm": 0.0003051013336516917, + "learning_rate": 0.03886524315708621, + "loss": 0.282, + "num_input_tokens_seen": 23355912, + "step": 30625 + }, + { + "epoch": 63.67983367983368, + "grad_norm": 0.0003926317731384188, + "learning_rate": 0.03882569021176255, + "loss": 0.2686, + "num_input_tokens_seen": 23359720, + "step": 30630 + }, + { + "epoch": 63.69022869022869, + "grad_norm": 0.0003085850621573627, + "learning_rate": 0.038786154410914535, + "loss": 0.2688, + "num_input_tokens_seen": 23363432, + "step": 30635 + }, + { + "epoch": 63.7006237006237, + "grad_norm": 0.0001126511488109827, + "learning_rate": 0.03874663576063917, + "loss": 0.2733, + "num_input_tokens_seen": 23367240, + "step": 30640 + }, + { + "epoch": 63.71101871101871, + "grad_norm": 0.0003682049282360822, + "learning_rate": 0.038707134267030624, + "loss": 0.2633, + "num_input_tokens_seen": 23371016, + "step": 30645 + }, + { + "epoch": 63.72141372141372, + "grad_norm": 0.00016524641250725836, + "learning_rate": 0.038667649936180555, + "loss": 0.2532, + "num_input_tokens_seen": 23374856, + "step": 30650 + }, + { + "epoch": 63.731808731808734, + "grad_norm": 0.0005512066418305039, + "learning_rate": 0.038628182774178, + "loss": 0.2502, + "num_input_tokens_seen": 23378696, + "step": 30655 + }, + { + "epoch": 63.74220374220374, + "grad_norm": 8.064149733399972e-05, + "learning_rate": 0.038588732787109226, + "loss": 0.208, + "num_input_tokens_seen": 23382600, + "step": 30660 + }, + { + "epoch": 63.75259875259875, + "grad_norm": 0.000163375458214432, + "learning_rate": 0.03854929998105795, + "loss": 0.2827, + "num_input_tokens_seen": 23386376, + "step": 30665 + }, + { + "epoch": 63.762993762993766, + "grad_norm": 0.00027057400438934565, + "learning_rate": 0.03850988436210518, + "loss": 0.2534, + "num_input_tokens_seen": 23390184, + "step": 30670 + }, + { + "epoch": 63.773388773388774, + "grad_norm": 0.0002026645961450413, + "learning_rate": 0.03847048593632933, + "loss": 0.2715, + "num_input_tokens_seen": 23393896, + "step": 30675 + }, + { + "epoch": 63.78378378378378, + "grad_norm": 0.0003511596005409956, + "learning_rate": 0.038431104709806096, + "loss": 0.2964, + "num_input_tokens_seen": 23397768, + "step": 30680 + }, + { + "epoch": 63.79417879417879, + "grad_norm": 0.00010664512228686363, + "learning_rate": 0.0383917406886086, + "loss": 0.251, + "num_input_tokens_seen": 23401704, + "step": 30685 + }, + { + "epoch": 63.804573804573806, + "grad_norm": 0.0003200983046554029, + "learning_rate": 0.03835239387880722, + "loss": 0.2759, + "num_input_tokens_seen": 23405544, + "step": 30690 + }, + { + "epoch": 63.814968814968815, + "grad_norm": 0.00023513525957241654, + "learning_rate": 0.03831306428646979, + "loss": 0.2628, + "num_input_tokens_seen": 23409352, + "step": 30695 + }, + { + "epoch": 63.82536382536382, + "grad_norm": 0.00034564806264825165, + "learning_rate": 0.03827375191766135, + "loss": 0.2608, + "num_input_tokens_seen": 23413192, + "step": 30700 + }, + { + "epoch": 63.83575883575884, + "grad_norm": 0.0002446005237288773, + "learning_rate": 0.03823445677844446, + "loss": 0.2692, + "num_input_tokens_seen": 23417320, + "step": 30705 + }, + { + "epoch": 63.84615384615385, + "grad_norm": 0.0004821166512556374, + "learning_rate": 0.03819517887487881, + "loss": 0.2616, + "num_input_tokens_seen": 23421096, + "step": 30710 + }, + { + "epoch": 63.856548856548855, + "grad_norm": 0.0003033368266187608, + "learning_rate": 0.03815591821302161, + "loss": 0.2618, + "num_input_tokens_seen": 23424968, + "step": 30715 + }, + { + "epoch": 63.86694386694387, + "grad_norm": 0.00035398584441281855, + "learning_rate": 0.03811667479892739, + "loss": 0.2496, + "num_input_tokens_seen": 23428808, + "step": 30720 + }, + { + "epoch": 63.87733887733888, + "grad_norm": 0.0001340054877800867, + "learning_rate": 0.03807744863864788, + "loss": 0.2722, + "num_input_tokens_seen": 23432488, + "step": 30725 + }, + { + "epoch": 63.88773388773389, + "grad_norm": 0.00022208612062968314, + "learning_rate": 0.03803823973823229, + "loss": 0.2709, + "num_input_tokens_seen": 23436328, + "step": 30730 + }, + { + "epoch": 63.898128898128896, + "grad_norm": 0.0001387394586345181, + "learning_rate": 0.03799904810372719, + "loss": 0.2674, + "num_input_tokens_seen": 23440168, + "step": 30735 + }, + { + "epoch": 63.90852390852391, + "grad_norm": 0.000780729460529983, + "learning_rate": 0.03795987374117632, + "loss": 0.2339, + "num_input_tokens_seen": 23443848, + "step": 30740 + }, + { + "epoch": 63.91891891891892, + "grad_norm": 0.0001924822136061266, + "learning_rate": 0.03792071665662093, + "loss": 0.2713, + "num_input_tokens_seen": 23447752, + "step": 30745 + }, + { + "epoch": 63.92931392931393, + "grad_norm": 0.00019927808898501098, + "learning_rate": 0.03788157685609952, + "loss": 0.2752, + "num_input_tokens_seen": 23451496, + "step": 30750 + }, + { + "epoch": 63.93970893970894, + "grad_norm": 0.0011418135836720467, + "learning_rate": 0.037842454345647876, + "loss": 0.2657, + "num_input_tokens_seen": 23455240, + "step": 30755 + }, + { + "epoch": 63.95010395010395, + "grad_norm": 0.00015185806842055172, + "learning_rate": 0.03780334913129929, + "loss": 0.2805, + "num_input_tokens_seen": 23459016, + "step": 30760 + }, + { + "epoch": 63.96049896049896, + "grad_norm": 0.0014124070294201374, + "learning_rate": 0.037764261219084175, + "loss": 0.2666, + "num_input_tokens_seen": 23462760, + "step": 30765 + }, + { + "epoch": 63.97089397089397, + "grad_norm": 0.0005831573507748544, + "learning_rate": 0.037725190615030414, + "loss": 0.2548, + "num_input_tokens_seen": 23466504, + "step": 30770 + }, + { + "epoch": 63.981288981288984, + "grad_norm": 0.0003159177431371063, + "learning_rate": 0.037686137325163224, + "loss": 0.2724, + "num_input_tokens_seen": 23470312, + "step": 30775 + }, + { + "epoch": 63.99168399168399, + "grad_norm": 0.0002493836800567806, + "learning_rate": 0.037647101355505065, + "loss": 0.2551, + "num_input_tokens_seen": 23474152, + "step": 30780 + }, + { + "epoch": 64.002079002079, + "grad_norm": 0.0005750659620389342, + "learning_rate": 0.03760808271207581, + "loss": 0.2265, + "num_input_tokens_seen": 23478072, + "step": 30785 + }, + { + "epoch": 64.01247401247402, + "grad_norm": 0.00030464300652965903, + "learning_rate": 0.03756908140089258, + "loss": 0.258, + "num_input_tokens_seen": 23481912, + "step": 30790 + }, + { + "epoch": 64.02286902286902, + "grad_norm": 0.0003069822851102799, + "learning_rate": 0.03753009742796989, + "loss": 0.2443, + "num_input_tokens_seen": 23485624, + "step": 30795 + }, + { + "epoch": 64.03326403326403, + "grad_norm": 0.00019055369193665683, + "learning_rate": 0.037491130799319615, + "loss": 0.2354, + "num_input_tokens_seen": 23489432, + "step": 30800 + }, + { + "epoch": 64.03326403326403, + "eval_loss": 0.24824325740337372, + "eval_runtime": 13.4147, + "eval_samples_per_second": 63.81, + "eval_steps_per_second": 15.953, + "num_input_tokens_seen": 23489432, + "step": 30800 + }, + { + "epoch": 64.04365904365905, + "grad_norm": 0.00020071028848178685, + "learning_rate": 0.03745218152095079, + "loss": 0.2389, + "num_input_tokens_seen": 23493208, + "step": 30805 + }, + { + "epoch": 64.05405405405405, + "grad_norm": 0.0004476974718272686, + "learning_rate": 0.037413249598869935, + "loss": 0.242, + "num_input_tokens_seen": 23497112, + "step": 30810 + }, + { + "epoch": 64.06444906444906, + "grad_norm": 0.0002993455855175853, + "learning_rate": 0.037374335039080886, + "loss": 0.2917, + "num_input_tokens_seen": 23500824, + "step": 30815 + }, + { + "epoch": 64.07484407484408, + "grad_norm": 0.00014793386799283326, + "learning_rate": 0.037335437847584724, + "loss": 0.2443, + "num_input_tokens_seen": 23504536, + "step": 30820 + }, + { + "epoch": 64.08523908523908, + "grad_norm": 0.0008046767907217145, + "learning_rate": 0.03729655803037983, + "loss": 0.2566, + "num_input_tokens_seen": 23508280, + "step": 30825 + }, + { + "epoch": 64.0956340956341, + "grad_norm": 0.00010719981219153851, + "learning_rate": 0.03725769559346207, + "loss": 0.2564, + "num_input_tokens_seen": 23512120, + "step": 30830 + }, + { + "epoch": 64.10602910602911, + "grad_norm": 0.00033316988265141845, + "learning_rate": 0.03721885054282439, + "loss": 0.2662, + "num_input_tokens_seen": 23515960, + "step": 30835 + }, + { + "epoch": 64.11642411642411, + "grad_norm": 8.170922956196591e-05, + "learning_rate": 0.03718002288445731, + "loss": 0.2375, + "num_input_tokens_seen": 23519864, + "step": 30840 + }, + { + "epoch": 64.12681912681913, + "grad_norm": 0.0003092300903517753, + "learning_rate": 0.03714121262434844, + "loss": 0.2636, + "num_input_tokens_seen": 23523480, + "step": 30845 + }, + { + "epoch": 64.13721413721414, + "grad_norm": 0.0001744226028677076, + "learning_rate": 0.037102419768482844, + "loss": 0.2501, + "num_input_tokens_seen": 23527320, + "step": 30850 + }, + { + "epoch": 64.14760914760915, + "grad_norm": 0.0006735483766533434, + "learning_rate": 0.03706364432284293, + "loss": 0.2805, + "num_input_tokens_seen": 23531096, + "step": 30855 + }, + { + "epoch": 64.15800415800416, + "grad_norm": 0.0006111849215812981, + "learning_rate": 0.03702488629340828, + "loss": 0.2558, + "num_input_tokens_seen": 23534840, + "step": 30860 + }, + { + "epoch": 64.16839916839916, + "grad_norm": 4.212605927023105e-05, + "learning_rate": 0.036986145686155915, + "loss": 0.2687, + "num_input_tokens_seen": 23538840, + "step": 30865 + }, + { + "epoch": 64.17879417879418, + "grad_norm": 6.692374881822616e-05, + "learning_rate": 0.036947422507060075, + "loss": 0.2598, + "num_input_tokens_seen": 23542776, + "step": 30870 + }, + { + "epoch": 64.1891891891892, + "grad_norm": 6.341277912724763e-05, + "learning_rate": 0.0369087167620924, + "loss": 0.2662, + "num_input_tokens_seen": 23546552, + "step": 30875 + }, + { + "epoch": 64.1995841995842, + "grad_norm": 0.0004200090479571372, + "learning_rate": 0.03687002845722183, + "loss": 0.2755, + "num_input_tokens_seen": 23550232, + "step": 30880 + }, + { + "epoch": 64.20997920997921, + "grad_norm": 0.00017318730533588678, + "learning_rate": 0.03683135759841451, + "loss": 0.2828, + "num_input_tokens_seen": 23554008, + "step": 30885 + }, + { + "epoch": 64.22037422037423, + "grad_norm": 0.00021113459661137313, + "learning_rate": 0.03679270419163406, + "loss": 0.2544, + "num_input_tokens_seen": 23557848, + "step": 30890 + }, + { + "epoch": 64.23076923076923, + "grad_norm": 0.0005112666985951364, + "learning_rate": 0.03675406824284127, + "loss": 0.2622, + "num_input_tokens_seen": 23561592, + "step": 30895 + }, + { + "epoch": 64.24116424116424, + "grad_norm": 0.0010388800874352455, + "learning_rate": 0.03671544975799425, + "loss": 0.2602, + "num_input_tokens_seen": 23565368, + "step": 30900 + }, + { + "epoch": 64.25155925155926, + "grad_norm": 0.00020329933613538742, + "learning_rate": 0.03667684874304854, + "loss": 0.2532, + "num_input_tokens_seen": 23569144, + "step": 30905 + }, + { + "epoch": 64.26195426195426, + "grad_norm": 0.00011772145080612972, + "learning_rate": 0.03663826520395683, + "loss": 0.2624, + "num_input_tokens_seen": 23572920, + "step": 30910 + }, + { + "epoch": 64.27234927234927, + "grad_norm": 6.005024624755606e-05, + "learning_rate": 0.03659969914666922, + "loss": 0.2896, + "num_input_tokens_seen": 23576536, + "step": 30915 + }, + { + "epoch": 64.28274428274429, + "grad_norm": 0.00010861018381547183, + "learning_rate": 0.036561150577133106, + "loss": 0.2777, + "num_input_tokens_seen": 23580120, + "step": 30920 + }, + { + "epoch": 64.29313929313929, + "grad_norm": 0.000434287911048159, + "learning_rate": 0.036522619501293103, + "loss": 0.2672, + "num_input_tokens_seen": 23584056, + "step": 30925 + }, + { + "epoch": 64.3035343035343, + "grad_norm": 0.0006371169583871961, + "learning_rate": 0.03648410592509122, + "loss": 0.2638, + "num_input_tokens_seen": 23587960, + "step": 30930 + }, + { + "epoch": 64.31392931392931, + "grad_norm": 0.0013753941748291254, + "learning_rate": 0.03644560985446676, + "loss": 0.2777, + "num_input_tokens_seen": 23591832, + "step": 30935 + }, + { + "epoch": 64.32432432432432, + "grad_norm": 0.0001845574879553169, + "learning_rate": 0.036407131295356256, + "loss": 0.2692, + "num_input_tokens_seen": 23595544, + "step": 30940 + }, + { + "epoch": 64.33471933471934, + "grad_norm": 0.00018170123803429306, + "learning_rate": 0.03636867025369362, + "loss": 0.2928, + "num_input_tokens_seen": 23599320, + "step": 30945 + }, + { + "epoch": 64.34511434511434, + "grad_norm": 0.000825926021207124, + "learning_rate": 0.03633022673540999, + "loss": 0.2886, + "num_input_tokens_seen": 23603192, + "step": 30950 + }, + { + "epoch": 64.35550935550935, + "grad_norm": 0.000158306400408037, + "learning_rate": 0.03629180074643385, + "loss": 0.2624, + "num_input_tokens_seen": 23607032, + "step": 30955 + }, + { + "epoch": 64.36590436590437, + "grad_norm": 0.0002621819730848074, + "learning_rate": 0.03625339229269102, + "loss": 0.2713, + "num_input_tokens_seen": 23611000, + "step": 30960 + }, + { + "epoch": 64.37629937629937, + "grad_norm": 0.0008049997850321233, + "learning_rate": 0.036215001380104535, + "loss": 0.2635, + "num_input_tokens_seen": 23614808, + "step": 30965 + }, + { + "epoch": 64.38669438669439, + "grad_norm": 0.0006757226074114442, + "learning_rate": 0.03617662801459471, + "loss": 0.2601, + "num_input_tokens_seen": 23618616, + "step": 30970 + }, + { + "epoch": 64.3970893970894, + "grad_norm": 0.0002419425145490095, + "learning_rate": 0.036138272202079276, + "loss": 0.2487, + "num_input_tokens_seen": 23622616, + "step": 30975 + }, + { + "epoch": 64.4074844074844, + "grad_norm": 0.00035483541432768106, + "learning_rate": 0.036099933948473106, + "loss": 0.2552, + "num_input_tokens_seen": 23626456, + "step": 30980 + }, + { + "epoch": 64.41787941787942, + "grad_norm": 5.000762394047342e-05, + "learning_rate": 0.03606161325968851, + "loss": 0.248, + "num_input_tokens_seen": 23630232, + "step": 30985 + }, + { + "epoch": 64.42827442827443, + "grad_norm": 0.0008898017113097012, + "learning_rate": 0.03602331014163496, + "loss": 0.2589, + "num_input_tokens_seen": 23633944, + "step": 30990 + }, + { + "epoch": 64.43866943866944, + "grad_norm": 0.00027639957261271775, + "learning_rate": 0.035985024600219295, + "loss": 0.2651, + "num_input_tokens_seen": 23637784, + "step": 30995 + }, + { + "epoch": 64.44906444906445, + "grad_norm": 6.263157411012799e-05, + "learning_rate": 0.03594675664134569, + "loss": 0.2766, + "num_input_tokens_seen": 23641496, + "step": 31000 + }, + { + "epoch": 64.44906444906445, + "eval_loss": 0.24707148969173431, + "eval_runtime": 13.3957, + "eval_samples_per_second": 63.901, + "eval_steps_per_second": 15.975, + "num_input_tokens_seen": 23641496, + "step": 31000 + }, + { + "epoch": 64.45945945945945, + "grad_norm": 0.0002089206245727837, + "learning_rate": 0.03590850627091545, + "loss": 0.2955, + "num_input_tokens_seen": 23645464, + "step": 31005 + }, + { + "epoch": 64.46985446985447, + "grad_norm": 0.00026809112750925124, + "learning_rate": 0.03587027349482731, + "loss": 0.2624, + "num_input_tokens_seen": 23649304, + "step": 31010 + }, + { + "epoch": 64.48024948024948, + "grad_norm": 0.0002953018993139267, + "learning_rate": 0.035832058318977275, + "loss": 0.2628, + "num_input_tokens_seen": 23653048, + "step": 31015 + }, + { + "epoch": 64.49064449064448, + "grad_norm": 0.0022743260487914085, + "learning_rate": 0.03579386074925853, + "loss": 0.2731, + "num_input_tokens_seen": 23656760, + "step": 31020 + }, + { + "epoch": 64.5010395010395, + "grad_norm": 0.0003832119982689619, + "learning_rate": 0.035755680791561696, + "loss": 0.2672, + "num_input_tokens_seen": 23660760, + "step": 31025 + }, + { + "epoch": 64.51143451143452, + "grad_norm": 0.0004797054280061275, + "learning_rate": 0.03571751845177454, + "loss": 0.2683, + "num_input_tokens_seen": 23664408, + "step": 31030 + }, + { + "epoch": 64.52182952182952, + "grad_norm": 0.00039895178633742034, + "learning_rate": 0.03567937373578225, + "loss": 0.2797, + "num_input_tokens_seen": 23668216, + "step": 31035 + }, + { + "epoch": 64.53222453222453, + "grad_norm": 0.0002463925047777593, + "learning_rate": 0.03564124664946711, + "loss": 0.2558, + "num_input_tokens_seen": 23672152, + "step": 31040 + }, + { + "epoch": 64.54261954261955, + "grad_norm": 0.0003663892566692084, + "learning_rate": 0.035603137198708924, + "loss": 0.2539, + "num_input_tokens_seen": 23675928, + "step": 31045 + }, + { + "epoch": 64.55301455301455, + "grad_norm": 0.0004516581248026341, + "learning_rate": 0.035565045389384514, + "loss": 0.2702, + "num_input_tokens_seen": 23679896, + "step": 31050 + }, + { + "epoch": 64.56340956340956, + "grad_norm": 0.0002812107268255204, + "learning_rate": 0.03552697122736823, + "loss": 0.252, + "num_input_tokens_seen": 23683768, + "step": 31055 + }, + { + "epoch": 64.57380457380458, + "grad_norm": 0.00047311969683505595, + "learning_rate": 0.03548891471853153, + "loss": 0.2852, + "num_input_tokens_seen": 23687480, + "step": 31060 + }, + { + "epoch": 64.58419958419958, + "grad_norm": 0.00019577915372792631, + "learning_rate": 0.03545087586874322, + "loss": 0.2663, + "num_input_tokens_seen": 23691416, + "step": 31065 + }, + { + "epoch": 64.5945945945946, + "grad_norm": 0.00027959910221397877, + "learning_rate": 0.03541285468386935, + "loss": 0.2552, + "num_input_tokens_seen": 23695192, + "step": 31070 + }, + { + "epoch": 64.60498960498961, + "grad_norm": 0.00040719486423768103, + "learning_rate": 0.03537485116977327, + "loss": 0.2453, + "num_input_tokens_seen": 23699128, + "step": 31075 + }, + { + "epoch": 64.61538461538461, + "grad_norm": 0.00043097263551317155, + "learning_rate": 0.03533686533231565, + "loss": 0.2841, + "num_input_tokens_seen": 23703000, + "step": 31080 + }, + { + "epoch": 64.62577962577963, + "grad_norm": 0.00019425485515967011, + "learning_rate": 0.0352988971773543, + "loss": 0.2904, + "num_input_tokens_seen": 23706808, + "step": 31085 + }, + { + "epoch": 64.63617463617463, + "grad_norm": 0.0002332548756385222, + "learning_rate": 0.03526094671074443, + "loss": 0.2699, + "num_input_tokens_seen": 23710552, + "step": 31090 + }, + { + "epoch": 64.64656964656965, + "grad_norm": 0.00013271684292703867, + "learning_rate": 0.03522301393833852, + "loss": 0.2704, + "num_input_tokens_seen": 23714360, + "step": 31095 + }, + { + "epoch": 64.65696465696466, + "grad_norm": 0.0008996139513328671, + "learning_rate": 0.035185098865986204, + "loss": 0.2673, + "num_input_tokens_seen": 23718328, + "step": 31100 + }, + { + "epoch": 64.66735966735966, + "grad_norm": 0.00011638574505923316, + "learning_rate": 0.03514720149953453, + "loss": 0.2728, + "num_input_tokens_seen": 23722104, + "step": 31105 + }, + { + "epoch": 64.67775467775468, + "grad_norm": 0.00015800939581822604, + "learning_rate": 0.03510932184482773, + "loss": 0.2621, + "num_input_tokens_seen": 23726040, + "step": 31110 + }, + { + "epoch": 64.6881496881497, + "grad_norm": 0.0001915096363518387, + "learning_rate": 0.03507145990770724, + "loss": 0.2627, + "num_input_tokens_seen": 23729880, + "step": 31115 + }, + { + "epoch": 64.6985446985447, + "grad_norm": 0.0006221798830665648, + "learning_rate": 0.035033615694011984, + "loss": 0.2645, + "num_input_tokens_seen": 23733624, + "step": 31120 + }, + { + "epoch": 64.70893970893971, + "grad_norm": 0.00021186945377849042, + "learning_rate": 0.03499578920957788, + "loss": 0.2537, + "num_input_tokens_seen": 23737464, + "step": 31125 + }, + { + "epoch": 64.71933471933473, + "grad_norm": 0.0002337249752599746, + "learning_rate": 0.034957980460238375, + "loss": 0.2596, + "num_input_tokens_seen": 23741400, + "step": 31130 + }, + { + "epoch": 64.72972972972973, + "grad_norm": 0.00018540085875429213, + "learning_rate": 0.03492018945182393, + "loss": 0.2783, + "num_input_tokens_seen": 23745208, + "step": 31135 + }, + { + "epoch": 64.74012474012474, + "grad_norm": 0.0003771416377276182, + "learning_rate": 0.03488241619016247, + "loss": 0.2755, + "num_input_tokens_seen": 23748952, + "step": 31140 + }, + { + "epoch": 64.75051975051976, + "grad_norm": 0.0003336940426379442, + "learning_rate": 0.03484466068107913, + "loss": 0.2746, + "num_input_tokens_seen": 23752632, + "step": 31145 + }, + { + "epoch": 64.76091476091476, + "grad_norm": 0.0002736676251515746, + "learning_rate": 0.034806922930396195, + "loss": 0.2422, + "num_input_tokens_seen": 23756376, + "step": 31150 + }, + { + "epoch": 64.77130977130977, + "grad_norm": 0.00015524720947723836, + "learning_rate": 0.03476920294393337, + "loss": 0.2533, + "num_input_tokens_seen": 23760152, + "step": 31155 + }, + { + "epoch": 64.78170478170478, + "grad_norm": 0.0006728442385792732, + "learning_rate": 0.03473150072750755, + "loss": 0.2553, + "num_input_tokens_seen": 23764120, + "step": 31160 + }, + { + "epoch": 64.79209979209979, + "grad_norm": 0.00021102912432979792, + "learning_rate": 0.03469381628693284, + "loss": 0.2672, + "num_input_tokens_seen": 23767928, + "step": 31165 + }, + { + "epoch": 64.8024948024948, + "grad_norm": 8.775474998401478e-05, + "learning_rate": 0.03465614962802072, + "loss": 0.2505, + "num_input_tokens_seen": 23771576, + "step": 31170 + }, + { + "epoch": 64.81288981288981, + "grad_norm": 0.00034374051028862596, + "learning_rate": 0.0346185007565798, + "loss": 0.2571, + "num_input_tokens_seen": 23775256, + "step": 31175 + }, + { + "epoch": 64.82328482328482, + "grad_norm": 0.00028751016361638904, + "learning_rate": 0.03458086967841609, + "loss": 0.2751, + "num_input_tokens_seen": 23779192, + "step": 31180 + }, + { + "epoch": 64.83367983367984, + "grad_norm": 0.00020720680186059326, + "learning_rate": 0.03454325639933266, + "loss": 0.268, + "num_input_tokens_seen": 23783064, + "step": 31185 + }, + { + "epoch": 64.84407484407484, + "grad_norm": 6.728695734636858e-05, + "learning_rate": 0.03450566092513007, + "loss": 0.2617, + "num_input_tokens_seen": 23786904, + "step": 31190 + }, + { + "epoch": 64.85446985446985, + "grad_norm": 0.0005479721585288644, + "learning_rate": 0.034468083261605914, + "loss": 0.245, + "num_input_tokens_seen": 23790808, + "step": 31195 + }, + { + "epoch": 64.86486486486487, + "grad_norm": 0.0003646282711997628, + "learning_rate": 0.03443052341455522, + "loss": 0.2808, + "num_input_tokens_seen": 23794744, + "step": 31200 + }, + { + "epoch": 64.86486486486487, + "eval_loss": 0.2504383325576782, + "eval_runtime": 13.3856, + "eval_samples_per_second": 63.949, + "eval_steps_per_second": 15.987, + "num_input_tokens_seen": 23794744, + "step": 31200 + }, + { + "epoch": 64.87525987525987, + "grad_norm": 6.249493890209123e-05, + "learning_rate": 0.0343929813897701, + "loss": 0.2945, + "num_input_tokens_seen": 23798648, + "step": 31205 + }, + { + "epoch": 64.88565488565489, + "grad_norm": 0.00011055656068492681, + "learning_rate": 0.034355457193040125, + "loss": 0.253, + "num_input_tokens_seen": 23802456, + "step": 31210 + }, + { + "epoch": 64.8960498960499, + "grad_norm": 0.0007412268896587193, + "learning_rate": 0.03431795083015186, + "loss": 0.2735, + "num_input_tokens_seen": 23806360, + "step": 31215 + }, + { + "epoch": 64.9064449064449, + "grad_norm": 0.0010586136486381292, + "learning_rate": 0.03428046230688936, + "loss": 0.2607, + "num_input_tokens_seen": 23810168, + "step": 31220 + }, + { + "epoch": 64.91683991683992, + "grad_norm": 0.00028996329638175666, + "learning_rate": 0.034242991629033805, + "loss": 0.2583, + "num_input_tokens_seen": 23813880, + "step": 31225 + }, + { + "epoch": 64.92723492723492, + "grad_norm": 0.00010236066009383649, + "learning_rate": 0.03420553880236362, + "loss": 0.2628, + "num_input_tokens_seen": 23817624, + "step": 31230 + }, + { + "epoch": 64.93762993762994, + "grad_norm": 0.00023848592536523938, + "learning_rate": 0.03416810383265449, + "loss": 0.2359, + "num_input_tokens_seen": 23821528, + "step": 31235 + }, + { + "epoch": 64.94802494802495, + "grad_norm": 0.0001134915801230818, + "learning_rate": 0.03413068672567944, + "loss": 0.2666, + "num_input_tokens_seen": 23825304, + "step": 31240 + }, + { + "epoch": 64.95841995841995, + "grad_norm": 5.729231270379387e-05, + "learning_rate": 0.034093287487208565, + "loss": 0.2592, + "num_input_tokens_seen": 23829016, + "step": 31245 + }, + { + "epoch": 64.96881496881497, + "grad_norm": 0.00024576118448749185, + "learning_rate": 0.03405590612300937, + "loss": 0.2611, + "num_input_tokens_seen": 23832792, + "step": 31250 + }, + { + "epoch": 64.97920997920998, + "grad_norm": 0.00017415838374290615, + "learning_rate": 0.03401854263884646, + "loss": 0.2606, + "num_input_tokens_seen": 23836728, + "step": 31255 + }, + { + "epoch": 64.98960498960498, + "grad_norm": 0.000707188795786351, + "learning_rate": 0.033981197040481824, + "loss": 0.2359, + "num_input_tokens_seen": 23840664, + "step": 31260 + }, + { + "epoch": 65.0, + "grad_norm": 0.00012469751527532935, + "learning_rate": 0.03394386933367459, + "loss": 0.2477, + "num_input_tokens_seen": 23844456, + "step": 31265 + }, + { + "epoch": 65.01039501039502, + "grad_norm": 0.0001524774415884167, + "learning_rate": 0.033906559524181104, + "loss": 0.244, + "num_input_tokens_seen": 23848424, + "step": 31270 + }, + { + "epoch": 65.02079002079002, + "grad_norm": 0.00013190801837481558, + "learning_rate": 0.033869267617755085, + "loss": 0.2729, + "num_input_tokens_seen": 23852328, + "step": 31275 + }, + { + "epoch": 65.03118503118503, + "grad_norm": 0.00011232318502152339, + "learning_rate": 0.0338319936201474, + "loss": 0.3129, + "num_input_tokens_seen": 23856136, + "step": 31280 + }, + { + "epoch": 65.04158004158005, + "grad_norm": 0.00018011359497904778, + "learning_rate": 0.033794737537106136, + "loss": 0.2651, + "num_input_tokens_seen": 23860008, + "step": 31285 + }, + { + "epoch": 65.05197505197505, + "grad_norm": 3.540679972502403e-05, + "learning_rate": 0.03375749937437671, + "loss": 0.2725, + "num_input_tokens_seen": 23863816, + "step": 31290 + }, + { + "epoch": 65.06237006237006, + "grad_norm": 6.644897803198546e-05, + "learning_rate": 0.033720279137701634, + "loss": 0.2741, + "num_input_tokens_seen": 23867592, + "step": 31295 + }, + { + "epoch": 65.07276507276508, + "grad_norm": 0.0003192275180481374, + "learning_rate": 0.03368307683282078, + "loss": 0.2493, + "num_input_tokens_seen": 23871336, + "step": 31300 + }, + { + "epoch": 65.08316008316008, + "grad_norm": 0.00016762847371865064, + "learning_rate": 0.033645892465471235, + "loss": 0.2565, + "num_input_tokens_seen": 23875144, + "step": 31305 + }, + { + "epoch": 65.0935550935551, + "grad_norm": 0.00027856751694343984, + "learning_rate": 0.03360872604138724, + "loss": 0.2485, + "num_input_tokens_seen": 23878920, + "step": 31310 + }, + { + "epoch": 65.1039501039501, + "grad_norm": 0.00011977836402365938, + "learning_rate": 0.03357157756630034, + "loss": 0.286, + "num_input_tokens_seen": 23882728, + "step": 31315 + }, + { + "epoch": 65.11434511434511, + "grad_norm": 0.0008518590475432575, + "learning_rate": 0.033534447045939365, + "loss": 0.2759, + "num_input_tokens_seen": 23886632, + "step": 31320 + }, + { + "epoch": 65.12474012474013, + "grad_norm": 0.00042627108632586896, + "learning_rate": 0.03349733448603026, + "loss": 0.2472, + "num_input_tokens_seen": 23890504, + "step": 31325 + }, + { + "epoch": 65.13513513513513, + "grad_norm": 0.00016863194468896836, + "learning_rate": 0.03346023989229619, + "loss": 0.266, + "num_input_tokens_seen": 23894312, + "step": 31330 + }, + { + "epoch": 65.14553014553015, + "grad_norm": 0.0003911313833668828, + "learning_rate": 0.03342316327045769, + "loss": 0.2524, + "num_input_tokens_seen": 23898056, + "step": 31335 + }, + { + "epoch": 65.15592515592516, + "grad_norm": 0.00034798047272488475, + "learning_rate": 0.033386104626232385, + "loss": 0.2778, + "num_input_tokens_seen": 23901864, + "step": 31340 + }, + { + "epoch": 65.16632016632016, + "grad_norm": 0.00010456281597726047, + "learning_rate": 0.03334906396533525, + "loss": 0.2606, + "num_input_tokens_seen": 23905704, + "step": 31345 + }, + { + "epoch": 65.17671517671518, + "grad_norm": 0.0003668017452582717, + "learning_rate": 0.033312041293478326, + "loss": 0.2268, + "num_input_tokens_seen": 23909576, + "step": 31350 + }, + { + "epoch": 65.18711018711019, + "grad_norm": 0.00019780281581915915, + "learning_rate": 0.03327503661637103, + "loss": 0.2475, + "num_input_tokens_seen": 23913288, + "step": 31355 + }, + { + "epoch": 65.1975051975052, + "grad_norm": 0.0003783702850341797, + "learning_rate": 0.03323804993971998, + "loss": 0.2727, + "num_input_tokens_seen": 23917128, + "step": 31360 + }, + { + "epoch": 65.20790020790021, + "grad_norm": 0.0013048759428784251, + "learning_rate": 0.033201081269228924, + "loss": 0.2404, + "num_input_tokens_seen": 23920968, + "step": 31365 + }, + { + "epoch": 65.21829521829522, + "grad_norm": 0.00026464962866157293, + "learning_rate": 0.03316413061059895, + "loss": 0.2331, + "num_input_tokens_seen": 23924680, + "step": 31370 + }, + { + "epoch": 65.22869022869023, + "grad_norm": 0.00011672815890051425, + "learning_rate": 0.03312719796952827, + "loss": 0.2622, + "num_input_tokens_seen": 23928584, + "step": 31375 + }, + { + "epoch": 65.23908523908524, + "grad_norm": 0.000176707559148781, + "learning_rate": 0.03309028335171236, + "loss": 0.2595, + "num_input_tokens_seen": 23932296, + "step": 31380 + }, + { + "epoch": 65.24948024948024, + "grad_norm": 0.0005229840753600001, + "learning_rate": 0.03305338676284398, + "loss": 0.2625, + "num_input_tokens_seen": 23936040, + "step": 31385 + }, + { + "epoch": 65.25987525987526, + "grad_norm": 0.000316180958179757, + "learning_rate": 0.03301650820861296, + "loss": 0.2847, + "num_input_tokens_seen": 23939784, + "step": 31390 + }, + { + "epoch": 65.27027027027027, + "grad_norm": 0.00020591021166183054, + "learning_rate": 0.03297964769470652, + "loss": 0.2757, + "num_input_tokens_seen": 23943720, + "step": 31395 + }, + { + "epoch": 65.28066528066527, + "grad_norm": 0.00044366164365783334, + "learning_rate": 0.032942805226808945, + "loss": 0.2555, + "num_input_tokens_seen": 23947688, + "step": 31400 + }, + { + "epoch": 65.28066528066527, + "eval_loss": 0.2505023181438446, + "eval_runtime": 13.4141, + "eval_samples_per_second": 63.814, + "eval_steps_per_second": 15.953, + "num_input_tokens_seen": 23947688, + "step": 31400 + }, + { + "epoch": 65.29106029106029, + "grad_norm": 0.00042196191498078406, + "learning_rate": 0.03290598081060187, + "loss": 0.259, + "num_input_tokens_seen": 23951464, + "step": 31405 + }, + { + "epoch": 65.3014553014553, + "grad_norm": 0.00043268216541036963, + "learning_rate": 0.03286917445176407, + "loss": 0.2726, + "num_input_tokens_seen": 23955144, + "step": 31410 + }, + { + "epoch": 65.3118503118503, + "grad_norm": 0.00017379199562128633, + "learning_rate": 0.032832386155971456, + "loss": 0.2837, + "num_input_tokens_seen": 23958888, + "step": 31415 + }, + { + "epoch": 65.32224532224532, + "grad_norm": 0.00010136065247934312, + "learning_rate": 0.032795615928897334, + "loss": 0.2562, + "num_input_tokens_seen": 23962536, + "step": 31420 + }, + { + "epoch": 65.33264033264034, + "grad_norm": 4.883684596279636e-05, + "learning_rate": 0.03275886377621215, + "loss": 0.2529, + "num_input_tokens_seen": 23966184, + "step": 31425 + }, + { + "epoch": 65.34303534303534, + "grad_norm": 0.00029319716850295663, + "learning_rate": 0.03272212970358348, + "loss": 0.2743, + "num_input_tokens_seen": 23969960, + "step": 31430 + }, + { + "epoch": 65.35343035343035, + "grad_norm": 7.706792530370876e-05, + "learning_rate": 0.032685413716676215, + "loss": 0.2501, + "num_input_tokens_seen": 23973832, + "step": 31435 + }, + { + "epoch": 65.36382536382537, + "grad_norm": 7.348047074628994e-05, + "learning_rate": 0.032648715821152474, + "loss": 0.2925, + "num_input_tokens_seen": 23977608, + "step": 31440 + }, + { + "epoch": 65.37422037422037, + "grad_norm": 0.00039305517566390336, + "learning_rate": 0.03261203602267143, + "loss": 0.2632, + "num_input_tokens_seen": 23981480, + "step": 31445 + }, + { + "epoch": 65.38461538461539, + "grad_norm": 0.0002983359736390412, + "learning_rate": 0.03257537432688966, + "loss": 0.276, + "num_input_tokens_seen": 23985288, + "step": 31450 + }, + { + "epoch": 65.39501039501039, + "grad_norm": 0.00035976088838651776, + "learning_rate": 0.03253873073946077, + "loss": 0.2736, + "num_input_tokens_seen": 23989064, + "step": 31455 + }, + { + "epoch": 65.4054054054054, + "grad_norm": 0.0003426525217946619, + "learning_rate": 0.03250210526603572, + "loss": 0.2422, + "num_input_tokens_seen": 23992808, + "step": 31460 + }, + { + "epoch": 65.41580041580042, + "grad_norm": 0.0002808624121826142, + "learning_rate": 0.03246549791226266, + "loss": 0.2662, + "num_input_tokens_seen": 23996552, + "step": 31465 + }, + { + "epoch": 65.42619542619542, + "grad_norm": 0.0008905503782443702, + "learning_rate": 0.03242890868378679, + "loss": 0.2898, + "num_input_tokens_seen": 24000264, + "step": 31470 + }, + { + "epoch": 65.43659043659044, + "grad_norm": 0.0008628521463833749, + "learning_rate": 0.03239233758625074, + "loss": 0.2732, + "num_input_tokens_seen": 24003976, + "step": 31475 + }, + { + "epoch": 65.44698544698545, + "grad_norm": 0.0001316967827733606, + "learning_rate": 0.032355784625294204, + "loss": 0.2615, + "num_input_tokens_seen": 24007752, + "step": 31480 + }, + { + "epoch": 65.45738045738045, + "grad_norm": 6.294428021647036e-05, + "learning_rate": 0.03231924980655402, + "loss": 0.2633, + "num_input_tokens_seen": 24011624, + "step": 31485 + }, + { + "epoch": 65.46777546777547, + "grad_norm": 0.0001941964728757739, + "learning_rate": 0.032282733135664446, + "loss": 0.2525, + "num_input_tokens_seen": 24015592, + "step": 31490 + }, + { + "epoch": 65.47817047817048, + "grad_norm": 7.37763402867131e-05, + "learning_rate": 0.03224623461825669, + "loss": 0.267, + "num_input_tokens_seen": 24019368, + "step": 31495 + }, + { + "epoch": 65.48856548856548, + "grad_norm": 0.0001615682995179668, + "learning_rate": 0.03220975425995937, + "loss": 0.2567, + "num_input_tokens_seen": 24023432, + "step": 31500 + }, + { + "epoch": 65.4989604989605, + "grad_norm": 0.00030263225198723376, + "learning_rate": 0.032173292066398206, + "loss": 0.2823, + "num_input_tokens_seen": 24027240, + "step": 31505 + }, + { + "epoch": 65.50935550935552, + "grad_norm": 0.00017948824097402394, + "learning_rate": 0.03213684804319606, + "loss": 0.2415, + "num_input_tokens_seen": 24031112, + "step": 31510 + }, + { + "epoch": 65.51975051975052, + "grad_norm": 0.0005189067451283336, + "learning_rate": 0.03210042219597312, + "loss": 0.2519, + "num_input_tokens_seen": 24034792, + "step": 31515 + }, + { + "epoch": 65.53014553014553, + "grad_norm": 0.0011047226144000888, + "learning_rate": 0.03206401453034675, + "loss": 0.2607, + "num_input_tokens_seen": 24038696, + "step": 31520 + }, + { + "epoch": 65.54054054054055, + "grad_norm": 0.0005004229606129229, + "learning_rate": 0.03202762505193136, + "loss": 0.2587, + "num_input_tokens_seen": 24042504, + "step": 31525 + }, + { + "epoch": 65.55093555093555, + "grad_norm": 0.00023637371486984193, + "learning_rate": 0.031991253766338754, + "loss": 0.275, + "num_input_tokens_seen": 24046344, + "step": 31530 + }, + { + "epoch": 65.56133056133056, + "grad_norm": 0.000183104959432967, + "learning_rate": 0.03195490067917778, + "loss": 0.2674, + "num_input_tokens_seen": 24050152, + "step": 31535 + }, + { + "epoch": 65.57172557172557, + "grad_norm": 8.366480324184522e-05, + "learning_rate": 0.03191856579605461, + "loss": 0.2332, + "num_input_tokens_seen": 24053832, + "step": 31540 + }, + { + "epoch": 65.58212058212058, + "grad_norm": 9.493053221376613e-05, + "learning_rate": 0.031882249122572454, + "loss": 0.269, + "num_input_tokens_seen": 24057640, + "step": 31545 + }, + { + "epoch": 65.5925155925156, + "grad_norm": 0.00033328335848636925, + "learning_rate": 0.03184595066433188, + "loss": 0.2791, + "num_input_tokens_seen": 24061576, + "step": 31550 + }, + { + "epoch": 65.6029106029106, + "grad_norm": 0.0005416363710537553, + "learning_rate": 0.03180967042693049, + "loss": 0.2388, + "num_input_tokens_seen": 24065384, + "step": 31555 + }, + { + "epoch": 65.61330561330561, + "grad_norm": 0.0009904925245791674, + "learning_rate": 0.03177340841596323, + "loss": 0.2642, + "num_input_tokens_seen": 24069256, + "step": 31560 + }, + { + "epoch": 65.62370062370063, + "grad_norm": 8.842509123496711e-05, + "learning_rate": 0.03173716463702209, + "loss": 0.2589, + "num_input_tokens_seen": 24073000, + "step": 31565 + }, + { + "epoch": 65.63409563409563, + "grad_norm": 0.00023978106037247926, + "learning_rate": 0.03170093909569638, + "loss": 0.2735, + "num_input_tokens_seen": 24076776, + "step": 31570 + }, + { + "epoch": 65.64449064449065, + "grad_norm": 8.37829356896691e-05, + "learning_rate": 0.03166473179757246, + "loss": 0.2379, + "num_input_tokens_seen": 24080520, + "step": 31575 + }, + { + "epoch": 65.65488565488566, + "grad_norm": 6.53962924843654e-05, + "learning_rate": 0.031628542748234005, + "loss": 0.2812, + "num_input_tokens_seen": 24084200, + "step": 31580 + }, + { + "epoch": 65.66528066528066, + "grad_norm": 0.00028854614356532693, + "learning_rate": 0.03159237195326184, + "loss": 0.2864, + "num_input_tokens_seen": 24088168, + "step": 31585 + }, + { + "epoch": 65.67567567567568, + "grad_norm": 0.00015648029511794448, + "learning_rate": 0.031556219418233875, + "loss": 0.2692, + "num_input_tokens_seen": 24092072, + "step": 31590 + }, + { + "epoch": 65.68607068607069, + "grad_norm": 0.00015175908629316837, + "learning_rate": 0.03152008514872533, + "loss": 0.2589, + "num_input_tokens_seen": 24095912, + "step": 31595 + }, + { + "epoch": 65.6964656964657, + "grad_norm": 0.00013459622277878225, + "learning_rate": 0.03148396915030862, + "loss": 0.2635, + "num_input_tokens_seen": 24099432, + "step": 31600 + }, + { + "epoch": 65.6964656964657, + "eval_loss": 0.2520678639411926, + "eval_runtime": 13.4145, + "eval_samples_per_second": 63.811, + "eval_steps_per_second": 15.953, + "num_input_tokens_seen": 24099432, + "step": 31600 + }, + { + "epoch": 65.70686070686071, + "grad_norm": 0.00014599690621253103, + "learning_rate": 0.03144787142855318, + "loss": 0.2675, + "num_input_tokens_seen": 24103304, + "step": 31605 + }, + { + "epoch": 65.71725571725571, + "grad_norm": 5.4997188271954656e-05, + "learning_rate": 0.031411791989025835, + "loss": 0.2709, + "num_input_tokens_seen": 24107144, + "step": 31610 + }, + { + "epoch": 65.72765072765073, + "grad_norm": 0.0001864663790911436, + "learning_rate": 0.031375730837290394, + "loss": 0.2567, + "num_input_tokens_seen": 24110984, + "step": 31615 + }, + { + "epoch": 65.73804573804574, + "grad_norm": 0.0001474396703997627, + "learning_rate": 0.031339687978908015, + "loss": 0.2641, + "num_input_tokens_seen": 24114760, + "step": 31620 + }, + { + "epoch": 65.74844074844074, + "grad_norm": 0.0005662969779223204, + "learning_rate": 0.03130366341943694, + "loss": 0.2626, + "num_input_tokens_seen": 24118568, + "step": 31625 + }, + { + "epoch": 65.75883575883576, + "grad_norm": 0.0003027894999831915, + "learning_rate": 0.031267657164432555, + "loss": 0.2869, + "num_input_tokens_seen": 24122376, + "step": 31630 + }, + { + "epoch": 65.76923076923077, + "grad_norm": 0.0007666753372177482, + "learning_rate": 0.03123166921944752, + "loss": 0.2793, + "num_input_tokens_seen": 24126376, + "step": 31635 + }, + { + "epoch": 65.77962577962577, + "grad_norm": 0.0009008229244500399, + "learning_rate": 0.031195699590031666, + "loss": 0.2781, + "num_input_tokens_seen": 24130248, + "step": 31640 + }, + { + "epoch": 65.79002079002079, + "grad_norm": 0.00015835900558158755, + "learning_rate": 0.031159748281731885, + "loss": 0.2746, + "num_input_tokens_seen": 24134056, + "step": 31645 + }, + { + "epoch": 65.8004158004158, + "grad_norm": 0.0001646050368435681, + "learning_rate": 0.031123815300092394, + "loss": 0.2703, + "num_input_tokens_seen": 24137960, + "step": 31650 + }, + { + "epoch": 65.8108108108108, + "grad_norm": 0.0004070030990988016, + "learning_rate": 0.031087900650654424, + "loss": 0.2657, + "num_input_tokens_seen": 24141736, + "step": 31655 + }, + { + "epoch": 65.82120582120582, + "grad_norm": 0.0005975066451355815, + "learning_rate": 0.031052004338956534, + "loss": 0.2381, + "num_input_tokens_seen": 24145416, + "step": 31660 + }, + { + "epoch": 65.83160083160084, + "grad_norm": 0.00011681431351462379, + "learning_rate": 0.031016126370534407, + "loss": 0.2582, + "num_input_tokens_seen": 24149096, + "step": 31665 + }, + { + "epoch": 65.84199584199584, + "grad_norm": 0.0004791850515175611, + "learning_rate": 0.030980266750920804, + "loss": 0.2788, + "num_input_tokens_seen": 24152904, + "step": 31670 + }, + { + "epoch": 65.85239085239085, + "grad_norm": 0.0003924318589270115, + "learning_rate": 0.030944425485645747, + "loss": 0.2599, + "num_input_tokens_seen": 24156680, + "step": 31675 + }, + { + "epoch": 65.86278586278586, + "grad_norm": 0.0002755174064077437, + "learning_rate": 0.03090860258023647, + "loss": 0.2404, + "num_input_tokens_seen": 24160328, + "step": 31680 + }, + { + "epoch": 65.87318087318087, + "grad_norm": 0.0006860743742436171, + "learning_rate": 0.030872798040217236, + "loss": 0.2804, + "num_input_tokens_seen": 24164104, + "step": 31685 + }, + { + "epoch": 65.88357588357589, + "grad_norm": 0.00036432233173400164, + "learning_rate": 0.03083701187110964, + "loss": 0.2637, + "num_input_tokens_seen": 24168008, + "step": 31690 + }, + { + "epoch": 65.89397089397089, + "grad_norm": 0.0005987982731312513, + "learning_rate": 0.030801244078432294, + "loss": 0.2408, + "num_input_tokens_seen": 24171848, + "step": 31695 + }, + { + "epoch": 65.9043659043659, + "grad_norm": 0.0009505784255452454, + "learning_rate": 0.030765494667701024, + "loss": 0.2704, + "num_input_tokens_seen": 24175560, + "step": 31700 + }, + { + "epoch": 65.91476091476092, + "grad_norm": 0.0006386747700162232, + "learning_rate": 0.030729763644428913, + "loss": 0.2357, + "num_input_tokens_seen": 24179336, + "step": 31705 + }, + { + "epoch": 65.92515592515592, + "grad_norm": 0.00023983839491847903, + "learning_rate": 0.030694051014126048, + "loss": 0.2712, + "num_input_tokens_seen": 24183144, + "step": 31710 + }, + { + "epoch": 65.93555093555094, + "grad_norm": 0.00028881296748295426, + "learning_rate": 0.030658356782299792, + "loss": 0.2641, + "num_input_tokens_seen": 24186920, + "step": 31715 + }, + { + "epoch": 65.94594594594595, + "grad_norm": 0.00025892938720062375, + "learning_rate": 0.030622680954454726, + "loss": 0.2747, + "num_input_tokens_seen": 24190792, + "step": 31720 + }, + { + "epoch": 65.95634095634095, + "grad_norm": 0.001398879219777882, + "learning_rate": 0.030587023536092398, + "loss": 0.272, + "num_input_tokens_seen": 24194472, + "step": 31725 + }, + { + "epoch": 65.96673596673597, + "grad_norm": 8.295443694805726e-05, + "learning_rate": 0.03055138453271171, + "loss": 0.2716, + "num_input_tokens_seen": 24198376, + "step": 31730 + }, + { + "epoch": 65.97713097713098, + "grad_norm": 0.0006298432708717883, + "learning_rate": 0.03051576394980858, + "loss": 0.2552, + "num_input_tokens_seen": 24202088, + "step": 31735 + }, + { + "epoch": 65.98752598752598, + "grad_norm": 0.0005156004917807877, + "learning_rate": 0.030480161792876187, + "loss": 0.2597, + "num_input_tokens_seen": 24205928, + "step": 31740 + }, + { + "epoch": 65.997920997921, + "grad_norm": 0.00019751819490920752, + "learning_rate": 0.030444578067404846, + "loss": 0.2539, + "num_input_tokens_seen": 24209832, + "step": 31745 + }, + { + "epoch": 66.00831600831602, + "grad_norm": 0.0001847275416366756, + "learning_rate": 0.030409012778881975, + "loss": 0.2679, + "num_input_tokens_seen": 24213408, + "step": 31750 + }, + { + "epoch": 66.01871101871102, + "grad_norm": 0.00043415476102381945, + "learning_rate": 0.030373465932792235, + "loss": 0.2394, + "num_input_tokens_seen": 24217344, + "step": 31755 + }, + { + "epoch": 66.02910602910603, + "grad_norm": 8.959716069512069e-05, + "learning_rate": 0.030337937534617342, + "loss": 0.2742, + "num_input_tokens_seen": 24221184, + "step": 31760 + }, + { + "epoch": 66.03950103950103, + "grad_norm": 0.0001079543653759174, + "learning_rate": 0.030302427589836277, + "loss": 0.2621, + "num_input_tokens_seen": 24224960, + "step": 31765 + }, + { + "epoch": 66.04989604989605, + "grad_norm": 0.0005546874017454684, + "learning_rate": 0.030266936103925095, + "loss": 0.2638, + "num_input_tokens_seen": 24228704, + "step": 31770 + }, + { + "epoch": 66.06029106029106, + "grad_norm": 0.0001257527037523687, + "learning_rate": 0.030231463082356982, + "loss": 0.2845, + "num_input_tokens_seen": 24232352, + "step": 31775 + }, + { + "epoch": 66.07068607068607, + "grad_norm": 0.0005177812417969108, + "learning_rate": 0.030196008530602367, + "loss": 0.2613, + "num_input_tokens_seen": 24236160, + "step": 31780 + }, + { + "epoch": 66.08108108108108, + "grad_norm": 0.00020069188030902296, + "learning_rate": 0.030160572454128842, + "loss": 0.251, + "num_input_tokens_seen": 24239840, + "step": 31785 + }, + { + "epoch": 66.0914760914761, + "grad_norm": 9.035712719196454e-05, + "learning_rate": 0.03012515485840098, + "loss": 0.2419, + "num_input_tokens_seen": 24243648, + "step": 31790 + }, + { + "epoch": 66.1018711018711, + "grad_norm": 0.0005617070128209889, + "learning_rate": 0.030089755748880734, + "loss": 0.2447, + "num_input_tokens_seen": 24247360, + "step": 31795 + }, + { + "epoch": 66.11226611226611, + "grad_norm": 0.0003881428565364331, + "learning_rate": 0.030054375131027003, + "loss": 0.2697, + "num_input_tokens_seen": 24251200, + "step": 31800 + }, + { + "epoch": 66.11226611226611, + "eval_loss": 0.24743887782096863, + "eval_runtime": 13.3938, + "eval_samples_per_second": 63.91, + "eval_steps_per_second": 15.977, + "num_input_tokens_seen": 24251200, + "step": 31800 + }, + { + "epoch": 66.12266112266113, + "grad_norm": 0.0002553602389525622, + "learning_rate": 0.030019013010295942, + "loss": 0.2487, + "num_input_tokens_seen": 24254848, + "step": 31805 + }, + { + "epoch": 66.13305613305613, + "grad_norm": 0.00024426504387520254, + "learning_rate": 0.029983669392140897, + "loss": 0.2614, + "num_input_tokens_seen": 24258752, + "step": 31810 + }, + { + "epoch": 66.14345114345114, + "grad_norm": 0.0008571365615352988, + "learning_rate": 0.029948344282012217, + "loss": 0.2402, + "num_input_tokens_seen": 24262560, + "step": 31815 + }, + { + "epoch": 66.15384615384616, + "grad_norm": 0.0002636099234223366, + "learning_rate": 0.029913037685357507, + "loss": 0.2807, + "num_input_tokens_seen": 24266464, + "step": 31820 + }, + { + "epoch": 66.16424116424116, + "grad_norm": 0.00023669969232287258, + "learning_rate": 0.029877749607621528, + "loss": 0.2622, + "num_input_tokens_seen": 24270304, + "step": 31825 + }, + { + "epoch": 66.17463617463618, + "grad_norm": 0.00017646161722950637, + "learning_rate": 0.029842480054246077, + "loss": 0.2308, + "num_input_tokens_seen": 24274080, + "step": 31830 + }, + { + "epoch": 66.18503118503118, + "grad_norm": 0.00021775079949293286, + "learning_rate": 0.02980722903067022, + "loss": 0.2647, + "num_input_tokens_seen": 24277920, + "step": 31835 + }, + { + "epoch": 66.1954261954262, + "grad_norm": 0.0003477302088867873, + "learning_rate": 0.029771996542330113, + "loss": 0.2751, + "num_input_tokens_seen": 24281760, + "step": 31840 + }, + { + "epoch": 66.20582120582121, + "grad_norm": 0.00024201312044169754, + "learning_rate": 0.029736782594658954, + "loss": 0.2653, + "num_input_tokens_seen": 24285728, + "step": 31845 + }, + { + "epoch": 66.21621621621621, + "grad_norm": 0.00037786460597999394, + "learning_rate": 0.029701587193087284, + "loss": 0.2597, + "num_input_tokens_seen": 24289632, + "step": 31850 + }, + { + "epoch": 66.22661122661123, + "grad_norm": 0.00016534920723643154, + "learning_rate": 0.0296664103430426, + "loss": 0.2595, + "num_input_tokens_seen": 24293504, + "step": 31855 + }, + { + "epoch": 66.23700623700624, + "grad_norm": 0.00017315348668489605, + "learning_rate": 0.029631252049949652, + "loss": 0.2762, + "num_input_tokens_seen": 24297344, + "step": 31860 + }, + { + "epoch": 66.24740124740124, + "grad_norm": 0.0012824817094951868, + "learning_rate": 0.02959611231923031, + "loss": 0.2526, + "num_input_tokens_seen": 24301280, + "step": 31865 + }, + { + "epoch": 66.25779625779626, + "grad_norm": 0.00015988053928595036, + "learning_rate": 0.029560991156303507, + "loss": 0.265, + "num_input_tokens_seen": 24305408, + "step": 31870 + }, + { + "epoch": 66.26819126819127, + "grad_norm": 0.00023079382663127035, + "learning_rate": 0.02952588856658544, + "loss": 0.2746, + "num_input_tokens_seen": 24309280, + "step": 31875 + }, + { + "epoch": 66.27858627858627, + "grad_norm": 0.0002841517562046647, + "learning_rate": 0.029490804555489296, + "loss": 0.2556, + "num_input_tokens_seen": 24312864, + "step": 31880 + }, + { + "epoch": 66.28898128898129, + "grad_norm": 0.0015741126844659448, + "learning_rate": 0.029455739128425484, + "loss": 0.2719, + "num_input_tokens_seen": 24316576, + "step": 31885 + }, + { + "epoch": 66.2993762993763, + "grad_norm": 0.00044265022734180093, + "learning_rate": 0.029420692290801607, + "loss": 0.2593, + "num_input_tokens_seen": 24320384, + "step": 31890 + }, + { + "epoch": 66.3097713097713, + "grad_norm": 0.00036997589631937444, + "learning_rate": 0.02938566404802223, + "loss": 0.2677, + "num_input_tokens_seen": 24324256, + "step": 31895 + }, + { + "epoch": 66.32016632016632, + "grad_norm": 0.0003390981873963028, + "learning_rate": 0.029350654405489195, + "loss": 0.2797, + "num_input_tokens_seen": 24328000, + "step": 31900 + }, + { + "epoch": 66.33056133056132, + "grad_norm": 0.0001893117296276614, + "learning_rate": 0.02931566336860145, + "loss": 0.266, + "num_input_tokens_seen": 24331936, + "step": 31905 + }, + { + "epoch": 66.34095634095634, + "grad_norm": 0.000436124304542318, + "learning_rate": 0.02928069094275505, + "loss": 0.2683, + "num_input_tokens_seen": 24335680, + "step": 31910 + }, + { + "epoch": 66.35135135135135, + "grad_norm": 0.0002594403631519526, + "learning_rate": 0.02924573713334314, + "loss": 0.2758, + "num_input_tokens_seen": 24339456, + "step": 31915 + }, + { + "epoch": 66.36174636174636, + "grad_norm": 0.0004278372216504067, + "learning_rate": 0.02921080194575603, + "loss": 0.227, + "num_input_tokens_seen": 24343264, + "step": 31920 + }, + { + "epoch": 66.37214137214137, + "grad_norm": 0.0004695228999480605, + "learning_rate": 0.029175885385381177, + "loss": 0.2849, + "num_input_tokens_seen": 24347040, + "step": 31925 + }, + { + "epoch": 66.38253638253639, + "grad_norm": 0.00029203714802861214, + "learning_rate": 0.029140987457603223, + "loss": 0.2881, + "num_input_tokens_seen": 24350816, + "step": 31930 + }, + { + "epoch": 66.39293139293139, + "grad_norm": 0.0006393062067218125, + "learning_rate": 0.029106108167803763, + "loss": 0.2925, + "num_input_tokens_seen": 24354720, + "step": 31935 + }, + { + "epoch": 66.4033264033264, + "grad_norm": 0.000825194176286459, + "learning_rate": 0.029071247521361674, + "loss": 0.2599, + "num_input_tokens_seen": 24358528, + "step": 31940 + }, + { + "epoch": 66.41372141372142, + "grad_norm": 4.994422852178104e-05, + "learning_rate": 0.029036405523652945, + "loss": 0.2655, + "num_input_tokens_seen": 24362336, + "step": 31945 + }, + { + "epoch": 66.42411642411642, + "grad_norm": 0.0003835965762846172, + "learning_rate": 0.029001582180050577, + "loss": 0.2558, + "num_input_tokens_seen": 24366208, + "step": 31950 + }, + { + "epoch": 66.43451143451144, + "grad_norm": 0.00031185458647087216, + "learning_rate": 0.02896677749592482, + "loss": 0.2455, + "num_input_tokens_seen": 24370048, + "step": 31955 + }, + { + "epoch": 66.44490644490645, + "grad_norm": 0.0009959047893062234, + "learning_rate": 0.028931991476642938, + "loss": 0.2773, + "num_input_tokens_seen": 24373952, + "step": 31960 + }, + { + "epoch": 66.45530145530145, + "grad_norm": 0.00012873091327492148, + "learning_rate": 0.028897224127569412, + "loss": 0.2712, + "num_input_tokens_seen": 24377728, + "step": 31965 + }, + { + "epoch": 66.46569646569647, + "grad_norm": 0.0004577824438456446, + "learning_rate": 0.028862475454065832, + "loss": 0.2457, + "num_input_tokens_seen": 24381696, + "step": 31970 + }, + { + "epoch": 66.47609147609148, + "grad_norm": 0.0003507114597596228, + "learning_rate": 0.028827745461490806, + "loss": 0.2685, + "num_input_tokens_seen": 24385664, + "step": 31975 + }, + { + "epoch": 66.48648648648648, + "grad_norm": 0.0003083197516389191, + "learning_rate": 0.028793034155200212, + "loss": 0.295, + "num_input_tokens_seen": 24389536, + "step": 31980 + }, + { + "epoch": 66.4968814968815, + "grad_norm": 0.00040116204763762653, + "learning_rate": 0.028758341540546944, + "loss": 0.2801, + "num_input_tokens_seen": 24393344, + "step": 31985 + }, + { + "epoch": 66.5072765072765, + "grad_norm": 0.0008647387730889022, + "learning_rate": 0.02872366762288098, + "loss": 0.2591, + "num_input_tokens_seen": 24397280, + "step": 31990 + }, + { + "epoch": 66.51767151767152, + "grad_norm": 0.00014302368799690157, + "learning_rate": 0.028689012407549567, + "loss": 0.2456, + "num_input_tokens_seen": 24401024, + "step": 31995 + }, + { + "epoch": 66.52806652806653, + "grad_norm": 0.00019364868057891726, + "learning_rate": 0.028654375899896892, + "loss": 0.2765, + "num_input_tokens_seen": 24404736, + "step": 32000 + }, + { + "epoch": 66.52806652806653, + "eval_loss": 0.24981892108917236, + "eval_runtime": 13.3965, + "eval_samples_per_second": 63.897, + "eval_steps_per_second": 15.974, + "num_input_tokens_seen": 24404736, + "step": 32000 + }, + { + "epoch": 66.53846153846153, + "grad_norm": 0.00015548663213849068, + "learning_rate": 0.02861975810526437, + "loss": 0.2516, + "num_input_tokens_seen": 24408672, + "step": 32005 + }, + { + "epoch": 66.54885654885655, + "grad_norm": 0.0002616024576127529, + "learning_rate": 0.02858515902899056, + "loss": 0.2473, + "num_input_tokens_seen": 24412416, + "step": 32010 + }, + { + "epoch": 66.55925155925156, + "grad_norm": 0.00019993921159766614, + "learning_rate": 0.028550578676410976, + "loss": 0.2418, + "num_input_tokens_seen": 24416032, + "step": 32015 + }, + { + "epoch": 66.56964656964657, + "grad_norm": 0.00020499563834164292, + "learning_rate": 0.02851601705285837, + "loss": 0.2864, + "num_input_tokens_seen": 24419904, + "step": 32020 + }, + { + "epoch": 66.58004158004158, + "grad_norm": 7.819920574547723e-05, + "learning_rate": 0.028481474163662666, + "loss": 0.2685, + "num_input_tokens_seen": 24423776, + "step": 32025 + }, + { + "epoch": 66.5904365904366, + "grad_norm": 0.00013305859465617687, + "learning_rate": 0.028446950014150683, + "loss": 0.2775, + "num_input_tokens_seen": 24427424, + "step": 32030 + }, + { + "epoch": 66.6008316008316, + "grad_norm": 0.0003858963609673083, + "learning_rate": 0.028412444609646596, + "loss": 0.2686, + "num_input_tokens_seen": 24431200, + "step": 32035 + }, + { + "epoch": 66.61122661122661, + "grad_norm": 0.0006701680831611156, + "learning_rate": 0.028377957955471465, + "loss": 0.2545, + "num_input_tokens_seen": 24435136, + "step": 32040 + }, + { + "epoch": 66.62162162162163, + "grad_norm": 0.00026531622279435396, + "learning_rate": 0.0283434900569436, + "loss": 0.2767, + "num_input_tokens_seen": 24438880, + "step": 32045 + }, + { + "epoch": 66.63201663201663, + "grad_norm": 0.0010543946409597993, + "learning_rate": 0.028309040919378456, + "loss": 0.2724, + "num_input_tokens_seen": 24443072, + "step": 32050 + }, + { + "epoch": 66.64241164241164, + "grad_norm": 0.00019822463218588382, + "learning_rate": 0.02827461054808848, + "loss": 0.2624, + "num_input_tokens_seen": 24446816, + "step": 32055 + }, + { + "epoch": 66.65280665280665, + "grad_norm": 0.0001548689033370465, + "learning_rate": 0.028240198948383186, + "loss": 0.2672, + "num_input_tokens_seen": 24450784, + "step": 32060 + }, + { + "epoch": 66.66320166320166, + "grad_norm": 0.00027787633007392287, + "learning_rate": 0.028205806125569402, + "loss": 0.2705, + "num_input_tokens_seen": 24454528, + "step": 32065 + }, + { + "epoch": 66.67359667359668, + "grad_norm": 0.0015721704112365842, + "learning_rate": 0.028171432084950834, + "loss": 0.2673, + "num_input_tokens_seen": 24458336, + "step": 32070 + }, + { + "epoch": 66.68399168399168, + "grad_norm": 0.00023947693989612162, + "learning_rate": 0.028137076831828478, + "loss": 0.2583, + "num_input_tokens_seen": 24462144, + "step": 32075 + }, + { + "epoch": 66.6943866943867, + "grad_norm": 4.022300709038973e-05, + "learning_rate": 0.028102740371500238, + "loss": 0.2547, + "num_input_tokens_seen": 24465952, + "step": 32080 + }, + { + "epoch": 66.70478170478171, + "grad_norm": 0.00010132717579836026, + "learning_rate": 0.0280684227092613, + "loss": 0.2825, + "num_input_tokens_seen": 24469696, + "step": 32085 + }, + { + "epoch": 66.71517671517671, + "grad_norm": 3.9046226447680965e-05, + "learning_rate": 0.02803412385040392, + "loss": 0.271, + "num_input_tokens_seen": 24473440, + "step": 32090 + }, + { + "epoch": 66.72557172557173, + "grad_norm": 0.0005616469425149262, + "learning_rate": 0.027999843800217306, + "loss": 0.2755, + "num_input_tokens_seen": 24477248, + "step": 32095 + }, + { + "epoch": 66.73596673596674, + "grad_norm": 0.00030461253481917083, + "learning_rate": 0.027965582563987932, + "loss": 0.26, + "num_input_tokens_seen": 24481056, + "step": 32100 + }, + { + "epoch": 66.74636174636174, + "grad_norm": 0.00039138534339144826, + "learning_rate": 0.027931340146999346, + "loss": 0.2375, + "num_input_tokens_seen": 24484928, + "step": 32105 + }, + { + "epoch": 66.75675675675676, + "grad_norm": 0.0006079064914956689, + "learning_rate": 0.02789711655453208, + "loss": 0.2644, + "num_input_tokens_seen": 24488928, + "step": 32110 + }, + { + "epoch": 66.76715176715177, + "grad_norm": 0.0001235850650118664, + "learning_rate": 0.02786291179186392, + "loss": 0.2711, + "num_input_tokens_seen": 24492800, + "step": 32115 + }, + { + "epoch": 66.77754677754677, + "grad_norm": 0.00036211672704666853, + "learning_rate": 0.02782872586426961, + "loss": 0.2593, + "num_input_tokens_seen": 24496608, + "step": 32120 + }, + { + "epoch": 66.78794178794179, + "grad_norm": 6.564981595147401e-05, + "learning_rate": 0.027794558777021083, + "loss": 0.2652, + "num_input_tokens_seen": 24500352, + "step": 32125 + }, + { + "epoch": 66.7983367983368, + "grad_norm": 0.0001920598588185385, + "learning_rate": 0.02776041053538734, + "loss": 0.2473, + "num_input_tokens_seen": 24504032, + "step": 32130 + }, + { + "epoch": 66.8087318087318, + "grad_norm": 0.00030004591098986566, + "learning_rate": 0.027726281144634407, + "loss": 0.2475, + "num_input_tokens_seen": 24507872, + "step": 32135 + }, + { + "epoch": 66.81912681912682, + "grad_norm": 0.00015475442341994494, + "learning_rate": 0.02769217061002552, + "loss": 0.252, + "num_input_tokens_seen": 24511552, + "step": 32140 + }, + { + "epoch": 66.82952182952182, + "grad_norm": 5.0014648877549917e-05, + "learning_rate": 0.027658078936820967, + "loss": 0.2843, + "num_input_tokens_seen": 24515296, + "step": 32145 + }, + { + "epoch": 66.83991683991684, + "grad_norm": 0.0004331866221036762, + "learning_rate": 0.02762400613027805, + "loss": 0.2649, + "num_input_tokens_seen": 24519008, + "step": 32150 + }, + { + "epoch": 66.85031185031185, + "grad_norm": 0.0003696487983688712, + "learning_rate": 0.027589952195651295, + "loss": 0.2858, + "num_input_tokens_seen": 24522816, + "step": 32155 + }, + { + "epoch": 66.86070686070686, + "grad_norm": 0.00017552592908032238, + "learning_rate": 0.027555917138192186, + "loss": 0.2725, + "num_input_tokens_seen": 24526624, + "step": 32160 + }, + { + "epoch": 66.87110187110187, + "grad_norm": 0.00044517702190205455, + "learning_rate": 0.027521900963149375, + "loss": 0.2354, + "num_input_tokens_seen": 24530304, + "step": 32165 + }, + { + "epoch": 66.88149688149689, + "grad_norm": 0.000365311570931226, + "learning_rate": 0.027487903675768633, + "loss": 0.2397, + "num_input_tokens_seen": 24534144, + "step": 32170 + }, + { + "epoch": 66.89189189189189, + "grad_norm": 0.000583117944188416, + "learning_rate": 0.027453925281292677, + "loss": 0.2682, + "num_input_tokens_seen": 24537888, + "step": 32175 + }, + { + "epoch": 66.9022869022869, + "grad_norm": 0.00010460512567078695, + "learning_rate": 0.027419965784961475, + "loss": 0.2603, + "num_input_tokens_seen": 24541728, + "step": 32180 + }, + { + "epoch": 66.91268191268192, + "grad_norm": 0.00010335264960303903, + "learning_rate": 0.027386025192012015, + "loss": 0.2621, + "num_input_tokens_seen": 24545664, + "step": 32185 + }, + { + "epoch": 66.92307692307692, + "grad_norm": 0.0001855616719694808, + "learning_rate": 0.027352103507678277, + "loss": 0.2715, + "num_input_tokens_seen": 24549344, + "step": 32190 + }, + { + "epoch": 66.93347193347194, + "grad_norm": 0.00020409049466252327, + "learning_rate": 0.027318200737191527, + "loss": 0.2524, + "num_input_tokens_seen": 24553216, + "step": 32195 + }, + { + "epoch": 66.94386694386695, + "grad_norm": 0.0019398892764002085, + "learning_rate": 0.027284316885779935, + "loss": 0.2743, + "num_input_tokens_seen": 24557120, + "step": 32200 + }, + { + "epoch": 66.94386694386695, + "eval_loss": 0.24870139360427856, + "eval_runtime": 13.4109, + "eval_samples_per_second": 63.829, + "eval_steps_per_second": 15.957, + "num_input_tokens_seen": 24557120, + "step": 32200 + }, + { + "epoch": 66.95426195426195, + "grad_norm": 0.00022466557857114822, + "learning_rate": 0.027250451958668785, + "loss": 0.2894, + "num_input_tokens_seen": 24560832, + "step": 32205 + }, + { + "epoch": 66.96465696465697, + "grad_norm": 0.00018743482360150665, + "learning_rate": 0.027216605961080536, + "loss": 0.2712, + "num_input_tokens_seen": 24564640, + "step": 32210 + }, + { + "epoch": 66.97505197505197, + "grad_norm": 0.00018502064631320536, + "learning_rate": 0.02718277889823461, + "loss": 0.2615, + "num_input_tokens_seen": 24568544, + "step": 32215 + }, + { + "epoch": 66.98544698544698, + "grad_norm": 0.0003447768685873598, + "learning_rate": 0.027148970775347604, + "loss": 0.2551, + "num_input_tokens_seen": 24572352, + "step": 32220 + }, + { + "epoch": 66.995841995842, + "grad_norm": 0.0014508496969938278, + "learning_rate": 0.027115181597633174, + "loss": 0.2482, + "num_input_tokens_seen": 24576192, + "step": 32225 + }, + { + "epoch": 67.006237006237, + "grad_norm": 0.000303589622490108, + "learning_rate": 0.027081411370301976, + "loss": 0.2531, + "num_input_tokens_seen": 24579952, + "step": 32230 + }, + { + "epoch": 67.01663201663202, + "grad_norm": 0.00034578194026835263, + "learning_rate": 0.027047660098561875, + "loss": 0.2696, + "num_input_tokens_seen": 24583696, + "step": 32235 + }, + { + "epoch": 67.02702702702703, + "grad_norm": 0.00034689498716033995, + "learning_rate": 0.02701392778761766, + "loss": 0.238, + "num_input_tokens_seen": 24587504, + "step": 32240 + }, + { + "epoch": 67.03742203742203, + "grad_norm": 0.00010837531590368599, + "learning_rate": 0.02698021444267133, + "loss": 0.2666, + "num_input_tokens_seen": 24591280, + "step": 32245 + }, + { + "epoch": 67.04781704781705, + "grad_norm": 6.753968773409724e-05, + "learning_rate": 0.026946520068921915, + "loss": 0.2935, + "num_input_tokens_seen": 24595216, + "step": 32250 + }, + { + "epoch": 67.05821205821206, + "grad_norm": 0.00015085379709489644, + "learning_rate": 0.02691284467156547, + "loss": 0.2825, + "num_input_tokens_seen": 24599184, + "step": 32255 + }, + { + "epoch": 67.06860706860707, + "grad_norm": 0.0005276617011986673, + "learning_rate": 0.026879188255795182, + "loss": 0.2639, + "num_input_tokens_seen": 24603056, + "step": 32260 + }, + { + "epoch": 67.07900207900208, + "grad_norm": 0.0005873062764294446, + "learning_rate": 0.026845550826801328, + "loss": 0.2494, + "num_input_tokens_seen": 24606864, + "step": 32265 + }, + { + "epoch": 67.0893970893971, + "grad_norm": 0.00013601582031697035, + "learning_rate": 0.02681193238977121, + "loss": 0.2586, + "num_input_tokens_seen": 24610768, + "step": 32270 + }, + { + "epoch": 67.0997920997921, + "grad_norm": 0.00027289893478155136, + "learning_rate": 0.026778332949889145, + "loss": 0.2576, + "num_input_tokens_seen": 24614672, + "step": 32275 + }, + { + "epoch": 67.11018711018711, + "grad_norm": 0.0005473297787830234, + "learning_rate": 0.026744752512336673, + "loss": 0.2923, + "num_input_tokens_seen": 24618480, + "step": 32280 + }, + { + "epoch": 67.12058212058211, + "grad_norm": 4.2316532926633954e-05, + "learning_rate": 0.02671119108229225, + "loss": 0.2524, + "num_input_tokens_seen": 24622192, + "step": 32285 + }, + { + "epoch": 67.13097713097713, + "grad_norm": 0.0005457705701701343, + "learning_rate": 0.026677648664931556, + "loss": 0.2837, + "num_input_tokens_seen": 24626000, + "step": 32290 + }, + { + "epoch": 67.14137214137214, + "grad_norm": 9.365333971800283e-05, + "learning_rate": 0.026644125265427154, + "loss": 0.2595, + "num_input_tokens_seen": 24629904, + "step": 32295 + }, + { + "epoch": 67.15176715176715, + "grad_norm": 0.00046885712072253227, + "learning_rate": 0.026610620888948822, + "loss": 0.2512, + "num_input_tokens_seen": 24633584, + "step": 32300 + }, + { + "epoch": 67.16216216216216, + "grad_norm": 0.0002616364508867264, + "learning_rate": 0.026577135540663408, + "loss": 0.2613, + "num_input_tokens_seen": 24637424, + "step": 32305 + }, + { + "epoch": 67.17255717255718, + "grad_norm": 0.00028624775586649776, + "learning_rate": 0.026543669225734673, + "loss": 0.2809, + "num_input_tokens_seen": 24641200, + "step": 32310 + }, + { + "epoch": 67.18295218295218, + "grad_norm": 2.9039969376754016e-05, + "learning_rate": 0.02651022194932363, + "loss": 0.2681, + "num_input_tokens_seen": 24645008, + "step": 32315 + }, + { + "epoch": 67.1933471933472, + "grad_norm": 0.0010189771419391036, + "learning_rate": 0.026476793716588194, + "loss": 0.2543, + "num_input_tokens_seen": 24648880, + "step": 32320 + }, + { + "epoch": 67.20374220374221, + "grad_norm": 4.967190761817619e-05, + "learning_rate": 0.026443384532683467, + "loss": 0.2734, + "num_input_tokens_seen": 24652560, + "step": 32325 + }, + { + "epoch": 67.21413721413721, + "grad_norm": 0.0008077262318693101, + "learning_rate": 0.026409994402761584, + "loss": 0.2487, + "num_input_tokens_seen": 24656336, + "step": 32330 + }, + { + "epoch": 67.22453222453223, + "grad_norm": 0.0003014679823536426, + "learning_rate": 0.026376623331971653, + "loss": 0.2654, + "num_input_tokens_seen": 24660208, + "step": 32335 + }, + { + "epoch": 67.23492723492724, + "grad_norm": 0.00021866391762159765, + "learning_rate": 0.026343271325459997, + "loss": 0.2661, + "num_input_tokens_seen": 24664080, + "step": 32340 + }, + { + "epoch": 67.24532224532224, + "grad_norm": 0.00032542619737796485, + "learning_rate": 0.02630993838836987, + "loss": 0.2697, + "num_input_tokens_seen": 24667632, + "step": 32345 + }, + { + "epoch": 67.25571725571726, + "grad_norm": 0.00046098572784103453, + "learning_rate": 0.026276624525841584, + "loss": 0.2612, + "num_input_tokens_seen": 24671568, + "step": 32350 + }, + { + "epoch": 67.26611226611226, + "grad_norm": 0.00020881762611679733, + "learning_rate": 0.026243329743012637, + "loss": 0.2425, + "num_input_tokens_seen": 24675472, + "step": 32355 + }, + { + "epoch": 67.27650727650727, + "grad_norm": 0.0003716912760864943, + "learning_rate": 0.026210054045017438, + "loss": 0.2433, + "num_input_tokens_seen": 24679248, + "step": 32360 + }, + { + "epoch": 67.28690228690229, + "grad_norm": 5.221865649218671e-05, + "learning_rate": 0.02617679743698755, + "loss": 0.2802, + "num_input_tokens_seen": 24683056, + "step": 32365 + }, + { + "epoch": 67.29729729729729, + "grad_norm": 0.0004487780388444662, + "learning_rate": 0.02614355992405158, + "loss": 0.2839, + "num_input_tokens_seen": 24686736, + "step": 32370 + }, + { + "epoch": 67.3076923076923, + "grad_norm": 0.0006595671875402331, + "learning_rate": 0.026110341511335115, + "loss": 0.278, + "num_input_tokens_seen": 24690480, + "step": 32375 + }, + { + "epoch": 67.31808731808732, + "grad_norm": 0.0002938820980489254, + "learning_rate": 0.02607714220396093, + "loss": 0.275, + "num_input_tokens_seen": 24694256, + "step": 32380 + }, + { + "epoch": 67.32848232848232, + "grad_norm": 0.000242730719037354, + "learning_rate": 0.02604396200704869, + "loss": 0.2515, + "num_input_tokens_seen": 24698096, + "step": 32385 + }, + { + "epoch": 67.33887733887734, + "grad_norm": 0.00043907048529945314, + "learning_rate": 0.02601080092571523, + "loss": 0.2596, + "num_input_tokens_seen": 24701904, + "step": 32390 + }, + { + "epoch": 67.34927234927235, + "grad_norm": 6.275655323406681e-05, + "learning_rate": 0.025977658965074455, + "loss": 0.2681, + "num_input_tokens_seen": 24705712, + "step": 32395 + }, + { + "epoch": 67.35966735966736, + "grad_norm": 0.00016123673412948847, + "learning_rate": 0.02594453613023719, + "loss": 0.2775, + "num_input_tokens_seen": 24709616, + "step": 32400 + }, + { + "epoch": 67.35966735966736, + "eval_loss": 0.2504204511642456, + "eval_runtime": 13.3757, + "eval_samples_per_second": 63.997, + "eval_steps_per_second": 15.999, + "num_input_tokens_seen": 24709616, + "step": 32400 + }, + { + "epoch": 67.37006237006237, + "grad_norm": 0.0005676941364072263, + "learning_rate": 0.025911432426311443, + "loss": 0.2469, + "num_input_tokens_seen": 24713488, + "step": 32405 + }, + { + "epoch": 67.38045738045739, + "grad_norm": 0.00014894260675646365, + "learning_rate": 0.025878347858402234, + "loss": 0.2874, + "num_input_tokens_seen": 24717328, + "step": 32410 + }, + { + "epoch": 67.39085239085239, + "grad_norm": 0.00030095753027126193, + "learning_rate": 0.025845282431611598, + "loss": 0.2804, + "num_input_tokens_seen": 24721168, + "step": 32415 + }, + { + "epoch": 67.4012474012474, + "grad_norm": 0.00035193029907532036, + "learning_rate": 0.025812236151038608, + "loss": 0.2741, + "num_input_tokens_seen": 24724880, + "step": 32420 + }, + { + "epoch": 67.41164241164242, + "grad_norm": 0.00010354749974794686, + "learning_rate": 0.025779209021779468, + "loss": 0.2638, + "num_input_tokens_seen": 24728560, + "step": 32425 + }, + { + "epoch": 67.42203742203742, + "grad_norm": 0.0006758659146726131, + "learning_rate": 0.025746201048927324, + "loss": 0.2568, + "num_input_tokens_seen": 24732240, + "step": 32430 + }, + { + "epoch": 67.43243243243244, + "grad_norm": 9.829413465922698e-05, + "learning_rate": 0.025713212237572485, + "loss": 0.271, + "num_input_tokens_seen": 24736112, + "step": 32435 + }, + { + "epoch": 67.44282744282744, + "grad_norm": 0.0009382665739394724, + "learning_rate": 0.025680242592802164, + "loss": 0.2727, + "num_input_tokens_seen": 24739984, + "step": 32440 + }, + { + "epoch": 67.45322245322245, + "grad_norm": 0.0003254359762649983, + "learning_rate": 0.02564729211970073, + "loss": 0.2462, + "num_input_tokens_seen": 24743824, + "step": 32445 + }, + { + "epoch": 67.46361746361747, + "grad_norm": 0.0004774636763613671, + "learning_rate": 0.025614360823349617, + "loss": 0.2755, + "num_input_tokens_seen": 24747856, + "step": 32450 + }, + { + "epoch": 67.47401247401247, + "grad_norm": 0.0007760768639855087, + "learning_rate": 0.025581448708827146, + "loss": 0.254, + "num_input_tokens_seen": 24751600, + "step": 32455 + }, + { + "epoch": 67.48440748440748, + "grad_norm": 0.0002239678578916937, + "learning_rate": 0.025548555781208876, + "loss": 0.2394, + "num_input_tokens_seen": 24755280, + "step": 32460 + }, + { + "epoch": 67.4948024948025, + "grad_norm": 0.0011168154887855053, + "learning_rate": 0.02551568204556721, + "loss": 0.2905, + "num_input_tokens_seen": 24759120, + "step": 32465 + }, + { + "epoch": 67.5051975051975, + "grad_norm": 0.0004033400327898562, + "learning_rate": 0.02548282750697173, + "loss": 0.2322, + "num_input_tokens_seen": 24762960, + "step": 32470 + }, + { + "epoch": 67.51559251559252, + "grad_norm": 0.0005359951173886657, + "learning_rate": 0.02544999217048909, + "loss": 0.2612, + "num_input_tokens_seen": 24766800, + "step": 32475 + }, + { + "epoch": 67.52598752598753, + "grad_norm": 0.00034861263702623546, + "learning_rate": 0.025417176041182793, + "loss": 0.2685, + "num_input_tokens_seen": 24770832, + "step": 32480 + }, + { + "epoch": 67.53638253638253, + "grad_norm": 0.0002008568262681365, + "learning_rate": 0.025384379124113596, + "loss": 0.2678, + "num_input_tokens_seen": 24774576, + "step": 32485 + }, + { + "epoch": 67.54677754677755, + "grad_norm": 0.000246087001869455, + "learning_rate": 0.025351601424339124, + "loss": 0.2325, + "num_input_tokens_seen": 24778480, + "step": 32490 + }, + { + "epoch": 67.55717255717256, + "grad_norm": 0.00024455785751342773, + "learning_rate": 0.025318842946914184, + "loss": 0.247, + "num_input_tokens_seen": 24782224, + "step": 32495 + }, + { + "epoch": 67.56756756756756, + "grad_norm": 0.0003561387420631945, + "learning_rate": 0.025286103696890494, + "loss": 0.2718, + "num_input_tokens_seen": 24785968, + "step": 32500 + }, + { + "epoch": 67.57796257796258, + "grad_norm": 0.0010462590726092458, + "learning_rate": 0.025253383679316836, + "loss": 0.2729, + "num_input_tokens_seen": 24789744, + "step": 32505 + }, + { + "epoch": 67.58835758835758, + "grad_norm": 0.0003633305022958666, + "learning_rate": 0.025220682899239077, + "loss": 0.271, + "num_input_tokens_seen": 24793488, + "step": 32510 + }, + { + "epoch": 67.5987525987526, + "grad_norm": 0.0001137854196713306, + "learning_rate": 0.02518800136170013, + "loss": 0.264, + "num_input_tokens_seen": 24797424, + "step": 32515 + }, + { + "epoch": 67.60914760914761, + "grad_norm": 0.0008328195544891059, + "learning_rate": 0.02515533907173981, + "loss": 0.2957, + "num_input_tokens_seen": 24801424, + "step": 32520 + }, + { + "epoch": 67.61954261954261, + "grad_norm": 0.0006467492785304785, + "learning_rate": 0.025122696034395115, + "loss": 0.2696, + "num_input_tokens_seen": 24805328, + "step": 32525 + }, + { + "epoch": 67.62993762993763, + "grad_norm": 0.0005138764390721917, + "learning_rate": 0.025090072254700023, + "loss": 0.2668, + "num_input_tokens_seen": 24808944, + "step": 32530 + }, + { + "epoch": 67.64033264033264, + "grad_norm": 0.00035850287531502545, + "learning_rate": 0.025057467737685468, + "loss": 0.2633, + "num_input_tokens_seen": 24812848, + "step": 32535 + }, + { + "epoch": 67.65072765072765, + "grad_norm": 0.0002940946724265814, + "learning_rate": 0.025024882488379557, + "loss": 0.2671, + "num_input_tokens_seen": 24816496, + "step": 32540 + }, + { + "epoch": 67.66112266112266, + "grad_norm": 0.00015440104471053928, + "learning_rate": 0.02499231651180727, + "loss": 0.2492, + "num_input_tokens_seen": 24820304, + "step": 32545 + }, + { + "epoch": 67.67151767151768, + "grad_norm": 0.00019005945068784058, + "learning_rate": 0.024959769812990713, + "loss": 0.263, + "num_input_tokens_seen": 24824112, + "step": 32550 + }, + { + "epoch": 67.68191268191268, + "grad_norm": 0.0002867590810637921, + "learning_rate": 0.024927242396949045, + "loss": 0.2684, + "num_input_tokens_seen": 24827856, + "step": 32555 + }, + { + "epoch": 67.6923076923077, + "grad_norm": 0.00021186417143326253, + "learning_rate": 0.02489473426869836, + "loss": 0.2481, + "num_input_tokens_seen": 24831600, + "step": 32560 + }, + { + "epoch": 67.70270270270271, + "grad_norm": 0.00025059765903279185, + "learning_rate": 0.024862245433251776, + "loss": 0.2273, + "num_input_tokens_seen": 24835312, + "step": 32565 + }, + { + "epoch": 67.71309771309771, + "grad_norm": 0.0003760459367185831, + "learning_rate": 0.024829775895619577, + "loss": 0.2765, + "num_input_tokens_seen": 24839280, + "step": 32570 + }, + { + "epoch": 67.72349272349273, + "grad_norm": 0.0001189417889690958, + "learning_rate": 0.024797325660808882, + "loss": 0.2667, + "num_input_tokens_seen": 24843088, + "step": 32575 + }, + { + "epoch": 67.73388773388774, + "grad_norm": 7.178210944402963e-05, + "learning_rate": 0.02476489473382401, + "loss": 0.2661, + "num_input_tokens_seen": 24846800, + "step": 32580 + }, + { + "epoch": 67.74428274428274, + "grad_norm": 0.0004469225532375276, + "learning_rate": 0.024732483119666127, + "loss": 0.2699, + "num_input_tokens_seen": 24850736, + "step": 32585 + }, + { + "epoch": 67.75467775467776, + "grad_norm": 0.00019983664969913661, + "learning_rate": 0.024700090823333548, + "loss": 0.2463, + "num_input_tokens_seen": 24854480, + "step": 32590 + }, + { + "epoch": 67.76507276507276, + "grad_norm": 0.00039125647163018584, + "learning_rate": 0.02466771784982163, + "loss": 0.282, + "num_input_tokens_seen": 24858288, + "step": 32595 + }, + { + "epoch": 67.77546777546777, + "grad_norm": 0.00014569636550731957, + "learning_rate": 0.024635364204122594, + "loss": 0.2678, + "num_input_tokens_seen": 24862224, + "step": 32600 + }, + { + "epoch": 67.77546777546777, + "eval_loss": 0.24993912875652313, + "eval_runtime": 13.3855, + "eval_samples_per_second": 63.95, + "eval_steps_per_second": 15.987, + "num_input_tokens_seen": 24862224, + "step": 32600 + }, + { + "epoch": 67.78586278586279, + "grad_norm": 0.0004920413484796882, + "learning_rate": 0.024603029891225852, + "loss": 0.2744, + "num_input_tokens_seen": 24866160, + "step": 32605 + }, + { + "epoch": 67.79625779625779, + "grad_norm": 0.0010251975618302822, + "learning_rate": 0.024570714916117748, + "loss": 0.2826, + "num_input_tokens_seen": 24869968, + "step": 32610 + }, + { + "epoch": 67.8066528066528, + "grad_norm": 0.0005064397118985653, + "learning_rate": 0.024538419283781625, + "loss": 0.2782, + "num_input_tokens_seen": 24873712, + "step": 32615 + }, + { + "epoch": 67.81704781704782, + "grad_norm": 0.00044494454050436616, + "learning_rate": 0.024506142999197938, + "loss": 0.2618, + "num_input_tokens_seen": 24877712, + "step": 32620 + }, + { + "epoch": 67.82744282744282, + "grad_norm": 0.00012076876737410203, + "learning_rate": 0.024473886067344002, + "loss": 0.2752, + "num_input_tokens_seen": 24881360, + "step": 32625 + }, + { + "epoch": 67.83783783783784, + "grad_norm": 0.0001230451016454026, + "learning_rate": 0.02444164849319434, + "loss": 0.2518, + "num_input_tokens_seen": 24885296, + "step": 32630 + }, + { + "epoch": 67.84823284823285, + "grad_norm": 0.00011341610661474988, + "learning_rate": 0.024409430281720306, + "loss": 0.2569, + "num_input_tokens_seen": 24889072, + "step": 32635 + }, + { + "epoch": 67.85862785862786, + "grad_norm": 0.0005860665114596486, + "learning_rate": 0.024377231437890428, + "loss": 0.2652, + "num_input_tokens_seen": 24892944, + "step": 32640 + }, + { + "epoch": 67.86902286902287, + "grad_norm": 0.0006074450793676078, + "learning_rate": 0.024345051966670115, + "loss": 0.2492, + "num_input_tokens_seen": 24896624, + "step": 32645 + }, + { + "epoch": 67.87941787941789, + "grad_norm": 0.00017936533549800515, + "learning_rate": 0.024312891873021884, + "loss": 0.2419, + "num_input_tokens_seen": 24900432, + "step": 32650 + }, + { + "epoch": 67.88981288981289, + "grad_norm": 0.00037500725011341274, + "learning_rate": 0.024280751161905183, + "loss": 0.2381, + "num_input_tokens_seen": 24904272, + "step": 32655 + }, + { + "epoch": 67.9002079002079, + "grad_norm": 0.00024127135111484677, + "learning_rate": 0.02424862983827658, + "loss": 0.2435, + "num_input_tokens_seen": 24908048, + "step": 32660 + }, + { + "epoch": 67.9106029106029, + "grad_norm": 0.00033339319634251297, + "learning_rate": 0.024216527907089495, + "loss": 0.256, + "num_input_tokens_seen": 24911920, + "step": 32665 + }, + { + "epoch": 67.92099792099792, + "grad_norm": 0.001408817945048213, + "learning_rate": 0.024184445373294505, + "loss": 0.2452, + "num_input_tokens_seen": 24915664, + "step": 32670 + }, + { + "epoch": 67.93139293139293, + "grad_norm": 0.0003269794979132712, + "learning_rate": 0.02415238224183918, + "loss": 0.2764, + "num_input_tokens_seen": 24919440, + "step": 32675 + }, + { + "epoch": 67.94178794178794, + "grad_norm": 0.0001744201872497797, + "learning_rate": 0.024120338517667973, + "loss": 0.2621, + "num_input_tokens_seen": 24923376, + "step": 32680 + }, + { + "epoch": 67.95218295218295, + "grad_norm": 0.0002957224496640265, + "learning_rate": 0.02408831420572247, + "loss": 0.2896, + "num_input_tokens_seen": 24927024, + "step": 32685 + }, + { + "epoch": 67.96257796257797, + "grad_norm": 0.000302391272271052, + "learning_rate": 0.024056309310941264, + "loss": 0.2399, + "num_input_tokens_seen": 24930832, + "step": 32690 + }, + { + "epoch": 67.97297297297297, + "grad_norm": 8.811803127173334e-05, + "learning_rate": 0.02402432383825982, + "loss": 0.2718, + "num_input_tokens_seen": 24934672, + "step": 32695 + }, + { + "epoch": 67.98336798336798, + "grad_norm": 0.00011341377103235573, + "learning_rate": 0.023992357792610792, + "loss": 0.2405, + "num_input_tokens_seen": 24938480, + "step": 32700 + }, + { + "epoch": 67.993762993763, + "grad_norm": 0.0002103781735058874, + "learning_rate": 0.0239604111789237, + "loss": 0.2615, + "num_input_tokens_seen": 24942256, + "step": 32705 + }, + { + "epoch": 68.004158004158, + "grad_norm": 0.00018646445823833346, + "learning_rate": 0.023928484002125095, + "loss": 0.2501, + "num_input_tokens_seen": 24946272, + "step": 32710 + }, + { + "epoch": 68.01455301455302, + "grad_norm": 0.00011237052240176126, + "learning_rate": 0.023896576267138595, + "loss": 0.245, + "num_input_tokens_seen": 24950080, + "step": 32715 + }, + { + "epoch": 68.02494802494803, + "grad_norm": 8.91788222361356e-05, + "learning_rate": 0.02386468797888471, + "loss": 0.2589, + "num_input_tokens_seen": 24953984, + "step": 32720 + }, + { + "epoch": 68.03534303534303, + "grad_norm": 0.00012821443669963628, + "learning_rate": 0.023832819142281057, + "loss": 0.2761, + "num_input_tokens_seen": 24957696, + "step": 32725 + }, + { + "epoch": 68.04573804573805, + "grad_norm": 0.00021301429660525173, + "learning_rate": 0.02380096976224225, + "loss": 0.2588, + "num_input_tokens_seen": 24961504, + "step": 32730 + }, + { + "epoch": 68.05613305613305, + "grad_norm": 6.909112562425435e-05, + "learning_rate": 0.023769139843679777, + "loss": 0.2511, + "num_input_tokens_seen": 24965248, + "step": 32735 + }, + { + "epoch": 68.06652806652806, + "grad_norm": 0.0005465694703161716, + "learning_rate": 0.023737329391502287, + "loss": 0.2597, + "num_input_tokens_seen": 24968960, + "step": 32740 + }, + { + "epoch": 68.07692307692308, + "grad_norm": 0.00013574378681369126, + "learning_rate": 0.023705538410615293, + "loss": 0.253, + "num_input_tokens_seen": 24972928, + "step": 32745 + }, + { + "epoch": 68.08731808731808, + "grad_norm": 0.00020517090160865337, + "learning_rate": 0.023673766905921396, + "loss": 0.2598, + "num_input_tokens_seen": 24976768, + "step": 32750 + }, + { + "epoch": 68.0977130977131, + "grad_norm": 0.0009704021504148841, + "learning_rate": 0.0236420148823202, + "loss": 0.3023, + "num_input_tokens_seen": 24980704, + "step": 32755 + }, + { + "epoch": 68.10810810810811, + "grad_norm": 0.00021930535149294883, + "learning_rate": 0.02361028234470816, + "loss": 0.269, + "num_input_tokens_seen": 24984704, + "step": 32760 + }, + { + "epoch": 68.11850311850311, + "grad_norm": 0.0001617156231077388, + "learning_rate": 0.023578569297978913, + "loss": 0.2598, + "num_input_tokens_seen": 24988384, + "step": 32765 + }, + { + "epoch": 68.12889812889813, + "grad_norm": 6.561173358932137e-05, + "learning_rate": 0.023546875747023025, + "loss": 0.2815, + "num_input_tokens_seen": 24992096, + "step": 32770 + }, + { + "epoch": 68.13929313929314, + "grad_norm": 0.00018323773110751063, + "learning_rate": 0.02351520169672801, + "loss": 0.258, + "num_input_tokens_seen": 24995936, + "step": 32775 + }, + { + "epoch": 68.14968814968815, + "grad_norm": 0.00039092014776542783, + "learning_rate": 0.023483547151978357, + "loss": 0.2766, + "num_input_tokens_seen": 24999744, + "step": 32780 + }, + { + "epoch": 68.16008316008316, + "grad_norm": 0.0001044395612552762, + "learning_rate": 0.023451912117655675, + "loss": 0.2503, + "num_input_tokens_seen": 25003584, + "step": 32785 + }, + { + "epoch": 68.17047817047818, + "grad_norm": 0.0005508205504156649, + "learning_rate": 0.023420296598638417, + "loss": 0.2441, + "num_input_tokens_seen": 25007424, + "step": 32790 + }, + { + "epoch": 68.18087318087318, + "grad_norm": 0.00016276107635349035, + "learning_rate": 0.023388700599802165, + "loss": 0.2743, + "num_input_tokens_seen": 25011328, + "step": 32795 + }, + { + "epoch": 68.1912681912682, + "grad_norm": 0.002063712803646922, + "learning_rate": 0.023357124126019334, + "loss": 0.2658, + "num_input_tokens_seen": 25015296, + "step": 32800 + }, + { + "epoch": 68.1912681912682, + "eval_loss": 0.2485274374485016, + "eval_runtime": 13.3989, + "eval_samples_per_second": 63.886, + "eval_steps_per_second": 15.972, + "num_input_tokens_seen": 25015296, + "step": 32800 + }, + { + "epoch": 68.20166320166321, + "grad_norm": 0.0008883222471922636, + "learning_rate": 0.02332556718215945, + "loss": 0.2653, + "num_input_tokens_seen": 25019072, + "step": 32805 + }, + { + "epoch": 68.21205821205821, + "grad_norm": 0.0003568447136785835, + "learning_rate": 0.023294029773089035, + "loss": 0.2608, + "num_input_tokens_seen": 25023104, + "step": 32810 + }, + { + "epoch": 68.22245322245323, + "grad_norm": 0.0004084985994268209, + "learning_rate": 0.023262511903671484, + "loss": 0.2631, + "num_input_tokens_seen": 25026912, + "step": 32815 + }, + { + "epoch": 68.23284823284823, + "grad_norm": 0.0003029360668733716, + "learning_rate": 0.023231013578767324, + "loss": 0.2559, + "num_input_tokens_seen": 25030528, + "step": 32820 + }, + { + "epoch": 68.24324324324324, + "grad_norm": 0.00018699109205044806, + "learning_rate": 0.0231995348032339, + "loss": 0.2592, + "num_input_tokens_seen": 25034304, + "step": 32825 + }, + { + "epoch": 68.25363825363826, + "grad_norm": 0.0002158731804229319, + "learning_rate": 0.023168075581925685, + "loss": 0.2458, + "num_input_tokens_seen": 25038176, + "step": 32830 + }, + { + "epoch": 68.26403326403326, + "grad_norm": 0.00014035528874956071, + "learning_rate": 0.023136635919694126, + "loss": 0.2699, + "num_input_tokens_seen": 25042208, + "step": 32835 + }, + { + "epoch": 68.27442827442827, + "grad_norm": 0.00022370957594830543, + "learning_rate": 0.02310521582138753, + "loss": 0.2888, + "num_input_tokens_seen": 25045984, + "step": 32840 + }, + { + "epoch": 68.28482328482329, + "grad_norm": 0.00040909109520725906, + "learning_rate": 0.023073815291851357, + "loss": 0.2568, + "num_input_tokens_seen": 25049856, + "step": 32845 + }, + { + "epoch": 68.29521829521829, + "grad_norm": 0.0003214692114852369, + "learning_rate": 0.02304243433592788, + "loss": 0.263, + "num_input_tokens_seen": 25053536, + "step": 32850 + }, + { + "epoch": 68.3056133056133, + "grad_norm": 0.00039924561860971153, + "learning_rate": 0.023011072958456513, + "loss": 0.2526, + "num_input_tokens_seen": 25057248, + "step": 32855 + }, + { + "epoch": 68.31600831600832, + "grad_norm": 0.00018264653044752777, + "learning_rate": 0.022979731164273536, + "loss": 0.2811, + "num_input_tokens_seen": 25060960, + "step": 32860 + }, + { + "epoch": 68.32640332640332, + "grad_norm": 0.000957558280788362, + "learning_rate": 0.022948408958212218, + "loss": 0.2623, + "num_input_tokens_seen": 25064800, + "step": 32865 + }, + { + "epoch": 68.33679833679834, + "grad_norm": 0.0003596000315155834, + "learning_rate": 0.022917106345102876, + "loss": 0.2737, + "num_input_tokens_seen": 25068672, + "step": 32870 + }, + { + "epoch": 68.34719334719335, + "grad_norm": 0.00018152670236304402, + "learning_rate": 0.022885823329772785, + "loss": 0.2812, + "num_input_tokens_seen": 25072512, + "step": 32875 + }, + { + "epoch": 68.35758835758836, + "grad_norm": 0.0002636392309796065, + "learning_rate": 0.02285455991704612, + "loss": 0.2594, + "num_input_tokens_seen": 25076320, + "step": 32880 + }, + { + "epoch": 68.36798336798337, + "grad_norm": 0.00023016483464743942, + "learning_rate": 0.022823316111744117, + "loss": 0.275, + "num_input_tokens_seen": 25080064, + "step": 32885 + }, + { + "epoch": 68.37837837837837, + "grad_norm": 0.0004171333566773683, + "learning_rate": 0.022792091918685014, + "loss": 0.2454, + "num_input_tokens_seen": 25083936, + "step": 32890 + }, + { + "epoch": 68.38877338877339, + "grad_norm": 0.0001815782452467829, + "learning_rate": 0.022760887342683906, + "loss": 0.26, + "num_input_tokens_seen": 25087776, + "step": 32895 + }, + { + "epoch": 68.3991683991684, + "grad_norm": 0.0001997713843593374, + "learning_rate": 0.022729702388552975, + "loss": 0.2495, + "num_input_tokens_seen": 25091520, + "step": 32900 + }, + { + "epoch": 68.4095634095634, + "grad_norm": 8.17597538116388e-05, + "learning_rate": 0.022698537061101292, + "loss": 0.2798, + "num_input_tokens_seen": 25095360, + "step": 32905 + }, + { + "epoch": 68.41995841995842, + "grad_norm": 0.0002609244838822633, + "learning_rate": 0.022667391365134962, + "loss": 0.2579, + "num_input_tokens_seen": 25099232, + "step": 32910 + }, + { + "epoch": 68.43035343035343, + "grad_norm": 0.0004053963057231158, + "learning_rate": 0.022636265305457065, + "loss": 0.267, + "num_input_tokens_seen": 25103104, + "step": 32915 + }, + { + "epoch": 68.44074844074844, + "grad_norm": 7.771192758809775e-05, + "learning_rate": 0.02260515888686764, + "loss": 0.2625, + "num_input_tokens_seen": 25106976, + "step": 32920 + }, + { + "epoch": 68.45114345114345, + "grad_norm": 0.0001673835067776963, + "learning_rate": 0.022574072114163596, + "loss": 0.2698, + "num_input_tokens_seen": 25110912, + "step": 32925 + }, + { + "epoch": 68.46153846153847, + "grad_norm": 0.00011636003182502463, + "learning_rate": 0.022543004992139005, + "loss": 0.2857, + "num_input_tokens_seen": 25114688, + "step": 32930 + }, + { + "epoch": 68.47193347193347, + "grad_norm": 0.0008483059937134385, + "learning_rate": 0.022511957525584745, + "loss": 0.2512, + "num_input_tokens_seen": 25118464, + "step": 32935 + }, + { + "epoch": 68.48232848232848, + "grad_norm": 5.464487549033947e-05, + "learning_rate": 0.022480929719288778, + "loss": 0.2619, + "num_input_tokens_seen": 25122368, + "step": 32940 + }, + { + "epoch": 68.4927234927235, + "grad_norm": 0.0001387945667374879, + "learning_rate": 0.02244992157803592, + "loss": 0.2483, + "num_input_tokens_seen": 25126144, + "step": 32945 + }, + { + "epoch": 68.5031185031185, + "grad_norm": 0.0006583930226042867, + "learning_rate": 0.022418933106608047, + "loss": 0.2554, + "num_input_tokens_seen": 25129920, + "step": 32950 + }, + { + "epoch": 68.51351351351352, + "grad_norm": 7.203652057796717e-05, + "learning_rate": 0.022387964309784018, + "loss": 0.2769, + "num_input_tokens_seen": 25133792, + "step": 32955 + }, + { + "epoch": 68.52390852390852, + "grad_norm": 0.0002533787046559155, + "learning_rate": 0.022357015192339517, + "loss": 0.291, + "num_input_tokens_seen": 25137568, + "step": 32960 + }, + { + "epoch": 68.53430353430353, + "grad_norm": 0.0006018451531417668, + "learning_rate": 0.02232608575904734, + "loss": 0.2641, + "num_input_tokens_seen": 25141344, + "step": 32965 + }, + { + "epoch": 68.54469854469855, + "grad_norm": 0.00037659157533198595, + "learning_rate": 0.022295176014677225, + "loss": 0.245, + "num_input_tokens_seen": 25145216, + "step": 32970 + }, + { + "epoch": 68.55509355509355, + "grad_norm": 0.00038553026388399303, + "learning_rate": 0.02226428596399577, + "loss": 0.2328, + "num_input_tokens_seen": 25148736, + "step": 32975 + }, + { + "epoch": 68.56548856548856, + "grad_norm": 0.00038493535248562694, + "learning_rate": 0.02223341561176669, + "loss": 0.2736, + "num_input_tokens_seen": 25152576, + "step": 32980 + }, + { + "epoch": 68.57588357588358, + "grad_norm": 0.00020661762391682714, + "learning_rate": 0.0222025649627505, + "loss": 0.2703, + "num_input_tokens_seen": 25156384, + "step": 32985 + }, + { + "epoch": 68.58627858627858, + "grad_norm": 0.00038398735341615975, + "learning_rate": 0.022171734021704814, + "loss": 0.2699, + "num_input_tokens_seen": 25160128, + "step": 32990 + }, + { + "epoch": 68.5966735966736, + "grad_norm": 0.0002705395163502544, + "learning_rate": 0.022140922793384116, + "loss": 0.2506, + "num_input_tokens_seen": 25163904, + "step": 32995 + }, + { + "epoch": 68.60706860706861, + "grad_norm": 0.00017358586774207652, + "learning_rate": 0.022110131282539934, + "loss": 0.2544, + "num_input_tokens_seen": 25167744, + "step": 33000 + }, + { + "epoch": 68.60706860706861, + "eval_loss": 0.24799275398254395, + "eval_runtime": 13.4, + "eval_samples_per_second": 63.881, + "eval_steps_per_second": 15.97, + "num_input_tokens_seen": 25167744, + "step": 33000 + }, + { + "epoch": 68.61746361746361, + "grad_norm": 0.0007404489442706108, + "learning_rate": 0.022079359493920675, + "loss": 0.2498, + "num_input_tokens_seen": 25171456, + "step": 33005 + }, + { + "epoch": 68.62785862785863, + "grad_norm": 0.00015253355377353728, + "learning_rate": 0.02204860743227169, + "loss": 0.2264, + "num_input_tokens_seen": 25175264, + "step": 33010 + }, + { + "epoch": 68.63825363825364, + "grad_norm": 0.0002659687597770244, + "learning_rate": 0.022017875102335365, + "loss": 0.2727, + "num_input_tokens_seen": 25179072, + "step": 33015 + }, + { + "epoch": 68.64864864864865, + "grad_norm": 0.0003651087172329426, + "learning_rate": 0.02198716250885108, + "loss": 0.2525, + "num_input_tokens_seen": 25182944, + "step": 33020 + }, + { + "epoch": 68.65904365904366, + "grad_norm": 0.0003434156533330679, + "learning_rate": 0.021956469656555, + "loss": 0.2726, + "num_input_tokens_seen": 25186816, + "step": 33025 + }, + { + "epoch": 68.66943866943868, + "grad_norm": 0.0005010980530641973, + "learning_rate": 0.0219257965501804, + "loss": 0.2387, + "num_input_tokens_seen": 25190560, + "step": 33030 + }, + { + "epoch": 68.67983367983368, + "grad_norm": 0.0002615599369164556, + "learning_rate": 0.021895143194457494, + "loss": 0.2414, + "num_input_tokens_seen": 25194432, + "step": 33035 + }, + { + "epoch": 68.6902286902287, + "grad_norm": 0.00036230238038115203, + "learning_rate": 0.021864509594113322, + "loss": 0.2827, + "num_input_tokens_seen": 25198304, + "step": 33040 + }, + { + "epoch": 68.7006237006237, + "grad_norm": 0.0011337435571476817, + "learning_rate": 0.02183389575387207, + "loss": 0.2729, + "num_input_tokens_seen": 25202080, + "step": 33045 + }, + { + "epoch": 68.71101871101871, + "grad_norm": 0.0002490543120075017, + "learning_rate": 0.021803301678454682, + "loss": 0.265, + "num_input_tokens_seen": 25205760, + "step": 33050 + }, + { + "epoch": 68.72141372141373, + "grad_norm": 0.00026567679014988244, + "learning_rate": 0.021772727372579213, + "loss": 0.2789, + "num_input_tokens_seen": 25209472, + "step": 33055 + }, + { + "epoch": 68.73180873180873, + "grad_norm": 0.00033336292835883796, + "learning_rate": 0.02174217284096061, + "loss": 0.2713, + "num_input_tokens_seen": 25213280, + "step": 33060 + }, + { + "epoch": 68.74220374220374, + "grad_norm": 0.0003149026888422668, + "learning_rate": 0.0217116380883107, + "loss": 0.2696, + "num_input_tokens_seen": 25217344, + "step": 33065 + }, + { + "epoch": 68.75259875259876, + "grad_norm": 0.0007255618693307042, + "learning_rate": 0.021681123119338425, + "loss": 0.2921, + "num_input_tokens_seen": 25221088, + "step": 33070 + }, + { + "epoch": 68.76299376299376, + "grad_norm": 0.00048081984277814627, + "learning_rate": 0.02165062793874951, + "loss": 0.2862, + "num_input_tokens_seen": 25224832, + "step": 33075 + }, + { + "epoch": 68.77338877338877, + "grad_norm": 0.00022424502822104841, + "learning_rate": 0.021620152551246666, + "loss": 0.2738, + "num_input_tokens_seen": 25228768, + "step": 33080 + }, + { + "epoch": 68.78378378378379, + "grad_norm": 0.00021955319971311837, + "learning_rate": 0.02158969696152967, + "loss": 0.2548, + "num_input_tokens_seen": 25232512, + "step": 33085 + }, + { + "epoch": 68.79417879417879, + "grad_norm": 0.00022503330546896905, + "learning_rate": 0.021559261174295057, + "loss": 0.2496, + "num_input_tokens_seen": 25236608, + "step": 33090 + }, + { + "epoch": 68.8045738045738, + "grad_norm": 9.160657646134496e-05, + "learning_rate": 0.02152884519423646, + "loss": 0.2683, + "num_input_tokens_seen": 25240480, + "step": 33095 + }, + { + "epoch": 68.81496881496882, + "grad_norm": 0.00015715994231868535, + "learning_rate": 0.021498449026044447, + "loss": 0.2686, + "num_input_tokens_seen": 25244448, + "step": 33100 + }, + { + "epoch": 68.82536382536382, + "grad_norm": 0.00019138022616971284, + "learning_rate": 0.021468072674406414, + "loss": 0.2599, + "num_input_tokens_seen": 25248288, + "step": 33105 + }, + { + "epoch": 68.83575883575884, + "grad_norm": 0.00013834668789058924, + "learning_rate": 0.021437716144006795, + "loss": 0.2378, + "num_input_tokens_seen": 25252064, + "step": 33110 + }, + { + "epoch": 68.84615384615384, + "grad_norm": 0.0006101626786403358, + "learning_rate": 0.021407379439527002, + "loss": 0.304, + "num_input_tokens_seen": 25255712, + "step": 33115 + }, + { + "epoch": 68.85654885654886, + "grad_norm": 0.000161875577759929, + "learning_rate": 0.021377062565645255, + "loss": 0.2737, + "num_input_tokens_seen": 25259552, + "step": 33120 + }, + { + "epoch": 68.86694386694387, + "grad_norm": 0.0004340157029218972, + "learning_rate": 0.02134676552703688, + "loss": 0.228, + "num_input_tokens_seen": 25263360, + "step": 33125 + }, + { + "epoch": 68.87733887733887, + "grad_norm": 0.0006379140540957451, + "learning_rate": 0.02131648832837398, + "loss": 0.235, + "num_input_tokens_seen": 25267168, + "step": 33130 + }, + { + "epoch": 68.88773388773389, + "grad_norm": 0.0013916142052039504, + "learning_rate": 0.02128623097432574, + "loss": 0.2503, + "num_input_tokens_seen": 25270976, + "step": 33135 + }, + { + "epoch": 68.8981288981289, + "grad_norm": 0.0006769343162886798, + "learning_rate": 0.021255993469558192, + "loss": 0.2817, + "num_input_tokens_seen": 25274848, + "step": 33140 + }, + { + "epoch": 68.9085239085239, + "grad_norm": 5.576851253863424e-05, + "learning_rate": 0.021225775818734364, + "loss": 0.2525, + "num_input_tokens_seen": 25278720, + "step": 33145 + }, + { + "epoch": 68.91891891891892, + "grad_norm": 0.00017070425383280963, + "learning_rate": 0.021195578026514166, + "loss": 0.2817, + "num_input_tokens_seen": 25282496, + "step": 33150 + }, + { + "epoch": 68.92931392931393, + "grad_norm": 0.0003589412954170257, + "learning_rate": 0.02116540009755452, + "loss": 0.2542, + "num_input_tokens_seen": 25286336, + "step": 33155 + }, + { + "epoch": 68.93970893970894, + "grad_norm": 0.001447805087082088, + "learning_rate": 0.021135242036509173, + "loss": 0.2799, + "num_input_tokens_seen": 25290208, + "step": 33160 + }, + { + "epoch": 68.95010395010395, + "grad_norm": 0.00031258424860425293, + "learning_rate": 0.021105103848028967, + "loss": 0.2651, + "num_input_tokens_seen": 25293888, + "step": 33165 + }, + { + "epoch": 68.96049896049897, + "grad_norm": 0.0004397298616822809, + "learning_rate": 0.021074985536761504, + "loss": 0.2648, + "num_input_tokens_seen": 25297632, + "step": 33170 + }, + { + "epoch": 68.97089397089397, + "grad_norm": 0.00019049931142944843, + "learning_rate": 0.021044887107351435, + "loss": 0.26, + "num_input_tokens_seen": 25301280, + "step": 33175 + }, + { + "epoch": 68.98128898128898, + "grad_norm": 0.00013810255040880293, + "learning_rate": 0.021014808564440362, + "loss": 0.2806, + "num_input_tokens_seen": 25305088, + "step": 33180 + }, + { + "epoch": 68.99168399168398, + "grad_norm": 0.00013612354814540595, + "learning_rate": 0.02098474991266671, + "loss": 0.2687, + "num_input_tokens_seen": 25309024, + "step": 33185 + }, + { + "epoch": 69.002079002079, + "grad_norm": 0.00013730025966651738, + "learning_rate": 0.02095471115666592, + "loss": 0.2594, + "num_input_tokens_seen": 25312984, + "step": 33190 + }, + { + "epoch": 69.01247401247402, + "grad_norm": 0.0003259215736761689, + "learning_rate": 0.020924692301070406, + "loss": 0.2717, + "num_input_tokens_seen": 25317016, + "step": 33195 + }, + { + "epoch": 69.02286902286902, + "grad_norm": 0.00033645742223598063, + "learning_rate": 0.020894693350509346, + "loss": 0.2748, + "num_input_tokens_seen": 25321016, + "step": 33200 + }, + { + "epoch": 69.02286902286902, + "eval_loss": 0.2525767683982849, + "eval_runtime": 13.3975, + "eval_samples_per_second": 63.893, + "eval_steps_per_second": 15.973, + "num_input_tokens_seen": 25321016, + "step": 33200 + }, + { + "epoch": 69.03326403326403, + "grad_norm": 0.0004571609606500715, + "learning_rate": 0.020864714309609057, + "loss": 0.2902, + "num_input_tokens_seen": 25324824, + "step": 33205 + }, + { + "epoch": 69.04365904365905, + "grad_norm": 0.00010474302689544857, + "learning_rate": 0.020834755182992604, + "loss": 0.2771, + "num_input_tokens_seen": 25328440, + "step": 33210 + }, + { + "epoch": 69.05405405405405, + "grad_norm": 0.0005784222157672048, + "learning_rate": 0.02080481597528011, + "loss": 0.2734, + "num_input_tokens_seen": 25332312, + "step": 33215 + }, + { + "epoch": 69.06444906444906, + "grad_norm": 7.020789053058252e-05, + "learning_rate": 0.020774896691088583, + "loss": 0.2653, + "num_input_tokens_seen": 25336088, + "step": 33220 + }, + { + "epoch": 69.07484407484408, + "grad_norm": 0.00028458278393372893, + "learning_rate": 0.020744997335031882, + "loss": 0.2752, + "num_input_tokens_seen": 25339992, + "step": 33225 + }, + { + "epoch": 69.08523908523908, + "grad_norm": 0.0006704650004394352, + "learning_rate": 0.02071511791172092, + "loss": 0.2509, + "num_input_tokens_seen": 25343832, + "step": 33230 + }, + { + "epoch": 69.0956340956341, + "grad_norm": 0.0001475746976211667, + "learning_rate": 0.02068525842576351, + "loss": 0.2581, + "num_input_tokens_seen": 25347768, + "step": 33235 + }, + { + "epoch": 69.10602910602911, + "grad_norm": 0.00018695763719733804, + "learning_rate": 0.020655418881764264, + "loss": 0.2527, + "num_input_tokens_seen": 25351320, + "step": 33240 + }, + { + "epoch": 69.11642411642411, + "grad_norm": 0.0001395938452333212, + "learning_rate": 0.020625599284324923, + "loss": 0.2865, + "num_input_tokens_seen": 25355128, + "step": 33245 + }, + { + "epoch": 69.12681912681913, + "grad_norm": 0.00024050565843936056, + "learning_rate": 0.02059579963804396, + "loss": 0.2845, + "num_input_tokens_seen": 25358840, + "step": 33250 + }, + { + "epoch": 69.13721413721414, + "grad_norm": 7.251149509102106e-05, + "learning_rate": 0.02056601994751688, + "loss": 0.2569, + "num_input_tokens_seen": 25362616, + "step": 33255 + }, + { + "epoch": 69.14760914760915, + "grad_norm": 0.00027005522861145437, + "learning_rate": 0.02053626021733614, + "loss": 0.2432, + "num_input_tokens_seen": 25366296, + "step": 33260 + }, + { + "epoch": 69.15800415800416, + "grad_norm": 0.00025118948542512953, + "learning_rate": 0.02050652045209097, + "loss": 0.2638, + "num_input_tokens_seen": 25370392, + "step": 33265 + }, + { + "epoch": 69.16839916839916, + "grad_norm": 0.00043222654494456947, + "learning_rate": 0.020476800656367672, + "loss": 0.2266, + "num_input_tokens_seen": 25374104, + "step": 33270 + }, + { + "epoch": 69.17879417879418, + "grad_norm": 0.00012028025957988575, + "learning_rate": 0.020447100834749425, + "loss": 0.2646, + "num_input_tokens_seen": 25377912, + "step": 33275 + }, + { + "epoch": 69.1891891891892, + "grad_norm": 0.00024721797672100365, + "learning_rate": 0.02041742099181627, + "loss": 0.2498, + "num_input_tokens_seen": 25381848, + "step": 33280 + }, + { + "epoch": 69.1995841995842, + "grad_norm": 0.0002243156632175669, + "learning_rate": 0.02038776113214526, + "loss": 0.2511, + "num_input_tokens_seen": 25385752, + "step": 33285 + }, + { + "epoch": 69.20997920997921, + "grad_norm": 6.635630415985361e-05, + "learning_rate": 0.0203581212603103, + "loss": 0.2377, + "num_input_tokens_seen": 25389592, + "step": 33290 + }, + { + "epoch": 69.22037422037423, + "grad_norm": 0.0002868381852749735, + "learning_rate": 0.02032850138088219, + "loss": 0.2882, + "num_input_tokens_seen": 25393432, + "step": 33295 + }, + { + "epoch": 69.23076923076923, + "grad_norm": 0.00026127431192435324, + "learning_rate": 0.020298901498428754, + "loss": 0.2504, + "num_input_tokens_seen": 25397208, + "step": 33300 + }, + { + "epoch": 69.24116424116424, + "grad_norm": 0.0002085700980387628, + "learning_rate": 0.020269321617514595, + "loss": 0.3077, + "num_input_tokens_seen": 25401016, + "step": 33305 + }, + { + "epoch": 69.25155925155926, + "grad_norm": 0.0002756815229076892, + "learning_rate": 0.020239761742701343, + "loss": 0.2556, + "num_input_tokens_seen": 25404856, + "step": 33310 + }, + { + "epoch": 69.26195426195426, + "grad_norm": 8.619811706012115e-05, + "learning_rate": 0.02021022187854754, + "loss": 0.26, + "num_input_tokens_seen": 25408664, + "step": 33315 + }, + { + "epoch": 69.27234927234927, + "grad_norm": 0.0001901060895761475, + "learning_rate": 0.020180702029608522, + "loss": 0.2458, + "num_input_tokens_seen": 25412504, + "step": 33320 + }, + { + "epoch": 69.28274428274429, + "grad_norm": 0.00027060817228630185, + "learning_rate": 0.020151202200436695, + "loss": 0.2511, + "num_input_tokens_seen": 25416216, + "step": 33325 + }, + { + "epoch": 69.29313929313929, + "grad_norm": 0.0004160564858466387, + "learning_rate": 0.020121722395581226, + "loss": 0.2525, + "num_input_tokens_seen": 25420024, + "step": 33330 + }, + { + "epoch": 69.3035343035343, + "grad_norm": 0.0006290625897236168, + "learning_rate": 0.020092262619588342, + "loss": 0.2671, + "num_input_tokens_seen": 25423832, + "step": 33335 + }, + { + "epoch": 69.31392931392931, + "grad_norm": 0.00018528304644860327, + "learning_rate": 0.02006282287700109, + "loss": 0.2787, + "num_input_tokens_seen": 25427704, + "step": 33340 + }, + { + "epoch": 69.32432432432432, + "grad_norm": 0.0003635364701040089, + "learning_rate": 0.020033403172359427, + "loss": 0.2449, + "num_input_tokens_seen": 25431448, + "step": 33345 + }, + { + "epoch": 69.33471933471934, + "grad_norm": 0.0002578072890173644, + "learning_rate": 0.020004003510200284, + "loss": 0.2354, + "num_input_tokens_seen": 25435448, + "step": 33350 + }, + { + "epoch": 69.34511434511434, + "grad_norm": 0.00034021472674794495, + "learning_rate": 0.019974623895057407, + "loss": 0.281, + "num_input_tokens_seen": 25439256, + "step": 33355 + }, + { + "epoch": 69.35550935550935, + "grad_norm": 9.66677616816014e-05, + "learning_rate": 0.019945264331461553, + "loss": 0.2592, + "num_input_tokens_seen": 25442968, + "step": 33360 + }, + { + "epoch": 69.36590436590437, + "grad_norm": 0.0004060154897160828, + "learning_rate": 0.019915924823940317, + "loss": 0.2505, + "num_input_tokens_seen": 25446776, + "step": 33365 + }, + { + "epoch": 69.37629937629937, + "grad_norm": 0.0005675734137184918, + "learning_rate": 0.01988660537701816, + "loss": 0.2504, + "num_input_tokens_seen": 25450648, + "step": 33370 + }, + { + "epoch": 69.38669438669439, + "grad_norm": 0.0004298902931623161, + "learning_rate": 0.01985730599521659, + "loss": 0.2795, + "num_input_tokens_seen": 25454392, + "step": 33375 + }, + { + "epoch": 69.3970893970894, + "grad_norm": 0.00045510605559684336, + "learning_rate": 0.019828026683053918, + "loss": 0.2634, + "num_input_tokens_seen": 25458104, + "step": 33380 + }, + { + "epoch": 69.4074844074844, + "grad_norm": 0.0005070194602012634, + "learning_rate": 0.01979876744504535, + "loss": 0.2405, + "num_input_tokens_seen": 25461784, + "step": 33385 + }, + { + "epoch": 69.41787941787942, + "grad_norm": 0.00026510213501751423, + "learning_rate": 0.019769528285703046, + "loss": 0.2761, + "num_input_tokens_seen": 25465816, + "step": 33390 + }, + { + "epoch": 69.42827442827443, + "grad_norm": 0.00040704559069126844, + "learning_rate": 0.019740309209536098, + "loss": 0.2681, + "num_input_tokens_seen": 25469528, + "step": 33395 + }, + { + "epoch": 69.43866943866944, + "grad_norm": 0.00013673979265149683, + "learning_rate": 0.019711110221050387, + "loss": 0.2713, + "num_input_tokens_seen": 25473368, + "step": 33400 + }, + { + "epoch": 69.43866943866944, + "eval_loss": 0.24969600141048431, + "eval_runtime": 13.4019, + "eval_samples_per_second": 63.872, + "eval_steps_per_second": 15.968, + "num_input_tokens_seen": 25473368, + "step": 33400 + }, + { + "epoch": 69.44906444906445, + "grad_norm": 0.00023007711570244282, + "learning_rate": 0.019681931324748825, + "loss": 0.2683, + "num_input_tokens_seen": 25477240, + "step": 33405 + }, + { + "epoch": 69.45945945945945, + "grad_norm": 0.00042848187149502337, + "learning_rate": 0.019652772525131094, + "loss": 0.2879, + "num_input_tokens_seen": 25481144, + "step": 33410 + }, + { + "epoch": 69.46985446985447, + "grad_norm": 0.00060894968919456, + "learning_rate": 0.019623633826693885, + "loss": 0.2396, + "num_input_tokens_seen": 25484952, + "step": 33415 + }, + { + "epoch": 69.48024948024948, + "grad_norm": 0.0009679089416749775, + "learning_rate": 0.019594515233930788, + "loss": 0.2809, + "num_input_tokens_seen": 25488728, + "step": 33420 + }, + { + "epoch": 69.49064449064448, + "grad_norm": 0.00020810756541322917, + "learning_rate": 0.019565416751332186, + "loss": 0.2558, + "num_input_tokens_seen": 25492440, + "step": 33425 + }, + { + "epoch": 69.5010395010395, + "grad_norm": 0.00040405269828625023, + "learning_rate": 0.019536338383385497, + "loss": 0.2656, + "num_input_tokens_seen": 25496056, + "step": 33430 + }, + { + "epoch": 69.51143451143452, + "grad_norm": 0.00021596832084469497, + "learning_rate": 0.019507280134574933, + "loss": 0.2666, + "num_input_tokens_seen": 25499992, + "step": 33435 + }, + { + "epoch": 69.52182952182952, + "grad_norm": 0.00027918900013901293, + "learning_rate": 0.019478242009381624, + "loss": 0.2536, + "num_input_tokens_seen": 25503832, + "step": 33440 + }, + { + "epoch": 69.53222453222453, + "grad_norm": 0.000865179521497339, + "learning_rate": 0.01944922401228367, + "loss": 0.2857, + "num_input_tokens_seen": 25507608, + "step": 33445 + }, + { + "epoch": 69.54261954261955, + "grad_norm": 6.656336336163804e-05, + "learning_rate": 0.01942022614775593, + "loss": 0.2781, + "num_input_tokens_seen": 25511416, + "step": 33450 + }, + { + "epoch": 69.55301455301455, + "grad_norm": 0.0001633148203836754, + "learning_rate": 0.01939124842027029, + "loss": 0.2693, + "num_input_tokens_seen": 25515224, + "step": 33455 + }, + { + "epoch": 69.56340956340956, + "grad_norm": 0.00016383688489440829, + "learning_rate": 0.01936229083429551, + "loss": 0.2539, + "num_input_tokens_seen": 25519032, + "step": 33460 + }, + { + "epoch": 69.57380457380458, + "grad_norm": 0.00025212758919224143, + "learning_rate": 0.019333353394297148, + "loss": 0.2592, + "num_input_tokens_seen": 25522936, + "step": 33465 + }, + { + "epoch": 69.58419958419958, + "grad_norm": 0.00016369522199966013, + "learning_rate": 0.019304436104737754, + "loss": 0.2567, + "num_input_tokens_seen": 25526552, + "step": 33470 + }, + { + "epoch": 69.5945945945946, + "grad_norm": 0.00014449565787799656, + "learning_rate": 0.019275538970076778, + "loss": 0.2551, + "num_input_tokens_seen": 25530488, + "step": 33475 + }, + { + "epoch": 69.60498960498961, + "grad_norm": 0.0004477107140701264, + "learning_rate": 0.019246661994770434, + "loss": 0.2634, + "num_input_tokens_seen": 25534232, + "step": 33480 + }, + { + "epoch": 69.61538461538461, + "grad_norm": 0.00016228918684646487, + "learning_rate": 0.019217805183271985, + "loss": 0.2501, + "num_input_tokens_seen": 25538136, + "step": 33485 + }, + { + "epoch": 69.62577962577963, + "grad_norm": 0.0004337795835454017, + "learning_rate": 0.019188968540031465, + "loss": 0.269, + "num_input_tokens_seen": 25542104, + "step": 33490 + }, + { + "epoch": 69.63617463617463, + "grad_norm": 9.822798165259883e-05, + "learning_rate": 0.019160152069495867, + "loss": 0.2519, + "num_input_tokens_seen": 25545880, + "step": 33495 + }, + { + "epoch": 69.64656964656965, + "grad_norm": 0.00025473610730841756, + "learning_rate": 0.019131355776109103, + "loss": 0.2733, + "num_input_tokens_seen": 25549752, + "step": 33500 + }, + { + "epoch": 69.65696465696466, + "grad_norm": 0.0001143800254794769, + "learning_rate": 0.019102579664311857, + "loss": 0.2718, + "num_input_tokens_seen": 25553656, + "step": 33505 + }, + { + "epoch": 69.66735966735966, + "grad_norm": 0.00013031599519308656, + "learning_rate": 0.019073823738541763, + "loss": 0.2897, + "num_input_tokens_seen": 25557368, + "step": 33510 + }, + { + "epoch": 69.67775467775468, + "grad_norm": 0.00010841305629583076, + "learning_rate": 0.0190450880032334, + "loss": 0.2618, + "num_input_tokens_seen": 25561176, + "step": 33515 + }, + { + "epoch": 69.6881496881497, + "grad_norm": 0.00027366101858206093, + "learning_rate": 0.019016372462818114, + "loss": 0.287, + "num_input_tokens_seen": 25565112, + "step": 33520 + }, + { + "epoch": 69.6985446985447, + "grad_norm": 0.000278150342637673, + "learning_rate": 0.018987677121724278, + "loss": 0.259, + "num_input_tokens_seen": 25568984, + "step": 33525 + }, + { + "epoch": 69.70893970893971, + "grad_norm": 0.0014317523455247283, + "learning_rate": 0.018959001984377, + "loss": 0.2523, + "num_input_tokens_seen": 25572856, + "step": 33530 + }, + { + "epoch": 69.71933471933473, + "grad_norm": 0.0002784867538139224, + "learning_rate": 0.018930347055198377, + "loss": 0.2573, + "num_input_tokens_seen": 25576728, + "step": 33535 + }, + { + "epoch": 69.72972972972973, + "grad_norm": 0.00046777664101682603, + "learning_rate": 0.01890171233860739, + "loss": 0.2743, + "num_input_tokens_seen": 25580664, + "step": 33540 + }, + { + "epoch": 69.74012474012474, + "grad_norm": 0.00012935900304000825, + "learning_rate": 0.018873097839019807, + "loss": 0.2681, + "num_input_tokens_seen": 25584280, + "step": 33545 + }, + { + "epoch": 69.75051975051976, + "grad_norm": 0.00030318216886371374, + "learning_rate": 0.0188445035608484, + "loss": 0.2549, + "num_input_tokens_seen": 25588344, + "step": 33550 + }, + { + "epoch": 69.76091476091476, + "grad_norm": 0.0007277113618329167, + "learning_rate": 0.018815929508502777, + "loss": 0.23, + "num_input_tokens_seen": 25592088, + "step": 33555 + }, + { + "epoch": 69.77130977130977, + "grad_norm": 0.0006297500804066658, + "learning_rate": 0.01878737568638934, + "loss": 0.2896, + "num_input_tokens_seen": 25595992, + "step": 33560 + }, + { + "epoch": 69.78170478170478, + "grad_norm": 0.00021845132869202644, + "learning_rate": 0.01875884209891152, + "loss": 0.2534, + "num_input_tokens_seen": 25599864, + "step": 33565 + }, + { + "epoch": 69.79209979209979, + "grad_norm": 0.0002170616644434631, + "learning_rate": 0.018730328750469514, + "loss": 0.2761, + "num_input_tokens_seen": 25603704, + "step": 33570 + }, + { + "epoch": 69.8024948024948, + "grad_norm": 0.00029275112319737673, + "learning_rate": 0.018701835645460473, + "loss": 0.2182, + "num_input_tokens_seen": 25607480, + "step": 33575 + }, + { + "epoch": 69.81288981288981, + "grad_norm": 0.00010994344484061003, + "learning_rate": 0.01867336278827838, + "loss": 0.2717, + "num_input_tokens_seen": 25611288, + "step": 33580 + }, + { + "epoch": 69.82328482328482, + "grad_norm": 0.0003190810966771096, + "learning_rate": 0.018644910183314056, + "loss": 0.2645, + "num_input_tokens_seen": 25615032, + "step": 33585 + }, + { + "epoch": 69.83367983367984, + "grad_norm": 0.00012765720020979643, + "learning_rate": 0.01861647783495531, + "loss": 0.2697, + "num_input_tokens_seen": 25618904, + "step": 33590 + }, + { + "epoch": 69.84407484407484, + "grad_norm": 0.00042300455970689654, + "learning_rate": 0.01858806574758676, + "loss": 0.2662, + "num_input_tokens_seen": 25622776, + "step": 33595 + }, + { + "epoch": 69.85446985446985, + "grad_norm": 9.982017218135297e-05, + "learning_rate": 0.01855967392558988, + "loss": 0.2548, + "num_input_tokens_seen": 25626520, + "step": 33600 + }, + { + "epoch": 69.85446985446985, + "eval_loss": 0.24909155070781708, + "eval_runtime": 13.4131, + "eval_samples_per_second": 63.818, + "eval_steps_per_second": 15.955, + "num_input_tokens_seen": 25626520, + "step": 33600 + }, + { + "epoch": 69.86486486486487, + "grad_norm": 0.00029304978670552373, + "learning_rate": 0.018531302373343096, + "loss": 0.2869, + "num_input_tokens_seen": 25630200, + "step": 33605 + }, + { + "epoch": 69.87525987525987, + "grad_norm": 0.0001853680150816217, + "learning_rate": 0.018502951095221588, + "loss": 0.2572, + "num_input_tokens_seen": 25633784, + "step": 33610 + }, + { + "epoch": 69.88565488565489, + "grad_norm": 0.00012747083383146673, + "learning_rate": 0.01847462009559751, + "loss": 0.2685, + "num_input_tokens_seen": 25637592, + "step": 33615 + }, + { + "epoch": 69.8960498960499, + "grad_norm": 0.00031032529659569263, + "learning_rate": 0.01844630937883992, + "loss": 0.2615, + "num_input_tokens_seen": 25641592, + "step": 33620 + }, + { + "epoch": 69.9064449064449, + "grad_norm": 7.44626740925014e-05, + "learning_rate": 0.018418018949314573, + "loss": 0.2613, + "num_input_tokens_seen": 25645272, + "step": 33625 + }, + { + "epoch": 69.91683991683992, + "grad_norm": 9.610377310309559e-05, + "learning_rate": 0.018389748811384315, + "loss": 0.2417, + "num_input_tokens_seen": 25649048, + "step": 33630 + }, + { + "epoch": 69.92723492723492, + "grad_norm": 0.00020023660908918828, + "learning_rate": 0.018361498969408658, + "loss": 0.2639, + "num_input_tokens_seen": 25652792, + "step": 33635 + }, + { + "epoch": 69.93762993762994, + "grad_norm": 6.828092591604218e-05, + "learning_rate": 0.01833326942774415, + "loss": 0.2472, + "num_input_tokens_seen": 25656632, + "step": 33640 + }, + { + "epoch": 69.94802494802495, + "grad_norm": 0.00037896123831160367, + "learning_rate": 0.018305060190744155, + "loss": 0.2611, + "num_input_tokens_seen": 25660376, + "step": 33645 + }, + { + "epoch": 69.95841995841995, + "grad_norm": 0.00042398928781040013, + "learning_rate": 0.018276871262758846, + "loss": 0.2925, + "num_input_tokens_seen": 25664216, + "step": 33650 + }, + { + "epoch": 69.96881496881497, + "grad_norm": 0.00022856802388560027, + "learning_rate": 0.0182487026481353, + "loss": 0.2711, + "num_input_tokens_seen": 25668152, + "step": 33655 + }, + { + "epoch": 69.97920997920998, + "grad_norm": 7.218356768134981e-05, + "learning_rate": 0.018220554351217538, + "loss": 0.2519, + "num_input_tokens_seen": 25671896, + "step": 33660 + }, + { + "epoch": 69.98960498960498, + "grad_norm": 0.00014351400022860616, + "learning_rate": 0.01819242637634629, + "loss": 0.272, + "num_input_tokens_seen": 25675672, + "step": 33665 + }, + { + "epoch": 70.0, + "grad_norm": 0.00023627256450708956, + "learning_rate": 0.01816431872785933, + "loss": 0.2559, + "num_input_tokens_seen": 25679336, + "step": 33670 + }, + { + "epoch": 70.01039501039502, + "grad_norm": 0.0003869260544888675, + "learning_rate": 0.018136231410091148, + "loss": 0.2525, + "num_input_tokens_seen": 25682984, + "step": 33675 + }, + { + "epoch": 70.02079002079002, + "grad_norm": 0.00021054453100077808, + "learning_rate": 0.018108164427373175, + "loss": 0.2761, + "num_input_tokens_seen": 25686824, + "step": 33680 + }, + { + "epoch": 70.03118503118503, + "grad_norm": 0.00043784594163298607, + "learning_rate": 0.01808011778403375, + "loss": 0.2821, + "num_input_tokens_seen": 25690760, + "step": 33685 + }, + { + "epoch": 70.04158004158005, + "grad_norm": 0.002799914451315999, + "learning_rate": 0.01805209148439793, + "loss": 0.2769, + "num_input_tokens_seen": 25694536, + "step": 33690 + }, + { + "epoch": 70.05197505197505, + "grad_norm": 0.0004847442323807627, + "learning_rate": 0.018024085532787757, + "loss": 0.2585, + "num_input_tokens_seen": 25698248, + "step": 33695 + }, + { + "epoch": 70.06237006237006, + "grad_norm": 0.00048818832146935165, + "learning_rate": 0.017996099933522164, + "loss": 0.2619, + "num_input_tokens_seen": 25701992, + "step": 33700 + }, + { + "epoch": 70.07276507276508, + "grad_norm": 0.00061624008230865, + "learning_rate": 0.017968134690916775, + "loss": 0.2523, + "num_input_tokens_seen": 25705896, + "step": 33705 + }, + { + "epoch": 70.08316008316008, + "grad_norm": 9.043209865922108e-05, + "learning_rate": 0.017940189809284263, + "loss": 0.2424, + "num_input_tokens_seen": 25709512, + "step": 33710 + }, + { + "epoch": 70.0935550935551, + "grad_norm": 0.0007165076676756144, + "learning_rate": 0.017912265292934024, + "loss": 0.2675, + "num_input_tokens_seen": 25713512, + "step": 33715 + }, + { + "epoch": 70.1039501039501, + "grad_norm": 0.00010592116450425237, + "learning_rate": 0.017884361146172423, + "loss": 0.2537, + "num_input_tokens_seen": 25717256, + "step": 33720 + }, + { + "epoch": 70.11434511434511, + "grad_norm": 0.0005999957793392241, + "learning_rate": 0.01785647737330261, + "loss": 0.2705, + "num_input_tokens_seen": 25721256, + "step": 33725 + }, + { + "epoch": 70.12474012474013, + "grad_norm": 0.00021102858590893447, + "learning_rate": 0.017828613978624563, + "loss": 0.2554, + "num_input_tokens_seen": 25724936, + "step": 33730 + }, + { + "epoch": 70.13513513513513, + "grad_norm": 0.00010630543692968786, + "learning_rate": 0.01780077096643523, + "loss": 0.2591, + "num_input_tokens_seen": 25728680, + "step": 33735 + }, + { + "epoch": 70.14553014553015, + "grad_norm": 0.0003579895419534296, + "learning_rate": 0.017772948341028345, + "loss": 0.2687, + "num_input_tokens_seen": 25732360, + "step": 33740 + }, + { + "epoch": 70.15592515592516, + "grad_norm": 0.00012928416253998876, + "learning_rate": 0.01774514610669447, + "loss": 0.2752, + "num_input_tokens_seen": 25736040, + "step": 33745 + }, + { + "epoch": 70.16632016632016, + "grad_norm": 0.00032673278474248946, + "learning_rate": 0.017717364267721112, + "loss": 0.2497, + "num_input_tokens_seen": 25739848, + "step": 33750 + }, + { + "epoch": 70.17671517671518, + "grad_norm": 0.00011024600826203823, + "learning_rate": 0.017689602828392513, + "loss": 0.2418, + "num_input_tokens_seen": 25743688, + "step": 33755 + }, + { + "epoch": 70.18711018711019, + "grad_norm": 0.0003032666281796992, + "learning_rate": 0.017661861792989897, + "loss": 0.2616, + "num_input_tokens_seen": 25747688, + "step": 33760 + }, + { + "epoch": 70.1975051975052, + "grad_norm": 0.0001534395560156554, + "learning_rate": 0.017634141165791272, + "loss": 0.2664, + "num_input_tokens_seen": 25751496, + "step": 33765 + }, + { + "epoch": 70.20790020790021, + "grad_norm": 0.0003258484648540616, + "learning_rate": 0.017606440951071455, + "loss": 0.2597, + "num_input_tokens_seen": 25755368, + "step": 33770 + }, + { + "epoch": 70.21829521829522, + "grad_norm": 0.00011250966781517491, + "learning_rate": 0.017578761153102213, + "loss": 0.2546, + "num_input_tokens_seen": 25759080, + "step": 33775 + }, + { + "epoch": 70.22869022869023, + "grad_norm": 0.0002449020103085786, + "learning_rate": 0.017551101776152146, + "loss": 0.2516, + "num_input_tokens_seen": 25763080, + "step": 33780 + }, + { + "epoch": 70.23908523908524, + "grad_norm": 0.00046255573397502303, + "learning_rate": 0.017523462824486608, + "loss": 0.2833, + "num_input_tokens_seen": 25766920, + "step": 33785 + }, + { + "epoch": 70.24948024948024, + "grad_norm": 0.0005931609193794429, + "learning_rate": 0.01749584430236794, + "loss": 0.2769, + "num_input_tokens_seen": 25770792, + "step": 33790 + }, + { + "epoch": 70.25987525987526, + "grad_norm": 0.0002591744705568999, + "learning_rate": 0.01746824621405524, + "loss": 0.2603, + "num_input_tokens_seen": 25774632, + "step": 33795 + }, + { + "epoch": 70.27027027027027, + "grad_norm": 0.00041593145579099655, + "learning_rate": 0.017440668563804412, + "loss": 0.2589, + "num_input_tokens_seen": 25778248, + "step": 33800 + }, + { + "epoch": 70.27027027027027, + "eval_loss": 0.24918237328529358, + "eval_runtime": 13.4135, + "eval_samples_per_second": 63.816, + "eval_steps_per_second": 15.954, + "num_input_tokens_seen": 25778248, + "step": 33800 + }, + { + "epoch": 70.28066528066527, + "grad_norm": 0.00015579836326651275, + "learning_rate": 0.017413111355868392, + "loss": 0.2559, + "num_input_tokens_seen": 25781896, + "step": 33805 + }, + { + "epoch": 70.29106029106029, + "grad_norm": 0.0006690014852210879, + "learning_rate": 0.017385574594496748, + "loss": 0.2567, + "num_input_tokens_seen": 25785704, + "step": 33810 + }, + { + "epoch": 70.3014553014553, + "grad_norm": 0.0003354672808200121, + "learning_rate": 0.01735805828393605, + "loss": 0.2599, + "num_input_tokens_seen": 25789704, + "step": 33815 + }, + { + "epoch": 70.3118503118503, + "grad_norm": 0.00013841109466739, + "learning_rate": 0.017330562428429667, + "loss": 0.262, + "num_input_tokens_seen": 25793736, + "step": 33820 + }, + { + "epoch": 70.32224532224532, + "grad_norm": 0.0002482893178239465, + "learning_rate": 0.01730308703221776, + "loss": 0.2915, + "num_input_tokens_seen": 25797640, + "step": 33825 + }, + { + "epoch": 70.33264033264034, + "grad_norm": 0.00022579891083296388, + "learning_rate": 0.01727563209953744, + "loss": 0.2346, + "num_input_tokens_seen": 25801384, + "step": 33830 + }, + { + "epoch": 70.34303534303534, + "grad_norm": 0.0003142142086289823, + "learning_rate": 0.017248197634622535, + "loss": 0.2554, + "num_input_tokens_seen": 25805160, + "step": 33835 + }, + { + "epoch": 70.35343035343035, + "grad_norm": 0.00010582814138615504, + "learning_rate": 0.01722078364170383, + "loss": 0.2498, + "num_input_tokens_seen": 25809032, + "step": 33840 + }, + { + "epoch": 70.36382536382537, + "grad_norm": 0.0008987306500785053, + "learning_rate": 0.017193390125008905, + "loss": 0.261, + "num_input_tokens_seen": 25812936, + "step": 33845 + }, + { + "epoch": 70.37422037422037, + "grad_norm": 0.00034803838934749365, + "learning_rate": 0.017166017088762153, + "loss": 0.2736, + "num_input_tokens_seen": 25816744, + "step": 33850 + }, + { + "epoch": 70.38461538461539, + "grad_norm": 0.0008792418520897627, + "learning_rate": 0.017138664537184878, + "loss": 0.241, + "num_input_tokens_seen": 25820616, + "step": 33855 + }, + { + "epoch": 70.39501039501039, + "grad_norm": 0.00012005298776784912, + "learning_rate": 0.017111332474495172, + "loss": 0.2562, + "num_input_tokens_seen": 25824424, + "step": 33860 + }, + { + "epoch": 70.4054054054054, + "grad_norm": 0.00014786144311074167, + "learning_rate": 0.017084020904907998, + "loss": 0.2572, + "num_input_tokens_seen": 25828072, + "step": 33865 + }, + { + "epoch": 70.41580041580042, + "grad_norm": 0.00023796291498001665, + "learning_rate": 0.017056729832635103, + "loss": 0.2549, + "num_input_tokens_seen": 25831784, + "step": 33870 + }, + { + "epoch": 70.42619542619542, + "grad_norm": 0.0001081714071915485, + "learning_rate": 0.017029459261885153, + "loss": 0.2572, + "num_input_tokens_seen": 25835592, + "step": 33875 + }, + { + "epoch": 70.43659043659044, + "grad_norm": 0.00011292738054180518, + "learning_rate": 0.01700220919686359, + "loss": 0.2734, + "num_input_tokens_seen": 25839528, + "step": 33880 + }, + { + "epoch": 70.44698544698545, + "grad_norm": 0.0007623339770361781, + "learning_rate": 0.016974979641772723, + "loss": 0.2514, + "num_input_tokens_seen": 25843432, + "step": 33885 + }, + { + "epoch": 70.45738045738045, + "grad_norm": 0.001597981434315443, + "learning_rate": 0.01694777060081169, + "loss": 0.2723, + "num_input_tokens_seen": 25847240, + "step": 33890 + }, + { + "epoch": 70.46777546777547, + "grad_norm": 0.00038543209666386247, + "learning_rate": 0.016920582078176444, + "loss": 0.276, + "num_input_tokens_seen": 25851016, + "step": 33895 + }, + { + "epoch": 70.47817047817048, + "grad_norm": 0.0004299981810618192, + "learning_rate": 0.016893414078059863, + "loss": 0.2549, + "num_input_tokens_seen": 25854792, + "step": 33900 + }, + { + "epoch": 70.48856548856548, + "grad_norm": 0.0002463023702148348, + "learning_rate": 0.016866266604651535, + "loss": 0.2595, + "num_input_tokens_seen": 25858440, + "step": 33905 + }, + { + "epoch": 70.4989604989605, + "grad_norm": 0.0009466482442803681, + "learning_rate": 0.016839139662137976, + "loss": 0.2441, + "num_input_tokens_seen": 25862280, + "step": 33910 + }, + { + "epoch": 70.50935550935552, + "grad_norm": 0.00028445583302527666, + "learning_rate": 0.01681203325470245, + "loss": 0.2554, + "num_input_tokens_seen": 25865960, + "step": 33915 + }, + { + "epoch": 70.51975051975052, + "grad_norm": 0.0005626974161714315, + "learning_rate": 0.016784947386525157, + "loss": 0.2453, + "num_input_tokens_seen": 25869672, + "step": 33920 + }, + { + "epoch": 70.53014553014553, + "grad_norm": 0.0007579077500849962, + "learning_rate": 0.01675788206178308, + "loss": 0.2524, + "num_input_tokens_seen": 25873448, + "step": 33925 + }, + { + "epoch": 70.54054054054055, + "grad_norm": 9.533877164358273e-05, + "learning_rate": 0.016730837284649986, + "loss": 0.2904, + "num_input_tokens_seen": 25877544, + "step": 33930 + }, + { + "epoch": 70.55093555093555, + "grad_norm": 0.00034102812060154974, + "learning_rate": 0.016703813059296583, + "loss": 0.2602, + "num_input_tokens_seen": 25881480, + "step": 33935 + }, + { + "epoch": 70.56133056133056, + "grad_norm": 0.0007499619387090206, + "learning_rate": 0.016676809389890294, + "loss": 0.2399, + "num_input_tokens_seen": 25885288, + "step": 33940 + }, + { + "epoch": 70.57172557172557, + "grad_norm": 0.00011411232117097825, + "learning_rate": 0.016649826280595435, + "loss": 0.2574, + "num_input_tokens_seen": 25889256, + "step": 33945 + }, + { + "epoch": 70.58212058212058, + "grad_norm": 0.00017075546202249825, + "learning_rate": 0.016622863735573163, + "loss": 0.263, + "num_input_tokens_seen": 25893096, + "step": 33950 + }, + { + "epoch": 70.5925155925156, + "grad_norm": 0.0003233301977161318, + "learning_rate": 0.016595921758981395, + "loss": 0.2813, + "num_input_tokens_seen": 25896936, + "step": 33955 + }, + { + "epoch": 70.6029106029106, + "grad_norm": 0.00035117159131914377, + "learning_rate": 0.01656900035497495, + "loss": 0.2599, + "num_input_tokens_seen": 25900776, + "step": 33960 + }, + { + "epoch": 70.61330561330561, + "grad_norm": 0.0005780687788501382, + "learning_rate": 0.016542099527705485, + "loss": 0.2488, + "num_input_tokens_seen": 25904424, + "step": 33965 + }, + { + "epoch": 70.62370062370063, + "grad_norm": 0.00015171048289630562, + "learning_rate": 0.01651521928132138, + "loss": 0.2715, + "num_input_tokens_seen": 25908360, + "step": 33970 + }, + { + "epoch": 70.63409563409563, + "grad_norm": 0.0002794294850900769, + "learning_rate": 0.01648835961996794, + "loss": 0.2355, + "num_input_tokens_seen": 25912072, + "step": 33975 + }, + { + "epoch": 70.64449064449065, + "grad_norm": 0.00021177640883252025, + "learning_rate": 0.016461520547787285, + "loss": 0.2598, + "num_input_tokens_seen": 25915784, + "step": 33980 + }, + { + "epoch": 70.65488565488566, + "grad_norm": 0.0006138205644674599, + "learning_rate": 0.016434702068918266, + "loss": 0.2839, + "num_input_tokens_seen": 25919592, + "step": 33985 + }, + { + "epoch": 70.66528066528066, + "grad_norm": 0.0007318980642594397, + "learning_rate": 0.01640790418749673, + "loss": 0.2789, + "num_input_tokens_seen": 25923400, + "step": 33990 + }, + { + "epoch": 70.67567567567568, + "grad_norm": 4.5394597691483796e-05, + "learning_rate": 0.016381126907655134, + "loss": 0.2699, + "num_input_tokens_seen": 25927112, + "step": 33995 + }, + { + "epoch": 70.68607068607069, + "grad_norm": 0.00014760946214664727, + "learning_rate": 0.016354370233522948, + "loss": 0.2534, + "num_input_tokens_seen": 25930920, + "step": 34000 + }, + { + "epoch": 70.68607068607069, + "eval_loss": 0.25100159645080566, + "eval_runtime": 13.3894, + "eval_samples_per_second": 63.931, + "eval_steps_per_second": 15.983, + "num_input_tokens_seen": 25930920, + "step": 34000 + }, + { + "epoch": 70.6964656964657, + "grad_norm": 0.00033858089591376483, + "learning_rate": 0.016327634169226394, + "loss": 0.2876, + "num_input_tokens_seen": 25934856, + "step": 34005 + }, + { + "epoch": 70.70686070686071, + "grad_norm": 0.00034338614204898477, + "learning_rate": 0.016300918718888485, + "loss": 0.2696, + "num_input_tokens_seen": 25938504, + "step": 34010 + }, + { + "epoch": 70.71725571725571, + "grad_norm": 0.00013957686314824969, + "learning_rate": 0.016274223886629052, + "loss": 0.2706, + "num_input_tokens_seen": 25942216, + "step": 34015 + }, + { + "epoch": 70.72765072765073, + "grad_norm": 0.00036131704109720886, + "learning_rate": 0.01624754967656482, + "loss": 0.2809, + "num_input_tokens_seen": 25945864, + "step": 34020 + }, + { + "epoch": 70.73804573804574, + "grad_norm": 0.0002578433195594698, + "learning_rate": 0.016220896092809235, + "loss": 0.2777, + "num_input_tokens_seen": 25949864, + "step": 34025 + }, + { + "epoch": 70.74844074844074, + "grad_norm": 0.00037835861439816654, + "learning_rate": 0.01619426313947267, + "loss": 0.2586, + "num_input_tokens_seen": 25953832, + "step": 34030 + }, + { + "epoch": 70.75883575883576, + "grad_norm": 0.0002139493590220809, + "learning_rate": 0.016167650820662228, + "loss": 0.2615, + "num_input_tokens_seen": 25957576, + "step": 34035 + }, + { + "epoch": 70.76923076923077, + "grad_norm": 0.00018161052139475942, + "learning_rate": 0.016141059140481855, + "loss": 0.2618, + "num_input_tokens_seen": 25961384, + "step": 34040 + }, + { + "epoch": 70.77962577962577, + "grad_norm": 0.0018129326635971665, + "learning_rate": 0.016114488103032374, + "loss": 0.2947, + "num_input_tokens_seen": 25965224, + "step": 34045 + }, + { + "epoch": 70.79002079002079, + "grad_norm": 0.0005014881025999784, + "learning_rate": 0.016087937712411293, + "loss": 0.2835, + "num_input_tokens_seen": 25969032, + "step": 34050 + }, + { + "epoch": 70.8004158004158, + "grad_norm": 0.0007075099274516106, + "learning_rate": 0.01606140797271308, + "loss": 0.2711, + "num_input_tokens_seen": 25972712, + "step": 34055 + }, + { + "epoch": 70.8108108108108, + "grad_norm": 0.00033514745882712305, + "learning_rate": 0.01603489888802897, + "loss": 0.25, + "num_input_tokens_seen": 25976456, + "step": 34060 + }, + { + "epoch": 70.82120582120582, + "grad_norm": 0.00043173637823201716, + "learning_rate": 0.016008410462446918, + "loss": 0.2652, + "num_input_tokens_seen": 25980328, + "step": 34065 + }, + { + "epoch": 70.83160083160084, + "grad_norm": 0.0003789405745919794, + "learning_rate": 0.01598194270005185, + "loss": 0.2515, + "num_input_tokens_seen": 25984072, + "step": 34070 + }, + { + "epoch": 70.84199584199584, + "grad_norm": 0.00015172957500908524, + "learning_rate": 0.015955495604925356, + "loss": 0.2641, + "num_input_tokens_seen": 25987880, + "step": 34075 + }, + { + "epoch": 70.85239085239085, + "grad_norm": 0.00028241268591955304, + "learning_rate": 0.01592906918114598, + "loss": 0.265, + "num_input_tokens_seen": 25991656, + "step": 34080 + }, + { + "epoch": 70.86278586278586, + "grad_norm": 0.00026847838307730854, + "learning_rate": 0.015902663432788965, + "loss": 0.26, + "num_input_tokens_seen": 25995496, + "step": 34085 + }, + { + "epoch": 70.87318087318087, + "grad_norm": 0.0004988065920770168, + "learning_rate": 0.01587627836392643, + "loss": 0.2561, + "num_input_tokens_seen": 25999400, + "step": 34090 + }, + { + "epoch": 70.88357588357589, + "grad_norm": 0.0006295868661254644, + "learning_rate": 0.01584991397862726, + "loss": 0.2868, + "num_input_tokens_seen": 26003240, + "step": 34095 + }, + { + "epoch": 70.89397089397089, + "grad_norm": 0.0002003811241593212, + "learning_rate": 0.015823570280957214, + "loss": 0.2516, + "num_input_tokens_seen": 26007208, + "step": 34100 + }, + { + "epoch": 70.9043659043659, + "grad_norm": 0.0008161681471392512, + "learning_rate": 0.015797247274978766, + "loss": 0.2542, + "num_input_tokens_seen": 26011112, + "step": 34105 + }, + { + "epoch": 70.91476091476092, + "grad_norm": 0.00030335740302689373, + "learning_rate": 0.015770944964751326, + "loss": 0.2779, + "num_input_tokens_seen": 26015048, + "step": 34110 + }, + { + "epoch": 70.92515592515592, + "grad_norm": 0.00048233720008283854, + "learning_rate": 0.015744663354330956, + "loss": 0.2561, + "num_input_tokens_seen": 26018920, + "step": 34115 + }, + { + "epoch": 70.93555093555094, + "grad_norm": 0.00033505153260193765, + "learning_rate": 0.015718402447770664, + "loss": 0.2637, + "num_input_tokens_seen": 26022728, + "step": 34120 + }, + { + "epoch": 70.94594594594595, + "grad_norm": 0.000287884206045419, + "learning_rate": 0.015692162249120224, + "loss": 0.2479, + "num_input_tokens_seen": 26026504, + "step": 34125 + }, + { + "epoch": 70.95634095634095, + "grad_norm": 0.0008753057918511331, + "learning_rate": 0.01566594276242615, + "loss": 0.2744, + "num_input_tokens_seen": 26030280, + "step": 34130 + }, + { + "epoch": 70.96673596673597, + "grad_norm": 0.0003068975347559899, + "learning_rate": 0.015639743991731857, + "loss": 0.2639, + "num_input_tokens_seen": 26034088, + "step": 34135 + }, + { + "epoch": 70.97713097713098, + "grad_norm": 0.0009717169450595975, + "learning_rate": 0.01561356594107755, + "loss": 0.2511, + "num_input_tokens_seen": 26037704, + "step": 34140 + }, + { + "epoch": 70.98752598752598, + "grad_norm": 0.0003354522632434964, + "learning_rate": 0.015587408614500147, + "loss": 0.2676, + "num_input_tokens_seen": 26041384, + "step": 34145 + }, + { + "epoch": 70.997920997921, + "grad_norm": 0.0006731243338435888, + "learning_rate": 0.015561272016033505, + "loss": 0.281, + "num_input_tokens_seen": 26045256, + "step": 34150 + }, + { + "epoch": 71.00831600831602, + "grad_norm": 0.00039801179082132876, + "learning_rate": 0.015535156149708167, + "loss": 0.2796, + "num_input_tokens_seen": 26049024, + "step": 34155 + }, + { + "epoch": 71.01871101871102, + "grad_norm": 0.0002682908670976758, + "learning_rate": 0.015509061019551528, + "loss": 0.2726, + "num_input_tokens_seen": 26052832, + "step": 34160 + }, + { + "epoch": 71.02910602910603, + "grad_norm": 0.00048176926793530583, + "learning_rate": 0.015482986629587818, + "loss": 0.2483, + "num_input_tokens_seen": 26056640, + "step": 34165 + }, + { + "epoch": 71.03950103950103, + "grad_norm": 0.0019338406855240464, + "learning_rate": 0.01545693298383799, + "loss": 0.2396, + "num_input_tokens_seen": 26060448, + "step": 34170 + }, + { + "epoch": 71.04989604989605, + "grad_norm": 0.0019806839991360903, + "learning_rate": 0.015430900086319858, + "loss": 0.2481, + "num_input_tokens_seen": 26064160, + "step": 34175 + }, + { + "epoch": 71.06029106029106, + "grad_norm": 0.00238628638908267, + "learning_rate": 0.015404887941048084, + "loss": 0.283, + "num_input_tokens_seen": 26068032, + "step": 34180 + }, + { + "epoch": 71.07068607068607, + "grad_norm": 0.00018494039250072092, + "learning_rate": 0.01537889655203397, + "loss": 0.2604, + "num_input_tokens_seen": 26071840, + "step": 34185 + }, + { + "epoch": 71.08108108108108, + "grad_norm": 0.00013956715702079237, + "learning_rate": 0.015352925923285798, + "loss": 0.2637, + "num_input_tokens_seen": 26075680, + "step": 34190 + }, + { + "epoch": 71.0914760914761, + "grad_norm": 0.0005342590739019215, + "learning_rate": 0.015326976058808511, + "loss": 0.2686, + "num_input_tokens_seen": 26079584, + "step": 34195 + }, + { + "epoch": 71.1018711018711, + "grad_norm": 0.00026903097750619054, + "learning_rate": 0.015301046962603908, + "loss": 0.2532, + "num_input_tokens_seen": 26083456, + "step": 34200 + }, + { + "epoch": 71.1018711018711, + "eval_loss": 0.2491416186094284, + "eval_runtime": 13.4224, + "eval_samples_per_second": 63.774, + "eval_steps_per_second": 15.943, + "num_input_tokens_seen": 26083456, + "step": 34200 + }, + { + "epoch": 71.11226611226611, + "grad_norm": 0.00017995413509197533, + "learning_rate": 0.015275138638670626, + "loss": 0.2583, + "num_input_tokens_seen": 26087200, + "step": 34205 + }, + { + "epoch": 71.12266112266113, + "grad_norm": 0.0012385392328724265, + "learning_rate": 0.015249251091004001, + "loss": 0.2664, + "num_input_tokens_seen": 26090944, + "step": 34210 + }, + { + "epoch": 71.13305613305613, + "grad_norm": 0.00025588812422938645, + "learning_rate": 0.01522338432359624, + "loss": 0.2438, + "num_input_tokens_seen": 26094592, + "step": 34215 + }, + { + "epoch": 71.14345114345114, + "grad_norm": 4.990608795196749e-05, + "learning_rate": 0.01519753834043635, + "loss": 0.2677, + "num_input_tokens_seen": 26098400, + "step": 34220 + }, + { + "epoch": 71.15384615384616, + "grad_norm": 5.714313738280907e-05, + "learning_rate": 0.015171713145510095, + "loss": 0.2545, + "num_input_tokens_seen": 26102368, + "step": 34225 + }, + { + "epoch": 71.16424116424116, + "grad_norm": 0.0004953845636919141, + "learning_rate": 0.01514590874279999, + "loss": 0.265, + "num_input_tokens_seen": 26106080, + "step": 34230 + }, + { + "epoch": 71.17463617463618, + "grad_norm": 0.0009671784937381744, + "learning_rate": 0.015120125136285467, + "loss": 0.2871, + "num_input_tokens_seen": 26109792, + "step": 34235 + }, + { + "epoch": 71.18503118503118, + "grad_norm": 0.001206830726005137, + "learning_rate": 0.015094362329942629, + "loss": 0.2504, + "num_input_tokens_seen": 26113536, + "step": 34240 + }, + { + "epoch": 71.1954261954262, + "grad_norm": 0.0017787389224395156, + "learning_rate": 0.01506862032774448, + "loss": 0.275, + "num_input_tokens_seen": 26117312, + "step": 34245 + }, + { + "epoch": 71.20582120582121, + "grad_norm": 0.00016638028318993747, + "learning_rate": 0.015042899133660697, + "loss": 0.2516, + "num_input_tokens_seen": 26121184, + "step": 34250 + }, + { + "epoch": 71.21621621621621, + "grad_norm": 0.00012802726996596903, + "learning_rate": 0.01501719875165789, + "loss": 0.2856, + "num_input_tokens_seen": 26124864, + "step": 34255 + }, + { + "epoch": 71.22661122661123, + "grad_norm": 0.00014675762213300914, + "learning_rate": 0.014991519185699286, + "loss": 0.2743, + "num_input_tokens_seen": 26128640, + "step": 34260 + }, + { + "epoch": 71.23700623700624, + "grad_norm": 0.000537066429387778, + "learning_rate": 0.014965860439745054, + "loss": 0.2883, + "num_input_tokens_seen": 26132544, + "step": 34265 + }, + { + "epoch": 71.24740124740124, + "grad_norm": 0.0006858361884951591, + "learning_rate": 0.01494022251775211, + "loss": 0.2607, + "num_input_tokens_seen": 26136288, + "step": 34270 + }, + { + "epoch": 71.25779625779626, + "grad_norm": 0.000666318868752569, + "learning_rate": 0.014914605423674109, + "loss": 0.2555, + "num_input_tokens_seen": 26140160, + "step": 34275 + }, + { + "epoch": 71.26819126819127, + "grad_norm": 0.00018950880621559918, + "learning_rate": 0.014889009161461525, + "loss": 0.2612, + "num_input_tokens_seen": 26144064, + "step": 34280 + }, + { + "epoch": 71.27858627858627, + "grad_norm": 0.00047989084850996733, + "learning_rate": 0.014863433735061665, + "loss": 0.266, + "num_input_tokens_seen": 26147872, + "step": 34285 + }, + { + "epoch": 71.28898128898129, + "grad_norm": 0.00037910614628344774, + "learning_rate": 0.014837879148418541, + "loss": 0.2616, + "num_input_tokens_seen": 26151744, + "step": 34290 + }, + { + "epoch": 71.2993762993763, + "grad_norm": 0.00048052528291009367, + "learning_rate": 0.01481234540547302, + "loss": 0.2651, + "num_input_tokens_seen": 26155616, + "step": 34295 + }, + { + "epoch": 71.3097713097713, + "grad_norm": 0.0019880712497979403, + "learning_rate": 0.014786832510162717, + "loss": 0.2571, + "num_input_tokens_seen": 26159424, + "step": 34300 + }, + { + "epoch": 71.32016632016632, + "grad_norm": 0.00032300769817084074, + "learning_rate": 0.014761340466422017, + "loss": 0.2762, + "num_input_tokens_seen": 26163296, + "step": 34305 + }, + { + "epoch": 71.33056133056132, + "grad_norm": 0.0004088038404006511, + "learning_rate": 0.014735869278182144, + "loss": 0.2582, + "num_input_tokens_seen": 26167104, + "step": 34310 + }, + { + "epoch": 71.34095634095634, + "grad_norm": 0.00038851110730320215, + "learning_rate": 0.014710418949371057, + "loss": 0.2447, + "num_input_tokens_seen": 26170816, + "step": 34315 + }, + { + "epoch": 71.35135135135135, + "grad_norm": 0.0005060535622760653, + "learning_rate": 0.014684989483913495, + "loss": 0.2586, + "num_input_tokens_seen": 26174656, + "step": 34320 + }, + { + "epoch": 71.36174636174636, + "grad_norm": 0.0005553790833801031, + "learning_rate": 0.014659580885731077, + "loss": 0.2635, + "num_input_tokens_seen": 26178624, + "step": 34325 + }, + { + "epoch": 71.37214137214137, + "grad_norm": 0.0002516135573387146, + "learning_rate": 0.014634193158742047, + "loss": 0.2434, + "num_input_tokens_seen": 26182400, + "step": 34330 + }, + { + "epoch": 71.38253638253639, + "grad_norm": 0.00018362715491093695, + "learning_rate": 0.014608826306861576, + "loss": 0.2539, + "num_input_tokens_seen": 26186240, + "step": 34335 + }, + { + "epoch": 71.39293139293139, + "grad_norm": 0.0002084135339828208, + "learning_rate": 0.014583480334001486, + "loss": 0.2357, + "num_input_tokens_seen": 26190016, + "step": 34340 + }, + { + "epoch": 71.4033264033264, + "grad_norm": 0.0007845240761525929, + "learning_rate": 0.014558155244070496, + "loss": 0.2653, + "num_input_tokens_seen": 26193792, + "step": 34345 + }, + { + "epoch": 71.41372141372142, + "grad_norm": 0.00022965479001868516, + "learning_rate": 0.014532851040974036, + "loss": 0.2766, + "num_input_tokens_seen": 26197600, + "step": 34350 + }, + { + "epoch": 71.42411642411642, + "grad_norm": 0.00018276744231116027, + "learning_rate": 0.014507567728614335, + "loss": 0.2588, + "num_input_tokens_seen": 26201248, + "step": 34355 + }, + { + "epoch": 71.43451143451144, + "grad_norm": 0.00124224997125566, + "learning_rate": 0.01448230531089037, + "loss": 0.2618, + "num_input_tokens_seen": 26205120, + "step": 34360 + }, + { + "epoch": 71.44490644490645, + "grad_norm": 0.00011332490248605609, + "learning_rate": 0.014457063791697993, + "loss": 0.2733, + "num_input_tokens_seen": 26209024, + "step": 34365 + }, + { + "epoch": 71.45530145530145, + "grad_norm": 0.0004901313222944736, + "learning_rate": 0.01443184317492971, + "loss": 0.2671, + "num_input_tokens_seen": 26212960, + "step": 34370 + }, + { + "epoch": 71.46569646569647, + "grad_norm": 0.00019822089234367013, + "learning_rate": 0.014406643464474822, + "loss": 0.257, + "num_input_tokens_seen": 26216576, + "step": 34375 + }, + { + "epoch": 71.47609147609148, + "grad_norm": 0.0012744978303089738, + "learning_rate": 0.014381464664219539, + "loss": 0.2948, + "num_input_tokens_seen": 26220448, + "step": 34380 + }, + { + "epoch": 71.48648648648648, + "grad_norm": 0.0001673355873208493, + "learning_rate": 0.014356306778046656, + "loss": 0.2727, + "num_input_tokens_seen": 26224256, + "step": 34385 + }, + { + "epoch": 71.4968814968815, + "grad_norm": 0.00024221985950134695, + "learning_rate": 0.014331169809835885, + "loss": 0.2579, + "num_input_tokens_seen": 26228064, + "step": 34390 + }, + { + "epoch": 71.5072765072765, + "grad_norm": 0.000330470415065065, + "learning_rate": 0.014306053763463644, + "loss": 0.251, + "num_input_tokens_seen": 26231776, + "step": 34395 + }, + { + "epoch": 71.51767151767152, + "grad_norm": 0.000543308793567121, + "learning_rate": 0.014280958642803147, + "loss": 0.266, + "num_input_tokens_seen": 26235552, + "step": 34400 + }, + { + "epoch": 71.51767151767152, + "eval_loss": 0.2503880560398102, + "eval_runtime": 13.3923, + "eval_samples_per_second": 63.918, + "eval_steps_per_second": 15.979, + "num_input_tokens_seen": 26235552, + "step": 34400 + }, + { + "epoch": 71.52806652806653, + "grad_norm": 0.00027296121697872877, + "learning_rate": 0.014255884451724404, + "loss": 0.2518, + "num_input_tokens_seen": 26239424, + "step": 34405 + }, + { + "epoch": 71.53846153846153, + "grad_norm": 0.00045671811676584184, + "learning_rate": 0.014230831194094101, + "loss": 0.2564, + "num_input_tokens_seen": 26243200, + "step": 34410 + }, + { + "epoch": 71.54885654885655, + "grad_norm": 0.00017378877964802086, + "learning_rate": 0.014205798873775865, + "loss": 0.262, + "num_input_tokens_seen": 26247040, + "step": 34415 + }, + { + "epoch": 71.55925155925156, + "grad_norm": 9.670834697317332e-05, + "learning_rate": 0.014180787494629893, + "loss": 0.282, + "num_input_tokens_seen": 26250720, + "step": 34420 + }, + { + "epoch": 71.56964656964657, + "grad_norm": 7.944850949570537e-05, + "learning_rate": 0.014155797060513314, + "loss": 0.2491, + "num_input_tokens_seen": 26254432, + "step": 34425 + }, + { + "epoch": 71.58004158004158, + "grad_norm": 0.0001057119297911413, + "learning_rate": 0.014130827575279963, + "loss": 0.2513, + "num_input_tokens_seen": 26258304, + "step": 34430 + }, + { + "epoch": 71.5904365904366, + "grad_norm": 0.0004385255742818117, + "learning_rate": 0.014105879042780427, + "loss": 0.2339, + "num_input_tokens_seen": 26262144, + "step": 34435 + }, + { + "epoch": 71.6008316008316, + "grad_norm": 0.0001551207824377343, + "learning_rate": 0.014080951466862113, + "loss": 0.2548, + "num_input_tokens_seen": 26265888, + "step": 34440 + }, + { + "epoch": 71.61122661122661, + "grad_norm": 0.000586564710829407, + "learning_rate": 0.014056044851369126, + "loss": 0.2613, + "num_input_tokens_seen": 26269824, + "step": 34445 + }, + { + "epoch": 71.62162162162163, + "grad_norm": 0.0006920543382875621, + "learning_rate": 0.014031159200142428, + "loss": 0.2652, + "num_input_tokens_seen": 26273632, + "step": 34450 + }, + { + "epoch": 71.63201663201663, + "grad_norm": 0.00020421632507350296, + "learning_rate": 0.014006294517019667, + "loss": 0.2707, + "num_input_tokens_seen": 26277408, + "step": 34455 + }, + { + "epoch": 71.64241164241164, + "grad_norm": 0.0002219188609160483, + "learning_rate": 0.013981450805835276, + "loss": 0.2787, + "num_input_tokens_seen": 26281152, + "step": 34460 + }, + { + "epoch": 71.65280665280665, + "grad_norm": 9.26225766306743e-05, + "learning_rate": 0.01395662807042049, + "loss": 0.2348, + "num_input_tokens_seen": 26285088, + "step": 34465 + }, + { + "epoch": 71.66320166320166, + "grad_norm": 0.00014120752166491002, + "learning_rate": 0.013931826314603296, + "loss": 0.2689, + "num_input_tokens_seen": 26289024, + "step": 34470 + }, + { + "epoch": 71.67359667359668, + "grad_norm": 0.0002083756262436509, + "learning_rate": 0.013907045542208401, + "loss": 0.2576, + "num_input_tokens_seen": 26292896, + "step": 34475 + }, + { + "epoch": 71.68399168399168, + "grad_norm": 0.00022319567506201565, + "learning_rate": 0.013882285757057333, + "loss": 0.289, + "num_input_tokens_seen": 26296608, + "step": 34480 + }, + { + "epoch": 71.6943866943867, + "grad_norm": 0.00013077058247290552, + "learning_rate": 0.013857546962968403, + "loss": 0.2941, + "num_input_tokens_seen": 26300448, + "step": 34485 + }, + { + "epoch": 71.70478170478171, + "grad_norm": 0.0001975538325496018, + "learning_rate": 0.013832829163756577, + "loss": 0.2585, + "num_input_tokens_seen": 26304256, + "step": 34490 + }, + { + "epoch": 71.71517671517671, + "grad_norm": 0.00021075663971714675, + "learning_rate": 0.013808132363233689, + "loss": 0.2763, + "num_input_tokens_seen": 26308128, + "step": 34495 + }, + { + "epoch": 71.72557172557173, + "grad_norm": 0.0003411433717701584, + "learning_rate": 0.013783456565208256, + "loss": 0.2513, + "num_input_tokens_seen": 26312160, + "step": 34500 + }, + { + "epoch": 71.73596673596674, + "grad_norm": 0.0011928318999707699, + "learning_rate": 0.01375880177348564, + "loss": 0.2809, + "num_input_tokens_seen": 26315968, + "step": 34505 + }, + { + "epoch": 71.74636174636174, + "grad_norm": 0.0001411486155120656, + "learning_rate": 0.013734167991867928, + "loss": 0.2684, + "num_input_tokens_seen": 26319712, + "step": 34510 + }, + { + "epoch": 71.75675675675676, + "grad_norm": 0.0002316038589924574, + "learning_rate": 0.013709555224153935, + "loss": 0.2587, + "num_input_tokens_seen": 26323712, + "step": 34515 + }, + { + "epoch": 71.76715176715177, + "grad_norm": 0.00013614627823699266, + "learning_rate": 0.013684963474139222, + "loss": 0.2526, + "num_input_tokens_seen": 26327552, + "step": 34520 + }, + { + "epoch": 71.77754677754677, + "grad_norm": 0.00013767197378911078, + "learning_rate": 0.013660392745616224, + "loss": 0.2763, + "num_input_tokens_seen": 26331392, + "step": 34525 + }, + { + "epoch": 71.78794178794179, + "grad_norm": 0.00042263694922439754, + "learning_rate": 0.013635843042373974, + "loss": 0.2382, + "num_input_tokens_seen": 26335360, + "step": 34530 + }, + { + "epoch": 71.7983367983368, + "grad_norm": 0.00031298157409764826, + "learning_rate": 0.01361131436819843, + "loss": 0.2607, + "num_input_tokens_seen": 26339328, + "step": 34535 + }, + { + "epoch": 71.8087318087318, + "grad_norm": 0.00030167793738655746, + "learning_rate": 0.013586806726872147, + "loss": 0.268, + "num_input_tokens_seen": 26343200, + "step": 34540 + }, + { + "epoch": 71.81912681912682, + "grad_norm": 0.0003853478701785207, + "learning_rate": 0.013562320122174537, + "loss": 0.2548, + "num_input_tokens_seen": 26347040, + "step": 34545 + }, + { + "epoch": 71.82952182952182, + "grad_norm": 0.0003635018365457654, + "learning_rate": 0.013537854557881762, + "loss": 0.2707, + "num_input_tokens_seen": 26350912, + "step": 34550 + }, + { + "epoch": 71.83991683991684, + "grad_norm": 0.000908699061255902, + "learning_rate": 0.013513410037766687, + "loss": 0.2829, + "num_input_tokens_seen": 26354784, + "step": 34555 + }, + { + "epoch": 71.85031185031185, + "grad_norm": 0.00015832159260753542, + "learning_rate": 0.013488986565598998, + "loss": 0.2579, + "num_input_tokens_seen": 26358464, + "step": 34560 + }, + { + "epoch": 71.86070686070686, + "grad_norm": 0.0003910884843207896, + "learning_rate": 0.013464584145145097, + "loss": 0.285, + "num_input_tokens_seen": 26362208, + "step": 34565 + }, + { + "epoch": 71.87110187110187, + "grad_norm": 0.0005502876010723412, + "learning_rate": 0.013440202780168109, + "loss": 0.2534, + "num_input_tokens_seen": 26365888, + "step": 34570 + }, + { + "epoch": 71.88149688149689, + "grad_norm": 0.00015817715029697865, + "learning_rate": 0.01341584247442799, + "loss": 0.2693, + "num_input_tokens_seen": 26369696, + "step": 34575 + }, + { + "epoch": 71.89189189189189, + "grad_norm": 0.00044241794967092574, + "learning_rate": 0.013391503231681355, + "loss": 0.2641, + "num_input_tokens_seen": 26373472, + "step": 34580 + }, + { + "epoch": 71.9022869022869, + "grad_norm": 0.0001169347160612233, + "learning_rate": 0.013367185055681685, + "loss": 0.2939, + "num_input_tokens_seen": 26377248, + "step": 34585 + }, + { + "epoch": 71.91268191268192, + "grad_norm": 0.00021283843670971692, + "learning_rate": 0.013342887950179095, + "loss": 0.2864, + "num_input_tokens_seen": 26381088, + "step": 34590 + }, + { + "epoch": 71.92307692307692, + "grad_norm": 0.0006728462176397443, + "learning_rate": 0.013318611918920554, + "loss": 0.2641, + "num_input_tokens_seen": 26385088, + "step": 34595 + }, + { + "epoch": 71.93347193347194, + "grad_norm": 0.0004620038962457329, + "learning_rate": 0.01329435696564965, + "loss": 0.256, + "num_input_tokens_seen": 26388832, + "step": 34600 + }, + { + "epoch": 71.93347193347194, + "eval_loss": 0.25349533557891846, + "eval_runtime": 13.3973, + "eval_samples_per_second": 63.893, + "eval_steps_per_second": 15.973, + "num_input_tokens_seen": 26388832, + "step": 34600 + }, + { + "epoch": 71.94386694386695, + "grad_norm": 0.00020175377721898258, + "learning_rate": 0.013270123094106894, + "loss": 0.2508, + "num_input_tokens_seen": 26392544, + "step": 34605 + }, + { + "epoch": 71.95426195426195, + "grad_norm": 0.00018747425929177552, + "learning_rate": 0.013245910308029395, + "loss": 0.2795, + "num_input_tokens_seen": 26396320, + "step": 34610 + }, + { + "epoch": 71.96465696465697, + "grad_norm": 0.0013315289979800582, + "learning_rate": 0.0132217186111511, + "loss": 0.2578, + "num_input_tokens_seen": 26400064, + "step": 34615 + }, + { + "epoch": 71.97505197505197, + "grad_norm": 0.0004098436620552093, + "learning_rate": 0.013197548007202626, + "loss": 0.2609, + "num_input_tokens_seen": 26403840, + "step": 34620 + }, + { + "epoch": 71.98544698544698, + "grad_norm": 0.0003166873357258737, + "learning_rate": 0.01317339849991142, + "loss": 0.2604, + "num_input_tokens_seen": 26407680, + "step": 34625 + }, + { + "epoch": 71.995841995842, + "grad_norm": 0.0004101577214896679, + "learning_rate": 0.013149270093001675, + "loss": 0.2287, + "num_input_tokens_seen": 26411584, + "step": 34630 + }, + { + "epoch": 72.006237006237, + "grad_norm": 0.00016882021736819297, + "learning_rate": 0.013125162790194227, + "loss": 0.254, + "num_input_tokens_seen": 26415344, + "step": 34635 + }, + { + "epoch": 72.01663201663202, + "grad_norm": 0.00030382515978999436, + "learning_rate": 0.01310107659520674, + "loss": 0.2906, + "num_input_tokens_seen": 26419120, + "step": 34640 + }, + { + "epoch": 72.02702702702703, + "grad_norm": 0.00011020986130461097, + "learning_rate": 0.013077011511753655, + "loss": 0.2739, + "num_input_tokens_seen": 26423056, + "step": 34645 + }, + { + "epoch": 72.03742203742203, + "grad_norm": 0.0003761641273740679, + "learning_rate": 0.013052967543546056, + "loss": 0.2428, + "num_input_tokens_seen": 26427056, + "step": 34650 + }, + { + "epoch": 72.04781704781705, + "grad_norm": 8.697348675923422e-05, + "learning_rate": 0.01302894469429186, + "loss": 0.2513, + "num_input_tokens_seen": 26430864, + "step": 34655 + }, + { + "epoch": 72.05821205821206, + "grad_norm": 0.0002621742896735668, + "learning_rate": 0.013004942967695653, + "loss": 0.2484, + "num_input_tokens_seen": 26434672, + "step": 34660 + }, + { + "epoch": 72.06860706860707, + "grad_norm": 0.00011176583211636171, + "learning_rate": 0.012980962367458859, + "loss": 0.2544, + "num_input_tokens_seen": 26438480, + "step": 34665 + }, + { + "epoch": 72.07900207900208, + "grad_norm": 0.00014768738765269518, + "learning_rate": 0.012957002897279567, + "loss": 0.267, + "num_input_tokens_seen": 26442320, + "step": 34670 + }, + { + "epoch": 72.0893970893971, + "grad_norm": 0.001965432893484831, + "learning_rate": 0.012933064560852576, + "loss": 0.2506, + "num_input_tokens_seen": 26446096, + "step": 34675 + }, + { + "epoch": 72.0997920997921, + "grad_norm": 0.00010717882832977921, + "learning_rate": 0.012909147361869527, + "loss": 0.2678, + "num_input_tokens_seen": 26450064, + "step": 34680 + }, + { + "epoch": 72.11018711018711, + "grad_norm": 0.0012429632479324937, + "learning_rate": 0.012885251304018774, + "loss": 0.268, + "num_input_tokens_seen": 26453904, + "step": 34685 + }, + { + "epoch": 72.12058212058211, + "grad_norm": 0.0004679160483647138, + "learning_rate": 0.012861376390985335, + "loss": 0.2731, + "num_input_tokens_seen": 26457808, + "step": 34690 + }, + { + "epoch": 72.13097713097713, + "grad_norm": 0.00020628752827178687, + "learning_rate": 0.012837522626451063, + "loss": 0.2846, + "num_input_tokens_seen": 26461520, + "step": 34695 + }, + { + "epoch": 72.14137214137214, + "grad_norm": 0.00042619314626790583, + "learning_rate": 0.01281369001409447, + "loss": 0.2705, + "num_input_tokens_seen": 26465392, + "step": 34700 + }, + { + "epoch": 72.15176715176715, + "grad_norm": 0.0005593939567916095, + "learning_rate": 0.012789878557590877, + "loss": 0.269, + "num_input_tokens_seen": 26469040, + "step": 34705 + }, + { + "epoch": 72.16216216216216, + "grad_norm": 0.0006083393236622214, + "learning_rate": 0.012766088260612334, + "loss": 0.2298, + "num_input_tokens_seen": 26472784, + "step": 34710 + }, + { + "epoch": 72.17255717255718, + "grad_norm": 0.0003040891024284065, + "learning_rate": 0.012742319126827523, + "loss": 0.264, + "num_input_tokens_seen": 26476592, + "step": 34715 + }, + { + "epoch": 72.18295218295218, + "grad_norm": 0.00038121125544421375, + "learning_rate": 0.012718571159902008, + "loss": 0.2477, + "num_input_tokens_seen": 26480336, + "step": 34720 + }, + { + "epoch": 72.1933471933472, + "grad_norm": 0.0004151242901571095, + "learning_rate": 0.01269484436349803, + "loss": 0.2758, + "num_input_tokens_seen": 26484048, + "step": 34725 + }, + { + "epoch": 72.20374220374221, + "grad_norm": 0.00010516310430830345, + "learning_rate": 0.012671138741274528, + "loss": 0.2714, + "num_input_tokens_seen": 26487952, + "step": 34730 + }, + { + "epoch": 72.21413721413721, + "grad_norm": 0.00028693751664832234, + "learning_rate": 0.012647454296887194, + "loss": 0.264, + "num_input_tokens_seen": 26491888, + "step": 34735 + }, + { + "epoch": 72.22453222453223, + "grad_norm": 7.745195034658536e-05, + "learning_rate": 0.012623791033988507, + "loss": 0.2503, + "num_input_tokens_seen": 26495664, + "step": 34740 + }, + { + "epoch": 72.23492723492724, + "grad_norm": 0.0002716166200116277, + "learning_rate": 0.012600148956227597, + "loss": 0.2664, + "num_input_tokens_seen": 26499600, + "step": 34745 + }, + { + "epoch": 72.24532224532224, + "grad_norm": 0.00024009094340726733, + "learning_rate": 0.012576528067250414, + "loss": 0.2581, + "num_input_tokens_seen": 26503440, + "step": 34750 + }, + { + "epoch": 72.25571725571726, + "grad_norm": 0.0003295024507679045, + "learning_rate": 0.012552928370699561, + "loss": 0.2649, + "num_input_tokens_seen": 26507152, + "step": 34755 + }, + { + "epoch": 72.26611226611226, + "grad_norm": 0.0002833440084941685, + "learning_rate": 0.012529349870214411, + "loss": 0.2566, + "num_input_tokens_seen": 26510768, + "step": 34760 + }, + { + "epoch": 72.27650727650727, + "grad_norm": 0.00046820822171866894, + "learning_rate": 0.012505792569431106, + "loss": 0.272, + "num_input_tokens_seen": 26514832, + "step": 34765 + }, + { + "epoch": 72.28690228690229, + "grad_norm": 0.00015528849326074123, + "learning_rate": 0.012482256471982422, + "loss": 0.2613, + "num_input_tokens_seen": 26518608, + "step": 34770 + }, + { + "epoch": 72.29729729729729, + "grad_norm": 0.0003930999955628067, + "learning_rate": 0.012458741581497956, + "loss": 0.2159, + "num_input_tokens_seen": 26522480, + "step": 34775 + }, + { + "epoch": 72.3076923076923, + "grad_norm": 0.00011319175246171653, + "learning_rate": 0.012435247901603974, + "loss": 0.2578, + "num_input_tokens_seen": 26526192, + "step": 34780 + }, + { + "epoch": 72.31808731808732, + "grad_norm": 0.00017740637122187763, + "learning_rate": 0.012411775435923528, + "loss": 0.253, + "num_input_tokens_seen": 26530000, + "step": 34785 + }, + { + "epoch": 72.32848232848232, + "grad_norm": 0.0009421475697308779, + "learning_rate": 0.012388324188076354, + "loss": 0.2537, + "num_input_tokens_seen": 26533872, + "step": 34790 + }, + { + "epoch": 72.33887733887734, + "grad_norm": 0.00037671311292797327, + "learning_rate": 0.012364894161678913, + "loss": 0.2587, + "num_input_tokens_seen": 26537680, + "step": 34795 + }, + { + "epoch": 72.34927234927235, + "grad_norm": 8.218806760851294e-05, + "learning_rate": 0.012341485360344445, + "loss": 0.2666, + "num_input_tokens_seen": 26541680, + "step": 34800 + }, + { + "epoch": 72.34927234927235, + "eval_loss": 0.24757446348667145, + "eval_runtime": 13.4016, + "eval_samples_per_second": 63.873, + "eval_steps_per_second": 15.968, + "num_input_tokens_seen": 26541680, + "step": 34800 + }, + { + "epoch": 72.35966735966736, + "grad_norm": 0.0005148149211890996, + "learning_rate": 0.01231809778768283, + "loss": 0.2582, + "num_input_tokens_seen": 26545392, + "step": 34805 + }, + { + "epoch": 72.37006237006237, + "grad_norm": 9.076271089725196e-05, + "learning_rate": 0.012294731447300799, + "loss": 0.2493, + "num_input_tokens_seen": 26549232, + "step": 34810 + }, + { + "epoch": 72.38045738045739, + "grad_norm": 0.00010365769412601367, + "learning_rate": 0.012271386342801671, + "loss": 0.2515, + "num_input_tokens_seen": 26553008, + "step": 34815 + }, + { + "epoch": 72.39085239085239, + "grad_norm": 0.0005789612769149244, + "learning_rate": 0.012248062477785565, + "loss": 0.2392, + "num_input_tokens_seen": 26556880, + "step": 34820 + }, + { + "epoch": 72.4012474012474, + "grad_norm": 0.00039866394945420325, + "learning_rate": 0.012224759855849305, + "loss": 0.2665, + "num_input_tokens_seen": 26560752, + "step": 34825 + }, + { + "epoch": 72.41164241164242, + "grad_norm": 0.0005092307110317051, + "learning_rate": 0.012201478480586513, + "loss": 0.2497, + "num_input_tokens_seen": 26564496, + "step": 34830 + }, + { + "epoch": 72.42203742203742, + "grad_norm": 0.0001708591153146699, + "learning_rate": 0.012178218355587389, + "loss": 0.2572, + "num_input_tokens_seen": 26568304, + "step": 34835 + }, + { + "epoch": 72.43243243243244, + "grad_norm": 0.00021920054859947413, + "learning_rate": 0.01215497948443896, + "loss": 0.2578, + "num_input_tokens_seen": 26572016, + "step": 34840 + }, + { + "epoch": 72.44282744282744, + "grad_norm": 0.0016531936125829816, + "learning_rate": 0.012131761870724993, + "loss": 0.2752, + "num_input_tokens_seen": 26575952, + "step": 34845 + }, + { + "epoch": 72.45322245322245, + "grad_norm": 0.0010893176076933742, + "learning_rate": 0.012108565518025893, + "loss": 0.2259, + "num_input_tokens_seen": 26579984, + "step": 34850 + }, + { + "epoch": 72.46361746361747, + "grad_norm": 0.0002617994323372841, + "learning_rate": 0.012085390429918862, + "loss": 0.2733, + "num_input_tokens_seen": 26583888, + "step": 34855 + }, + { + "epoch": 72.47401247401247, + "grad_norm": 0.000740803312510252, + "learning_rate": 0.012062236609977744, + "loss": 0.2754, + "num_input_tokens_seen": 26587760, + "step": 34860 + }, + { + "epoch": 72.48440748440748, + "grad_norm": 0.0004866077797487378, + "learning_rate": 0.01203910406177318, + "loss": 0.2821, + "num_input_tokens_seen": 26591632, + "step": 34865 + }, + { + "epoch": 72.4948024948025, + "grad_norm": 0.0005302297067828476, + "learning_rate": 0.01201599278887252, + "loss": 0.2916, + "num_input_tokens_seen": 26595472, + "step": 34870 + }, + { + "epoch": 72.5051975051975, + "grad_norm": 0.00019920691556762904, + "learning_rate": 0.011992902794839744, + "loss": 0.2615, + "num_input_tokens_seen": 26599376, + "step": 34875 + }, + { + "epoch": 72.51559251559252, + "grad_norm": 0.00010601821122691035, + "learning_rate": 0.011969834083235703, + "loss": 0.2591, + "num_input_tokens_seen": 26603120, + "step": 34880 + }, + { + "epoch": 72.52598752598753, + "grad_norm": 0.0002832012833096087, + "learning_rate": 0.011946786657617836, + "loss": 0.2795, + "num_input_tokens_seen": 26606960, + "step": 34885 + }, + { + "epoch": 72.53638253638253, + "grad_norm": 0.000433982175309211, + "learning_rate": 0.011923760521540332, + "loss": 0.2535, + "num_input_tokens_seen": 26610768, + "step": 34890 + }, + { + "epoch": 72.54677754677755, + "grad_norm": 0.0022351054940372705, + "learning_rate": 0.011900755678554153, + "loss": 0.2539, + "num_input_tokens_seen": 26614768, + "step": 34895 + }, + { + "epoch": 72.55717255717256, + "grad_norm": 0.00026285924832336605, + "learning_rate": 0.011877772132206893, + "loss": 0.2682, + "num_input_tokens_seen": 26618640, + "step": 34900 + }, + { + "epoch": 72.56756756756756, + "grad_norm": 0.00027138934819959104, + "learning_rate": 0.011854809886042915, + "loss": 0.2584, + "num_input_tokens_seen": 26622320, + "step": 34905 + }, + { + "epoch": 72.57796257796258, + "grad_norm": 0.0017051596660166979, + "learning_rate": 0.011831868943603325, + "loss": 0.2262, + "num_input_tokens_seen": 26626064, + "step": 34910 + }, + { + "epoch": 72.58835758835758, + "grad_norm": 0.0004306395712774247, + "learning_rate": 0.011808949308425836, + "loss": 0.2477, + "num_input_tokens_seen": 26629968, + "step": 34915 + }, + { + "epoch": 72.5987525987526, + "grad_norm": 0.0005320144700817764, + "learning_rate": 0.01178605098404501, + "loss": 0.2788, + "num_input_tokens_seen": 26633712, + "step": 34920 + }, + { + "epoch": 72.60914760914761, + "grad_norm": 0.000345159147400409, + "learning_rate": 0.011763173973992002, + "loss": 0.2629, + "num_input_tokens_seen": 26637584, + "step": 34925 + }, + { + "epoch": 72.61954261954261, + "grad_norm": 0.0005791736184619367, + "learning_rate": 0.011740318281794776, + "loss": 0.2926, + "num_input_tokens_seen": 26641488, + "step": 34930 + }, + { + "epoch": 72.62993762993763, + "grad_norm": 0.002580442000180483, + "learning_rate": 0.01171748391097796, + "loss": 0.2751, + "num_input_tokens_seen": 26645200, + "step": 34935 + }, + { + "epoch": 72.64033264033264, + "grad_norm": 0.0007091288571245968, + "learning_rate": 0.011694670865062873, + "loss": 0.2533, + "num_input_tokens_seen": 26648880, + "step": 34940 + }, + { + "epoch": 72.65072765072765, + "grad_norm": 0.000738053466193378, + "learning_rate": 0.011671879147567616, + "loss": 0.2386, + "num_input_tokens_seen": 26652688, + "step": 34945 + }, + { + "epoch": 72.66112266112266, + "grad_norm": 0.0003458709397818893, + "learning_rate": 0.011649108762006893, + "loss": 0.2655, + "num_input_tokens_seen": 26656496, + "step": 34950 + }, + { + "epoch": 72.67151767151768, + "grad_norm": 0.00039297930197790265, + "learning_rate": 0.011626359711892265, + "loss": 0.2773, + "num_input_tokens_seen": 26660144, + "step": 34955 + }, + { + "epoch": 72.68191268191268, + "grad_norm": 0.00016521291399840266, + "learning_rate": 0.01160363200073189, + "loss": 0.2693, + "num_input_tokens_seen": 26664016, + "step": 34960 + }, + { + "epoch": 72.6923076923077, + "grad_norm": 9.515542478766292e-05, + "learning_rate": 0.011580925632030614, + "loss": 0.2542, + "num_input_tokens_seen": 26667856, + "step": 34965 + }, + { + "epoch": 72.70270270270271, + "grad_norm": 6.108672096161172e-05, + "learning_rate": 0.011558240609290104, + "loss": 0.2673, + "num_input_tokens_seen": 26671856, + "step": 34970 + }, + { + "epoch": 72.71309771309771, + "grad_norm": 3.3291846193606034e-05, + "learning_rate": 0.011535576936008679, + "loss": 0.2869, + "num_input_tokens_seen": 26675632, + "step": 34975 + }, + { + "epoch": 72.72349272349273, + "grad_norm": 0.00020035800116602331, + "learning_rate": 0.011512934615681309, + "loss": 0.2706, + "num_input_tokens_seen": 26679536, + "step": 34980 + }, + { + "epoch": 72.73388773388774, + "grad_norm": 0.0003292696492280811, + "learning_rate": 0.011490313651799765, + "loss": 0.2666, + "num_input_tokens_seen": 26683344, + "step": 34985 + }, + { + "epoch": 72.74428274428274, + "grad_norm": 0.0002502183197066188, + "learning_rate": 0.011467714047852512, + "loss": 0.2622, + "num_input_tokens_seen": 26687216, + "step": 34990 + }, + { + "epoch": 72.75467775467776, + "grad_norm": 0.00014884970732964575, + "learning_rate": 0.011445135807324624, + "loss": 0.2596, + "num_input_tokens_seen": 26691120, + "step": 34995 + }, + { + "epoch": 72.76507276507276, + "grad_norm": 0.00017719414609018713, + "learning_rate": 0.011422578933698002, + "loss": 0.2528, + "num_input_tokens_seen": 26694832, + "step": 35000 + }, + { + "epoch": 72.76507276507276, + "eval_loss": 0.2502278983592987, + "eval_runtime": 13.396, + "eval_samples_per_second": 63.9, + "eval_steps_per_second": 15.975, + "num_input_tokens_seen": 26694832, + "step": 35000 + }, + { + "epoch": 72.77546777546777, + "grad_norm": 0.00025865077623166144, + "learning_rate": 0.011400043430451161, + "loss": 0.2674, + "num_input_tokens_seen": 26698608, + "step": 35005 + }, + { + "epoch": 72.78586278586279, + "grad_norm": 0.00018091282981913537, + "learning_rate": 0.011377529301059392, + "loss": 0.2648, + "num_input_tokens_seen": 26702352, + "step": 35010 + }, + { + "epoch": 72.79625779625779, + "grad_norm": 3.604188532335684e-05, + "learning_rate": 0.011355036548994646, + "loss": 0.2652, + "num_input_tokens_seen": 26706064, + "step": 35015 + }, + { + "epoch": 72.8066528066528, + "grad_norm": 0.0003208151028957218, + "learning_rate": 0.011332565177725584, + "loss": 0.2929, + "num_input_tokens_seen": 26709808, + "step": 35020 + }, + { + "epoch": 72.81704781704782, + "grad_norm": 5.122231596033089e-05, + "learning_rate": 0.011310115190717585, + "loss": 0.2826, + "num_input_tokens_seen": 26713552, + "step": 35025 + }, + { + "epoch": 72.82744282744282, + "grad_norm": 0.000767905090469867, + "learning_rate": 0.01128768659143271, + "loss": 0.2594, + "num_input_tokens_seen": 26717456, + "step": 35030 + }, + { + "epoch": 72.83783783783784, + "grad_norm": 0.0005068514728918672, + "learning_rate": 0.011265279383329713, + "loss": 0.2711, + "num_input_tokens_seen": 26721200, + "step": 35035 + }, + { + "epoch": 72.84823284823285, + "grad_norm": 0.0001822207123041153, + "learning_rate": 0.01124289356986411, + "loss": 0.2526, + "num_input_tokens_seen": 26724880, + "step": 35040 + }, + { + "epoch": 72.85862785862786, + "grad_norm": 0.0009814156219363213, + "learning_rate": 0.011220529154488023, + "loss": 0.2644, + "num_input_tokens_seen": 26728656, + "step": 35045 + }, + { + "epoch": 72.86902286902287, + "grad_norm": 0.00026034426991827786, + "learning_rate": 0.011198186140650346, + "loss": 0.2699, + "num_input_tokens_seen": 26732432, + "step": 35050 + }, + { + "epoch": 72.87941787941789, + "grad_norm": 0.0008678545127622783, + "learning_rate": 0.011175864531796685, + "loss": 0.2621, + "num_input_tokens_seen": 26736272, + "step": 35055 + }, + { + "epoch": 72.88981288981289, + "grad_norm": 0.00023129310284275562, + "learning_rate": 0.011153564331369258, + "loss": 0.2609, + "num_input_tokens_seen": 26740016, + "step": 35060 + }, + { + "epoch": 72.9002079002079, + "grad_norm": 0.00033284572418779135, + "learning_rate": 0.011131285542807078, + "loss": 0.2636, + "num_input_tokens_seen": 26743728, + "step": 35065 + }, + { + "epoch": 72.9106029106029, + "grad_norm": 0.0004265870666131377, + "learning_rate": 0.011109028169545815, + "loss": 0.2488, + "num_input_tokens_seen": 26747568, + "step": 35070 + }, + { + "epoch": 72.92099792099792, + "grad_norm": 0.00025476107839494944, + "learning_rate": 0.011086792215017804, + "loss": 0.2733, + "num_input_tokens_seen": 26751440, + "step": 35075 + }, + { + "epoch": 72.93139293139293, + "grad_norm": 9.862656588666141e-05, + "learning_rate": 0.011064577682652137, + "loss": 0.2564, + "num_input_tokens_seen": 26755312, + "step": 35080 + }, + { + "epoch": 72.94178794178794, + "grad_norm": 0.0009094732231460512, + "learning_rate": 0.011042384575874559, + "loss": 0.2436, + "num_input_tokens_seen": 26759152, + "step": 35085 + }, + { + "epoch": 72.95218295218295, + "grad_norm": 0.00036140522570349276, + "learning_rate": 0.011020212898107512, + "loss": 0.2523, + "num_input_tokens_seen": 26762992, + "step": 35090 + }, + { + "epoch": 72.96257796257797, + "grad_norm": 0.00031674173078499734, + "learning_rate": 0.010998062652770197, + "loss": 0.2806, + "num_input_tokens_seen": 26766736, + "step": 35095 + }, + { + "epoch": 72.97297297297297, + "grad_norm": 0.00042393614421598613, + "learning_rate": 0.010975933843278428, + "loss": 0.2893, + "num_input_tokens_seen": 26770480, + "step": 35100 + }, + { + "epoch": 72.98336798336798, + "grad_norm": 0.0004447273095138371, + "learning_rate": 0.010953826473044714, + "loss": 0.256, + "num_input_tokens_seen": 26774480, + "step": 35105 + }, + { + "epoch": 72.993762993763, + "grad_norm": 8.807906851870939e-05, + "learning_rate": 0.010931740545478357, + "loss": 0.2752, + "num_input_tokens_seen": 26778256, + "step": 35110 + }, + { + "epoch": 73.004158004158, + "grad_norm": 0.0007353270193561912, + "learning_rate": 0.010909676063985218, + "loss": 0.2566, + "num_input_tokens_seen": 26781952, + "step": 35115 + }, + { + "epoch": 73.01455301455302, + "grad_norm": 0.00042906764429062605, + "learning_rate": 0.010887633031967974, + "loss": 0.2621, + "num_input_tokens_seen": 26785824, + "step": 35120 + }, + { + "epoch": 73.02494802494803, + "grad_norm": 0.0005121810827404261, + "learning_rate": 0.01086561145282589, + "loss": 0.2539, + "num_input_tokens_seen": 26789664, + "step": 35125 + }, + { + "epoch": 73.03534303534303, + "grad_norm": 0.00025208445731550455, + "learning_rate": 0.010843611329954983, + "loss": 0.2417, + "num_input_tokens_seen": 26793440, + "step": 35130 + }, + { + "epoch": 73.04573804573805, + "grad_norm": 0.0013416351284831762, + "learning_rate": 0.010821632666747988, + "loss": 0.2553, + "num_input_tokens_seen": 26797408, + "step": 35135 + }, + { + "epoch": 73.05613305613305, + "grad_norm": 0.0002883170382119715, + "learning_rate": 0.010799675466594244, + "loss": 0.261, + "num_input_tokens_seen": 26801280, + "step": 35140 + }, + { + "epoch": 73.06652806652806, + "grad_norm": 0.00020439941727090627, + "learning_rate": 0.010777739732879826, + "loss": 0.2691, + "num_input_tokens_seen": 26805088, + "step": 35145 + }, + { + "epoch": 73.07692307692308, + "grad_norm": 0.00012665581016335636, + "learning_rate": 0.010755825468987562, + "loss": 0.2714, + "num_input_tokens_seen": 26808960, + "step": 35150 + }, + { + "epoch": 73.08731808731808, + "grad_norm": 0.0001814990973798558, + "learning_rate": 0.010733932678296814, + "loss": 0.2468, + "num_input_tokens_seen": 26812832, + "step": 35155 + }, + { + "epoch": 73.0977130977131, + "grad_norm": 0.00015646299289073795, + "learning_rate": 0.010712061364183817, + "loss": 0.2685, + "num_input_tokens_seen": 26816736, + "step": 35160 + }, + { + "epoch": 73.10810810810811, + "grad_norm": 0.0002479937975294888, + "learning_rate": 0.010690211530021337, + "loss": 0.2608, + "num_input_tokens_seen": 26820768, + "step": 35165 + }, + { + "epoch": 73.11850311850311, + "grad_norm": 0.00042188348015770316, + "learning_rate": 0.01066838317917893, + "loss": 0.2432, + "num_input_tokens_seen": 26824640, + "step": 35170 + }, + { + "epoch": 73.12889812889813, + "grad_norm": 0.00025837679277174175, + "learning_rate": 0.010646576315022787, + "loss": 0.2348, + "num_input_tokens_seen": 26828384, + "step": 35175 + }, + { + "epoch": 73.13929313929314, + "grad_norm": 0.00011937220551772043, + "learning_rate": 0.010624790940915785, + "loss": 0.2774, + "num_input_tokens_seen": 26832192, + "step": 35180 + }, + { + "epoch": 73.14968814968815, + "grad_norm": 0.00022277713287621737, + "learning_rate": 0.0106030270602175, + "loss": 0.2651, + "num_input_tokens_seen": 26835968, + "step": 35185 + }, + { + "epoch": 73.16008316008316, + "grad_norm": 0.00046280043898150325, + "learning_rate": 0.010581284676284252, + "loss": 0.2747, + "num_input_tokens_seen": 26839776, + "step": 35190 + }, + { + "epoch": 73.17047817047818, + "grad_norm": 0.000174676941242069, + "learning_rate": 0.010559563792468923, + "loss": 0.2552, + "num_input_tokens_seen": 26843456, + "step": 35195 + }, + { + "epoch": 73.18087318087318, + "grad_norm": 0.00010934713645838201, + "learning_rate": 0.010537864412121217, + "loss": 0.263, + "num_input_tokens_seen": 26847168, + "step": 35200 + }, + { + "epoch": 73.18087318087318, + "eval_loss": 0.24831782281398773, + "eval_runtime": 13.4014, + "eval_samples_per_second": 63.874, + "eval_steps_per_second": 15.969, + "num_input_tokens_seen": 26847168, + "step": 35200 + }, + { + "epoch": 73.1912681912682, + "grad_norm": 0.0002878326049540192, + "learning_rate": 0.010516186538587357, + "loss": 0.2569, + "num_input_tokens_seen": 26850944, + "step": 35205 + }, + { + "epoch": 73.20166320166321, + "grad_norm": 0.0001266978360945359, + "learning_rate": 0.01049453017521042, + "loss": 0.242, + "num_input_tokens_seen": 26854720, + "step": 35210 + }, + { + "epoch": 73.21205821205821, + "grad_norm": 0.0001323156029684469, + "learning_rate": 0.010472895325330083, + "loss": 0.2573, + "num_input_tokens_seen": 26858720, + "step": 35215 + }, + { + "epoch": 73.22245322245323, + "grad_norm": 7.0307040004991e-05, + "learning_rate": 0.010451281992282662, + "loss": 0.256, + "num_input_tokens_seen": 26862432, + "step": 35220 + }, + { + "epoch": 73.23284823284823, + "grad_norm": 8.258586603915319e-05, + "learning_rate": 0.01042969017940124, + "loss": 0.2544, + "num_input_tokens_seen": 26865984, + "step": 35225 + }, + { + "epoch": 73.24324324324324, + "grad_norm": 0.0002530558267608285, + "learning_rate": 0.01040811989001557, + "loss": 0.2541, + "num_input_tokens_seen": 26869824, + "step": 35230 + }, + { + "epoch": 73.25363825363826, + "grad_norm": 0.0003990020486526191, + "learning_rate": 0.010386571127451992, + "loss": 0.2758, + "num_input_tokens_seen": 26873632, + "step": 35235 + }, + { + "epoch": 73.26403326403326, + "grad_norm": 0.000595601974055171, + "learning_rate": 0.010365043895033682, + "loss": 0.2667, + "num_input_tokens_seen": 26877632, + "step": 35240 + }, + { + "epoch": 73.27442827442827, + "grad_norm": 0.0005734565202146769, + "learning_rate": 0.010343538196080365, + "loss": 0.2695, + "num_input_tokens_seen": 26881504, + "step": 35245 + }, + { + "epoch": 73.28482328482329, + "grad_norm": 0.00043768910109065473, + "learning_rate": 0.010322054033908457, + "loss": 0.2622, + "num_input_tokens_seen": 26885280, + "step": 35250 + }, + { + "epoch": 73.29521829521829, + "grad_norm": 0.0010707448236644268, + "learning_rate": 0.010300591411831156, + "loss": 0.2869, + "num_input_tokens_seen": 26889184, + "step": 35255 + }, + { + "epoch": 73.3056133056133, + "grad_norm": 0.00043533314601518214, + "learning_rate": 0.010279150333158198, + "loss": 0.2641, + "num_input_tokens_seen": 26892896, + "step": 35260 + }, + { + "epoch": 73.31600831600832, + "grad_norm": 0.0002873156627174467, + "learning_rate": 0.010257730801196107, + "loss": 0.2677, + "num_input_tokens_seen": 26896768, + "step": 35265 + }, + { + "epoch": 73.32640332640332, + "grad_norm": 0.0003685600240714848, + "learning_rate": 0.010236332819248056, + "loss": 0.2886, + "num_input_tokens_seen": 26900544, + "step": 35270 + }, + { + "epoch": 73.33679833679834, + "grad_norm": 0.0005905661382712424, + "learning_rate": 0.010214956390613854, + "loss": 0.2893, + "num_input_tokens_seen": 26904352, + "step": 35275 + }, + { + "epoch": 73.34719334719335, + "grad_norm": 0.00015305352280847728, + "learning_rate": 0.010193601518590034, + "loss": 0.27, + "num_input_tokens_seen": 26908224, + "step": 35280 + }, + { + "epoch": 73.35758835758836, + "grad_norm": 0.00047347971121780574, + "learning_rate": 0.010172268206469758, + "loss": 0.2355, + "num_input_tokens_seen": 26912096, + "step": 35285 + }, + { + "epoch": 73.36798336798337, + "grad_norm": 0.00020713932462967932, + "learning_rate": 0.010150956457542897, + "loss": 0.2737, + "num_input_tokens_seen": 26916000, + "step": 35290 + }, + { + "epoch": 73.37837837837837, + "grad_norm": 0.00038774916902184486, + "learning_rate": 0.010129666275096054, + "loss": 0.2623, + "num_input_tokens_seen": 26920000, + "step": 35295 + }, + { + "epoch": 73.38877338877339, + "grad_norm": 3.673564060591161e-05, + "learning_rate": 0.010108397662412338, + "loss": 0.263, + "num_input_tokens_seen": 26923584, + "step": 35300 + }, + { + "epoch": 73.3991683991684, + "grad_norm": 0.0003616712347138673, + "learning_rate": 0.010087150622771707, + "loss": 0.2408, + "num_input_tokens_seen": 26927488, + "step": 35305 + }, + { + "epoch": 73.4095634095634, + "grad_norm": 0.00017674409900791943, + "learning_rate": 0.010065925159450739, + "loss": 0.2737, + "num_input_tokens_seen": 26931392, + "step": 35310 + }, + { + "epoch": 73.41995841995842, + "grad_norm": 0.0002015016507357359, + "learning_rate": 0.010044721275722618, + "loss": 0.2634, + "num_input_tokens_seen": 26935104, + "step": 35315 + }, + { + "epoch": 73.43035343035343, + "grad_norm": 0.0006162866484373808, + "learning_rate": 0.01002353897485726, + "loss": 0.2684, + "num_input_tokens_seen": 26938816, + "step": 35320 + }, + { + "epoch": 73.44074844074844, + "grad_norm": 0.00018490733054932207, + "learning_rate": 0.010002378260121236, + "loss": 0.2867, + "num_input_tokens_seen": 26942752, + "step": 35325 + }, + { + "epoch": 73.45114345114345, + "grad_norm": 0.0002533454680815339, + "learning_rate": 0.009981239134777786, + "loss": 0.253, + "num_input_tokens_seen": 26946624, + "step": 35330 + }, + { + "epoch": 73.46153846153847, + "grad_norm": 7.157825166359544e-05, + "learning_rate": 0.009960121602086884, + "loss": 0.2601, + "num_input_tokens_seen": 26950368, + "step": 35335 + }, + { + "epoch": 73.47193347193347, + "grad_norm": 0.0005854950868524611, + "learning_rate": 0.009939025665305062, + "loss": 0.2556, + "num_input_tokens_seen": 26954272, + "step": 35340 + }, + { + "epoch": 73.48232848232848, + "grad_norm": 0.00026317095034755766, + "learning_rate": 0.009917951327685597, + "loss": 0.2564, + "num_input_tokens_seen": 26958112, + "step": 35345 + }, + { + "epoch": 73.4927234927235, + "grad_norm": 0.002402608748525381, + "learning_rate": 0.009896898592478425, + "loss": 0.272, + "num_input_tokens_seen": 26961856, + "step": 35350 + }, + { + "epoch": 73.5031185031185, + "grad_norm": 0.0004182198317721486, + "learning_rate": 0.009875867462930132, + "loss": 0.2624, + "num_input_tokens_seen": 26965664, + "step": 35355 + }, + { + "epoch": 73.51351351351352, + "grad_norm": 0.00024241018400061876, + "learning_rate": 0.009854857942284006, + "loss": 0.2739, + "num_input_tokens_seen": 26969504, + "step": 35360 + }, + { + "epoch": 73.52390852390852, + "grad_norm": 0.00012619365588761866, + "learning_rate": 0.009833870033779923, + "loss": 0.2547, + "num_input_tokens_seen": 26973248, + "step": 35365 + }, + { + "epoch": 73.53430353430353, + "grad_norm": 0.0004259055422153324, + "learning_rate": 0.009812903740654527, + "loss": 0.2428, + "num_input_tokens_seen": 26976896, + "step": 35370 + }, + { + "epoch": 73.54469854469855, + "grad_norm": 0.0002757531765382737, + "learning_rate": 0.009791959066141097, + "loss": 0.2725, + "num_input_tokens_seen": 26980896, + "step": 35375 + }, + { + "epoch": 73.55509355509355, + "grad_norm": 8.038845408009365e-05, + "learning_rate": 0.009771036013469537, + "loss": 0.2368, + "num_input_tokens_seen": 26984736, + "step": 35380 + }, + { + "epoch": 73.56548856548856, + "grad_norm": 0.00015068287029862404, + "learning_rate": 0.00975013458586646, + "loss": 0.2394, + "num_input_tokens_seen": 26988448, + "step": 35385 + }, + { + "epoch": 73.57588357588358, + "grad_norm": 0.00041127673466689885, + "learning_rate": 0.009729254786555107, + "loss": 0.2655, + "num_input_tokens_seen": 26992192, + "step": 35390 + }, + { + "epoch": 73.58627858627858, + "grad_norm": 0.0004211366467643529, + "learning_rate": 0.009708396618755421, + "loss": 0.2577, + "num_input_tokens_seen": 26996128, + "step": 35395 + }, + { + "epoch": 73.5966735966736, + "grad_norm": 0.00023710649111308157, + "learning_rate": 0.009687560085683994, + "loss": 0.2583, + "num_input_tokens_seen": 27000096, + "step": 35400 + }, + { + "epoch": 73.5966735966736, + "eval_loss": 0.24813279509544373, + "eval_runtime": 13.4022, + "eval_samples_per_second": 63.87, + "eval_steps_per_second": 15.967, + "num_input_tokens_seen": 27000096, + "step": 35400 + }, + { + "epoch": 73.60706860706861, + "grad_norm": 0.0005935343215242028, + "learning_rate": 0.009666745190554054, + "loss": 0.2719, + "num_input_tokens_seen": 27003936, + "step": 35405 + }, + { + "epoch": 73.61746361746361, + "grad_norm": 0.0003768174210563302, + "learning_rate": 0.009645951936575553, + "loss": 0.2497, + "num_input_tokens_seen": 27007552, + "step": 35410 + }, + { + "epoch": 73.62785862785863, + "grad_norm": 0.0002991840592585504, + "learning_rate": 0.00962518032695509, + "loss": 0.2666, + "num_input_tokens_seen": 27011232, + "step": 35415 + }, + { + "epoch": 73.63825363825364, + "grad_norm": 8.902014087652788e-05, + "learning_rate": 0.009604430364895855, + "loss": 0.2591, + "num_input_tokens_seen": 27014944, + "step": 35420 + }, + { + "epoch": 73.64864864864865, + "grad_norm": 0.0010339658474549651, + "learning_rate": 0.00958370205359777, + "loss": 0.2538, + "num_input_tokens_seen": 27018688, + "step": 35425 + }, + { + "epoch": 73.65904365904366, + "grad_norm": 0.000444586796220392, + "learning_rate": 0.009562995396257445, + "loss": 0.2352, + "num_input_tokens_seen": 27022304, + "step": 35430 + }, + { + "epoch": 73.66943866943868, + "grad_norm": 0.0002212984545622021, + "learning_rate": 0.009542310396068026, + "loss": 0.2559, + "num_input_tokens_seen": 27026176, + "step": 35435 + }, + { + "epoch": 73.67983367983368, + "grad_norm": 0.00010784497862914577, + "learning_rate": 0.009521647056219495, + "loss": 0.2805, + "num_input_tokens_seen": 27029952, + "step": 35440 + }, + { + "epoch": 73.6902286902287, + "grad_norm": 0.000459110684460029, + "learning_rate": 0.00950100537989832, + "loss": 0.2795, + "num_input_tokens_seen": 27033792, + "step": 35445 + }, + { + "epoch": 73.7006237006237, + "grad_norm": 0.0003582810459192842, + "learning_rate": 0.00948038537028772, + "loss": 0.2777, + "num_input_tokens_seen": 27037600, + "step": 35450 + }, + { + "epoch": 73.71101871101871, + "grad_norm": 0.00031609792495146394, + "learning_rate": 0.009459787030567617, + "loss": 0.2465, + "num_input_tokens_seen": 27041472, + "step": 35455 + }, + { + "epoch": 73.72141372141373, + "grad_norm": 0.00017977422976400703, + "learning_rate": 0.00943921036391449, + "loss": 0.2599, + "num_input_tokens_seen": 27045216, + "step": 35460 + }, + { + "epoch": 73.73180873180873, + "grad_norm": 0.0003309495223220438, + "learning_rate": 0.009418655373501483, + "loss": 0.307, + "num_input_tokens_seen": 27049184, + "step": 35465 + }, + { + "epoch": 73.74220374220374, + "grad_norm": 0.00028912367997691035, + "learning_rate": 0.00939812206249851, + "loss": 0.2607, + "num_input_tokens_seen": 27053088, + "step": 35470 + }, + { + "epoch": 73.75259875259876, + "grad_norm": 0.00018243902013637125, + "learning_rate": 0.009377610434072004, + "loss": 0.2729, + "num_input_tokens_seen": 27056928, + "step": 35475 + }, + { + "epoch": 73.76299376299376, + "grad_norm": 0.0006715066265314817, + "learning_rate": 0.009357120491385167, + "loss": 0.2717, + "num_input_tokens_seen": 27060640, + "step": 35480 + }, + { + "epoch": 73.77338877338877, + "grad_norm": 0.00011176204861840233, + "learning_rate": 0.009336652237597743, + "loss": 0.2927, + "num_input_tokens_seen": 27064480, + "step": 35485 + }, + { + "epoch": 73.78378378378379, + "grad_norm": 0.00037878184230066836, + "learning_rate": 0.009316205675866251, + "loss": 0.2731, + "num_input_tokens_seen": 27068320, + "step": 35490 + }, + { + "epoch": 73.79417879417879, + "grad_norm": 0.0002593880635686219, + "learning_rate": 0.00929578080934379, + "loss": 0.2699, + "num_input_tokens_seen": 27072160, + "step": 35495 + }, + { + "epoch": 73.8045738045738, + "grad_norm": 0.0003337521920911968, + "learning_rate": 0.00927537764118012, + "loss": 0.2528, + "num_input_tokens_seen": 27075968, + "step": 35500 + }, + { + "epoch": 73.81496881496882, + "grad_norm": 0.0005197642603889108, + "learning_rate": 0.009254996174521678, + "loss": 0.2587, + "num_input_tokens_seen": 27079712, + "step": 35505 + }, + { + "epoch": 73.82536382536382, + "grad_norm": 0.00020184523600619286, + "learning_rate": 0.009234636412511531, + "loss": 0.2785, + "num_input_tokens_seen": 27083584, + "step": 35510 + }, + { + "epoch": 73.83575883575884, + "grad_norm": 0.0005622287862934172, + "learning_rate": 0.009214298358289418, + "loss": 0.2694, + "num_input_tokens_seen": 27087328, + "step": 35515 + }, + { + "epoch": 73.84615384615384, + "grad_norm": 0.0005193763645365834, + "learning_rate": 0.00919398201499173, + "loss": 0.2678, + "num_input_tokens_seen": 27091136, + "step": 35520 + }, + { + "epoch": 73.85654885654886, + "grad_norm": 0.0005127398180775344, + "learning_rate": 0.009173687385751495, + "loss": 0.2552, + "num_input_tokens_seen": 27094976, + "step": 35525 + }, + { + "epoch": 73.86694386694387, + "grad_norm": 0.00024458588450215757, + "learning_rate": 0.009153414473698407, + "loss": 0.2765, + "num_input_tokens_seen": 27098688, + "step": 35530 + }, + { + "epoch": 73.87733887733887, + "grad_norm": 0.000374671071767807, + "learning_rate": 0.009133163281958784, + "loss": 0.2609, + "num_input_tokens_seen": 27102560, + "step": 35535 + }, + { + "epoch": 73.88773388773389, + "grad_norm": 0.0005043321289122105, + "learning_rate": 0.009112933813655627, + "loss": 0.2696, + "num_input_tokens_seen": 27106400, + "step": 35540 + }, + { + "epoch": 73.8981288981289, + "grad_norm": 0.002000111388042569, + "learning_rate": 0.009092726071908573, + "loss": 0.2389, + "num_input_tokens_seen": 27110240, + "step": 35545 + }, + { + "epoch": 73.9085239085239, + "grad_norm": 0.00019785486801993102, + "learning_rate": 0.0090725400598339, + "loss": 0.2647, + "num_input_tokens_seen": 27113888, + "step": 35550 + }, + { + "epoch": 73.91891891891892, + "grad_norm": 0.00015866119065321982, + "learning_rate": 0.009052375780544563, + "loss": 0.2571, + "num_input_tokens_seen": 27117600, + "step": 35555 + }, + { + "epoch": 73.92931392931393, + "grad_norm": 0.00025204537087120116, + "learning_rate": 0.009032233237150144, + "loss": 0.2679, + "num_input_tokens_seen": 27121280, + "step": 35560 + }, + { + "epoch": 73.93970893970894, + "grad_norm": 0.0001546829444123432, + "learning_rate": 0.009012112432756875, + "loss": 0.2565, + "num_input_tokens_seen": 27125184, + "step": 35565 + }, + { + "epoch": 73.95010395010395, + "grad_norm": 0.00026677310233935714, + "learning_rate": 0.008992013370467605, + "loss": 0.2705, + "num_input_tokens_seen": 27129088, + "step": 35570 + }, + { + "epoch": 73.96049896049897, + "grad_norm": 0.0006169647094793618, + "learning_rate": 0.008971936053381924, + "loss": 0.2638, + "num_input_tokens_seen": 27133024, + "step": 35575 + }, + { + "epoch": 73.97089397089397, + "grad_norm": 0.0006587838288396597, + "learning_rate": 0.008951880484595953, + "loss": 0.277, + "num_input_tokens_seen": 27136800, + "step": 35580 + }, + { + "epoch": 73.98128898128898, + "grad_norm": 0.0009608623804524541, + "learning_rate": 0.008931846667202552, + "loss": 0.2309, + "num_input_tokens_seen": 27140544, + "step": 35585 + }, + { + "epoch": 73.99168399168398, + "grad_norm": 0.00029404196538962424, + "learning_rate": 0.008911834604291152, + "loss": 0.2647, + "num_input_tokens_seen": 27144352, + "step": 35590 + }, + { + "epoch": 74.002079002079, + "grad_norm": 0.000500818423461169, + "learning_rate": 0.008891844298947882, + "loss": 0.2463, + "num_input_tokens_seen": 27147960, + "step": 35595 + }, + { + "epoch": 74.01247401247402, + "grad_norm": 0.0007572741596959531, + "learning_rate": 0.008871875754255508, + "loss": 0.277, + "num_input_tokens_seen": 27151800, + "step": 35600 + }, + { + "epoch": 74.01247401247402, + "eval_loss": 0.24922853708267212, + "eval_runtime": 13.3916, + "eval_samples_per_second": 63.921, + "eval_steps_per_second": 15.98, + "num_input_tokens_seen": 27151800, + "step": 35600 + }, + { + "epoch": 74.02286902286902, + "grad_norm": 0.0003930544189643115, + "learning_rate": 0.008851928973293422, + "loss": 0.2666, + "num_input_tokens_seen": 27155736, + "step": 35605 + }, + { + "epoch": 74.03326403326403, + "grad_norm": 0.00026405713288113475, + "learning_rate": 0.00883200395913764, + "loss": 0.2664, + "num_input_tokens_seen": 27159672, + "step": 35610 + }, + { + "epoch": 74.04365904365905, + "grad_norm": 0.000271029508439824, + "learning_rate": 0.00881210071486091, + "loss": 0.2648, + "num_input_tokens_seen": 27163320, + "step": 35615 + }, + { + "epoch": 74.05405405405405, + "grad_norm": 0.0003909572260454297, + "learning_rate": 0.008792219243532505, + "loss": 0.2635, + "num_input_tokens_seen": 27167160, + "step": 35620 + }, + { + "epoch": 74.06444906444906, + "grad_norm": 0.0005478140083141625, + "learning_rate": 0.008772359548218428, + "loss": 0.259, + "num_input_tokens_seen": 27170968, + "step": 35625 + }, + { + "epoch": 74.07484407484408, + "grad_norm": 0.0003560836485121399, + "learning_rate": 0.008752521631981274, + "loss": 0.2608, + "num_input_tokens_seen": 27174776, + "step": 35630 + }, + { + "epoch": 74.08523908523908, + "grad_norm": 0.0005666264914907515, + "learning_rate": 0.008732705497880315, + "loss": 0.2632, + "num_input_tokens_seen": 27178808, + "step": 35635 + }, + { + "epoch": 74.0956340956341, + "grad_norm": 0.0003599419433157891, + "learning_rate": 0.008712911148971459, + "loss": 0.2802, + "num_input_tokens_seen": 27182520, + "step": 35640 + }, + { + "epoch": 74.10602910602911, + "grad_norm": 0.00027648115064948797, + "learning_rate": 0.008693138588307208, + "loss": 0.2531, + "num_input_tokens_seen": 27186168, + "step": 35645 + }, + { + "epoch": 74.11642411642411, + "grad_norm": 0.0002439768722979352, + "learning_rate": 0.008673387818936762, + "loss": 0.2913, + "num_input_tokens_seen": 27190040, + "step": 35650 + }, + { + "epoch": 74.12681912681913, + "grad_norm": 0.0001416257000528276, + "learning_rate": 0.008653658843905948, + "loss": 0.2622, + "num_input_tokens_seen": 27193848, + "step": 35655 + }, + { + "epoch": 74.13721413721414, + "grad_norm": 0.00029875419568270445, + "learning_rate": 0.0086339516662572, + "loss": 0.2816, + "num_input_tokens_seen": 27197752, + "step": 35660 + }, + { + "epoch": 74.14760914760915, + "grad_norm": 0.00013812282122671604, + "learning_rate": 0.008614266289029638, + "loss": 0.2542, + "num_input_tokens_seen": 27201688, + "step": 35665 + }, + { + "epoch": 74.15800415800416, + "grad_norm": 0.00044145682477392256, + "learning_rate": 0.008594602715258965, + "loss": 0.2661, + "num_input_tokens_seen": 27205368, + "step": 35670 + }, + { + "epoch": 74.16839916839916, + "grad_norm": 0.0007312553352676332, + "learning_rate": 0.008574960947977573, + "loss": 0.2739, + "num_input_tokens_seen": 27209240, + "step": 35675 + }, + { + "epoch": 74.17879417879418, + "grad_norm": 0.0006273904000408947, + "learning_rate": 0.008555340990214438, + "loss": 0.2498, + "num_input_tokens_seen": 27213048, + "step": 35680 + }, + { + "epoch": 74.1891891891892, + "grad_norm": 0.00042457215022295713, + "learning_rate": 0.008535742844995258, + "loss": 0.261, + "num_input_tokens_seen": 27216728, + "step": 35685 + }, + { + "epoch": 74.1995841995842, + "grad_norm": 0.00018563997582532465, + "learning_rate": 0.008516166515342266, + "loss": 0.2599, + "num_input_tokens_seen": 27220664, + "step": 35690 + }, + { + "epoch": 74.20997920997921, + "grad_norm": 0.001080030226148665, + "learning_rate": 0.008496612004274411, + "loss": 0.2649, + "num_input_tokens_seen": 27224600, + "step": 35695 + }, + { + "epoch": 74.22037422037423, + "grad_norm": 0.0003102063783444464, + "learning_rate": 0.008477079314807201, + "loss": 0.2412, + "num_input_tokens_seen": 27228312, + "step": 35700 + }, + { + "epoch": 74.23076923076923, + "grad_norm": 0.00018369620374869555, + "learning_rate": 0.008457568449952874, + "loss": 0.2463, + "num_input_tokens_seen": 27232088, + "step": 35705 + }, + { + "epoch": 74.24116424116424, + "grad_norm": 0.000136344155180268, + "learning_rate": 0.008438079412720189, + "loss": 0.2591, + "num_input_tokens_seen": 27235960, + "step": 35710 + }, + { + "epoch": 74.25155925155926, + "grad_norm": 0.00017369123816024512, + "learning_rate": 0.00841861220611466, + "loss": 0.2742, + "num_input_tokens_seen": 27239896, + "step": 35715 + }, + { + "epoch": 74.26195426195426, + "grad_norm": 0.00031644431874156, + "learning_rate": 0.008399166833138355, + "loss": 0.2879, + "num_input_tokens_seen": 27243736, + "step": 35720 + }, + { + "epoch": 74.27234927234927, + "grad_norm": 0.0001117918873205781, + "learning_rate": 0.008379743296789987, + "loss": 0.2422, + "num_input_tokens_seen": 27247576, + "step": 35725 + }, + { + "epoch": 74.28274428274429, + "grad_norm": 0.0002494224754627794, + "learning_rate": 0.008360341600064896, + "loss": 0.2592, + "num_input_tokens_seen": 27251320, + "step": 35730 + }, + { + "epoch": 74.29313929313929, + "grad_norm": 0.0003176149621140212, + "learning_rate": 0.008340961745955121, + "loss": 0.2499, + "num_input_tokens_seen": 27255160, + "step": 35735 + }, + { + "epoch": 74.3035343035343, + "grad_norm": 0.0001895570894703269, + "learning_rate": 0.008321603737449224, + "loss": 0.2687, + "num_input_tokens_seen": 27258936, + "step": 35740 + }, + { + "epoch": 74.31392931392931, + "grad_norm": 0.0013831006363034248, + "learning_rate": 0.008302267577532479, + "loss": 0.2325, + "num_input_tokens_seen": 27262776, + "step": 35745 + }, + { + "epoch": 74.32432432432432, + "grad_norm": 0.0002702161727938801, + "learning_rate": 0.008282953269186771, + "loss": 0.2621, + "num_input_tokens_seen": 27266680, + "step": 35750 + }, + { + "epoch": 74.33471933471934, + "grad_norm": 0.0006112667615525424, + "learning_rate": 0.008263660815390567, + "loss": 0.2681, + "num_input_tokens_seen": 27270296, + "step": 35755 + }, + { + "epoch": 74.34511434511434, + "grad_norm": 0.0010118919890373945, + "learning_rate": 0.008244390219119069, + "loss": 0.2742, + "num_input_tokens_seen": 27274104, + "step": 35760 + }, + { + "epoch": 74.35550935550935, + "grad_norm": 0.000422335957409814, + "learning_rate": 0.008225141483343967, + "loss": 0.2533, + "num_input_tokens_seen": 27277880, + "step": 35765 + }, + { + "epoch": 74.36590436590437, + "grad_norm": 0.00032449059654027224, + "learning_rate": 0.00820591461103372, + "loss": 0.2684, + "num_input_tokens_seen": 27281688, + "step": 35770 + }, + { + "epoch": 74.37629937629937, + "grad_norm": 0.000574265664909035, + "learning_rate": 0.008186709605153358, + "loss": 0.2815, + "num_input_tokens_seen": 27285496, + "step": 35775 + }, + { + "epoch": 74.38669438669439, + "grad_norm": 0.00011316668678773567, + "learning_rate": 0.008167526468664492, + "loss": 0.258, + "num_input_tokens_seen": 27289208, + "step": 35780 + }, + { + "epoch": 74.3970893970894, + "grad_norm": 0.0003479458100628108, + "learning_rate": 0.008148365204525443, + "loss": 0.2571, + "num_input_tokens_seen": 27292984, + "step": 35785 + }, + { + "epoch": 74.4074844074844, + "grad_norm": 0.0009247698471881449, + "learning_rate": 0.00812922581569106, + "loss": 0.2665, + "num_input_tokens_seen": 27296664, + "step": 35790 + }, + { + "epoch": 74.41787941787942, + "grad_norm": 0.00012059063010383397, + "learning_rate": 0.008110108305112934, + "loss": 0.2482, + "num_input_tokens_seen": 27300440, + "step": 35795 + }, + { + "epoch": 74.42827442827443, + "grad_norm": 0.0003978035820182413, + "learning_rate": 0.008091012675739223, + "loss": 0.2486, + "num_input_tokens_seen": 27304152, + "step": 35800 + }, + { + "epoch": 74.42827442827443, + "eval_loss": 0.2478523552417755, + "eval_runtime": 13.3972, + "eval_samples_per_second": 63.894, + "eval_steps_per_second": 15.973, + "num_input_tokens_seen": 27304152, + "step": 35800 + }, + { + "epoch": 74.43866943866944, + "grad_norm": 0.00023250629601534456, + "learning_rate": 0.008071938930514671, + "loss": 0.2332, + "num_input_tokens_seen": 27307832, + "step": 35805 + }, + { + "epoch": 74.44906444906445, + "grad_norm": 0.0001280247379327193, + "learning_rate": 0.008052887072380726, + "loss": 0.2913, + "num_input_tokens_seen": 27311672, + "step": 35810 + }, + { + "epoch": 74.45945945945945, + "grad_norm": 0.0006330225733108819, + "learning_rate": 0.008033857104275437, + "loss": 0.2752, + "num_input_tokens_seen": 27315736, + "step": 35815 + }, + { + "epoch": 74.46985446985447, + "grad_norm": 0.0016815828857943416, + "learning_rate": 0.008014849029133424, + "loss": 0.2735, + "num_input_tokens_seen": 27319384, + "step": 35820 + }, + { + "epoch": 74.48024948024948, + "grad_norm": 0.0002503742289263755, + "learning_rate": 0.007995862849885975, + "loss": 0.2281, + "num_input_tokens_seen": 27323128, + "step": 35825 + }, + { + "epoch": 74.49064449064448, + "grad_norm": 0.0003067964571528137, + "learning_rate": 0.007976898569461032, + "loss": 0.2824, + "num_input_tokens_seen": 27327128, + "step": 35830 + }, + { + "epoch": 74.5010395010395, + "grad_norm": 0.0007298548589460552, + "learning_rate": 0.007957956190783088, + "loss": 0.2702, + "num_input_tokens_seen": 27330872, + "step": 35835 + }, + { + "epoch": 74.51143451143452, + "grad_norm": 0.00034885347122326493, + "learning_rate": 0.007939035716773324, + "loss": 0.2668, + "num_input_tokens_seen": 27334648, + "step": 35840 + }, + { + "epoch": 74.52182952182952, + "grad_norm": 0.0004228349425829947, + "learning_rate": 0.007920137150349487, + "loss": 0.2624, + "num_input_tokens_seen": 27338520, + "step": 35845 + }, + { + "epoch": 74.53222453222453, + "grad_norm": 0.000176787783857435, + "learning_rate": 0.007901260494425981, + "loss": 0.2842, + "num_input_tokens_seen": 27342424, + "step": 35850 + }, + { + "epoch": 74.54261954261955, + "grad_norm": 0.0005410254816524684, + "learning_rate": 0.007882405751913861, + "loss": 0.2765, + "num_input_tokens_seen": 27346232, + "step": 35855 + }, + { + "epoch": 74.55301455301455, + "grad_norm": 0.0005940303090028465, + "learning_rate": 0.007863572925720702, + "loss": 0.2567, + "num_input_tokens_seen": 27349720, + "step": 35860 + }, + { + "epoch": 74.56340956340956, + "grad_norm": 0.0006077117286622524, + "learning_rate": 0.007844762018750827, + "loss": 0.2542, + "num_input_tokens_seen": 27353560, + "step": 35865 + }, + { + "epoch": 74.57380457380458, + "grad_norm": 0.0007424021605402231, + "learning_rate": 0.007825973033905054, + "loss": 0.2553, + "num_input_tokens_seen": 27357368, + "step": 35870 + }, + { + "epoch": 74.58419958419958, + "grad_norm": 0.0001844785874709487, + "learning_rate": 0.007807205974080927, + "loss": 0.2477, + "num_input_tokens_seen": 27361304, + "step": 35875 + }, + { + "epoch": 74.5945945945946, + "grad_norm": 0.00019676057854667306, + "learning_rate": 0.007788460842172551, + "loss": 0.2423, + "num_input_tokens_seen": 27365112, + "step": 35880 + }, + { + "epoch": 74.60498960498961, + "grad_norm": 0.00031554396264255047, + "learning_rate": 0.0077697376410706285, + "loss": 0.257, + "num_input_tokens_seen": 27368920, + "step": 35885 + }, + { + "epoch": 74.61538461538461, + "grad_norm": 0.0018940113950520754, + "learning_rate": 0.007751036373662567, + "loss": 0.2243, + "num_input_tokens_seen": 27372600, + "step": 35890 + }, + { + "epoch": 74.62577962577963, + "grad_norm": 0.0010034558363258839, + "learning_rate": 0.00773235704283231, + "loss": 0.2557, + "num_input_tokens_seen": 27376568, + "step": 35895 + }, + { + "epoch": 74.63617463617463, + "grad_norm": 0.0005155546241439879, + "learning_rate": 0.007713699651460437, + "loss": 0.2402, + "num_input_tokens_seen": 27380376, + "step": 35900 + }, + { + "epoch": 74.64656964656965, + "grad_norm": 0.0012954804114997387, + "learning_rate": 0.007695064202424162, + "loss": 0.2762, + "num_input_tokens_seen": 27384024, + "step": 35905 + }, + { + "epoch": 74.65696465696466, + "grad_norm": 0.00025558372726663947, + "learning_rate": 0.007676450698597286, + "loss": 0.2571, + "num_input_tokens_seen": 27387736, + "step": 35910 + }, + { + "epoch": 74.66735966735966, + "grad_norm": 0.00023000799410510808, + "learning_rate": 0.007657859142850265, + "loss": 0.2316, + "num_input_tokens_seen": 27391576, + "step": 35915 + }, + { + "epoch": 74.67775467775468, + "grad_norm": 0.001397228566929698, + "learning_rate": 0.0076392895380501535, + "loss": 0.3017, + "num_input_tokens_seen": 27395384, + "step": 35920 + }, + { + "epoch": 74.6881496881497, + "grad_norm": 0.0001082841627066955, + "learning_rate": 0.007620741887060611, + "loss": 0.2786, + "num_input_tokens_seen": 27399352, + "step": 35925 + }, + { + "epoch": 74.6985446985447, + "grad_norm": 0.00027650519041344523, + "learning_rate": 0.007602216192741901, + "loss": 0.2839, + "num_input_tokens_seen": 27403288, + "step": 35930 + }, + { + "epoch": 74.70893970893971, + "grad_norm": 0.0003521101316437125, + "learning_rate": 0.007583712457950969, + "loss": 0.2936, + "num_input_tokens_seen": 27407224, + "step": 35935 + }, + { + "epoch": 74.71933471933473, + "grad_norm": 0.0006344080320559442, + "learning_rate": 0.007565230685541269, + "loss": 0.2603, + "num_input_tokens_seen": 27411064, + "step": 35940 + }, + { + "epoch": 74.72972972972973, + "grad_norm": 0.00043430578080005944, + "learning_rate": 0.007546770878362968, + "loss": 0.2411, + "num_input_tokens_seen": 27414808, + "step": 35945 + }, + { + "epoch": 74.74012474012474, + "grad_norm": 0.00025257677771151066, + "learning_rate": 0.0075283330392627405, + "loss": 0.2412, + "num_input_tokens_seen": 27418712, + "step": 35950 + }, + { + "epoch": 74.75051975051976, + "grad_norm": 0.0006738262018188834, + "learning_rate": 0.007509917171083979, + "loss": 0.2301, + "num_input_tokens_seen": 27422552, + "step": 35955 + }, + { + "epoch": 74.76091476091476, + "grad_norm": 0.000822072965092957, + "learning_rate": 0.007491523276666662, + "loss": 0.267, + "num_input_tokens_seen": 27426328, + "step": 35960 + }, + { + "epoch": 74.77130977130977, + "grad_norm": 0.00027745121042244136, + "learning_rate": 0.007473151358847318, + "loss": 0.2645, + "num_input_tokens_seen": 27430072, + "step": 35965 + }, + { + "epoch": 74.78170478170478, + "grad_norm": 0.0002748681581579149, + "learning_rate": 0.007454801420459117, + "loss": 0.2651, + "num_input_tokens_seen": 27433880, + "step": 35970 + }, + { + "epoch": 74.79209979209979, + "grad_norm": 0.0006046561175026, + "learning_rate": 0.0074364734643319105, + "loss": 0.2885, + "num_input_tokens_seen": 27437592, + "step": 35975 + }, + { + "epoch": 74.8024948024948, + "grad_norm": 0.00027396451332606375, + "learning_rate": 0.007418167493292022, + "loss": 0.2618, + "num_input_tokens_seen": 27441400, + "step": 35980 + }, + { + "epoch": 74.81288981288981, + "grad_norm": 0.00034973659785464406, + "learning_rate": 0.0073998835101625245, + "loss": 0.2874, + "num_input_tokens_seen": 27445080, + "step": 35985 + }, + { + "epoch": 74.82328482328482, + "grad_norm": 0.00021457408729474992, + "learning_rate": 0.007381621517762998, + "loss": 0.2869, + "num_input_tokens_seen": 27449048, + "step": 35990 + }, + { + "epoch": 74.83367983367984, + "grad_norm": 0.0010342495515942574, + "learning_rate": 0.007363381518909689, + "loss": 0.2674, + "num_input_tokens_seen": 27452984, + "step": 35995 + }, + { + "epoch": 74.84407484407484, + "grad_norm": 9.518170554656535e-05, + "learning_rate": 0.007345163516415448, + "loss": 0.2598, + "num_input_tokens_seen": 27456856, + "step": 36000 + }, + { + "epoch": 74.84407484407484, + "eval_loss": 0.2507597804069519, + "eval_runtime": 13.3904, + "eval_samples_per_second": 63.927, + "eval_steps_per_second": 15.982, + "num_input_tokens_seen": 27456856, + "step": 36000 + }, + { + "epoch": 74.85446985446985, + "grad_norm": 0.00029960134997963905, + "learning_rate": 0.007326967513089693, + "loss": 0.2433, + "num_input_tokens_seen": 27460760, + "step": 36005 + }, + { + "epoch": 74.86486486486487, + "grad_norm": 0.00023461967066396028, + "learning_rate": 0.0073087935117384815, + "loss": 0.2673, + "num_input_tokens_seen": 27464728, + "step": 36010 + }, + { + "epoch": 74.87525987525987, + "grad_norm": 0.00046880284207873046, + "learning_rate": 0.007290641515164503, + "loss": 0.2282, + "num_input_tokens_seen": 27468600, + "step": 36015 + }, + { + "epoch": 74.88565488565489, + "grad_norm": 0.0017863644752651453, + "learning_rate": 0.007272511526166986, + "loss": 0.2732, + "num_input_tokens_seen": 27472440, + "step": 36020 + }, + { + "epoch": 74.8960498960499, + "grad_norm": 0.00017287139780819416, + "learning_rate": 0.0072544035475418265, + "loss": 0.271, + "num_input_tokens_seen": 27476248, + "step": 36025 + }, + { + "epoch": 74.9064449064449, + "grad_norm": 0.0004917072947137058, + "learning_rate": 0.007236317582081475, + "loss": 0.265, + "num_input_tokens_seen": 27480120, + "step": 36030 + }, + { + "epoch": 74.91683991683992, + "grad_norm": 0.0001281521690543741, + "learning_rate": 0.007218253632575066, + "loss": 0.2537, + "num_input_tokens_seen": 27484120, + "step": 36035 + }, + { + "epoch": 74.92723492723492, + "grad_norm": 0.0003193325537722558, + "learning_rate": 0.007200211701808223, + "loss": 0.2564, + "num_input_tokens_seen": 27487800, + "step": 36040 + }, + { + "epoch": 74.93762993762994, + "grad_norm": 0.00043910779641009867, + "learning_rate": 0.007182191792563286, + "loss": 0.2801, + "num_input_tokens_seen": 27491544, + "step": 36045 + }, + { + "epoch": 74.94802494802495, + "grad_norm": 0.00034964369842782617, + "learning_rate": 0.0071641939076191145, + "loss": 0.2752, + "num_input_tokens_seen": 27495352, + "step": 36050 + }, + { + "epoch": 74.95841995841995, + "grad_norm": 0.00023063960543368012, + "learning_rate": 0.007146218049751257, + "loss": 0.2686, + "num_input_tokens_seen": 27499416, + "step": 36055 + }, + { + "epoch": 74.96881496881497, + "grad_norm": 0.0002742695214692503, + "learning_rate": 0.0071282642217317775, + "loss": 0.2578, + "num_input_tokens_seen": 27503224, + "step": 36060 + }, + { + "epoch": 74.97920997920998, + "grad_norm": 0.003657560097053647, + "learning_rate": 0.007110332426329396, + "loss": 0.2863, + "num_input_tokens_seen": 27507224, + "step": 36065 + }, + { + "epoch": 74.98960498960498, + "grad_norm": 0.0004119256045669317, + "learning_rate": 0.007092422666309417, + "loss": 0.2443, + "num_input_tokens_seen": 27511160, + "step": 36070 + }, + { + "epoch": 75.0, + "grad_norm": 0.00017305923392996192, + "learning_rate": 0.0070745349444337295, + "loss": 0.2649, + "num_input_tokens_seen": 27514792, + "step": 36075 + }, + { + "epoch": 75.01039501039502, + "grad_norm": 0.0001678390835877508, + "learning_rate": 0.007056669263460913, + "loss": 0.2543, + "num_input_tokens_seen": 27518632, + "step": 36080 + }, + { + "epoch": 75.02079002079002, + "grad_norm": 0.0006255152984522283, + "learning_rate": 0.007038825626145995, + "loss": 0.2587, + "num_input_tokens_seen": 27522408, + "step": 36085 + }, + { + "epoch": 75.03118503118503, + "grad_norm": 0.0003666336415335536, + "learning_rate": 0.007021004035240724, + "loss": 0.2499, + "num_input_tokens_seen": 27526280, + "step": 36090 + }, + { + "epoch": 75.04158004158005, + "grad_norm": 0.00011675449059111997, + "learning_rate": 0.007003204493493453, + "loss": 0.2808, + "num_input_tokens_seen": 27530088, + "step": 36095 + }, + { + "epoch": 75.05197505197505, + "grad_norm": 0.0002810081350617111, + "learning_rate": 0.006985427003649036, + "loss": 0.2589, + "num_input_tokens_seen": 27534088, + "step": 36100 + }, + { + "epoch": 75.06237006237006, + "grad_norm": 0.00029885617550462484, + "learning_rate": 0.006967671568449013, + "loss": 0.2763, + "num_input_tokens_seen": 27537832, + "step": 36105 + }, + { + "epoch": 75.07276507276508, + "grad_norm": 0.0006143073551356792, + "learning_rate": 0.006949938190631511, + "loss": 0.2445, + "num_input_tokens_seen": 27541512, + "step": 36110 + }, + { + "epoch": 75.08316008316008, + "grad_norm": 0.0012581292539834976, + "learning_rate": 0.0069322268729311905, + "loss": 0.2561, + "num_input_tokens_seen": 27545448, + "step": 36115 + }, + { + "epoch": 75.0935550935551, + "grad_norm": 0.0007887134561315179, + "learning_rate": 0.006914537618079403, + "loss": 0.2629, + "num_input_tokens_seen": 27549224, + "step": 36120 + }, + { + "epoch": 75.1039501039501, + "grad_norm": 0.000439716357504949, + "learning_rate": 0.006896870428804031, + "loss": 0.2571, + "num_input_tokens_seen": 27553096, + "step": 36125 + }, + { + "epoch": 75.11434511434511, + "grad_norm": 0.00021960939920973033, + "learning_rate": 0.006879225307829595, + "loss": 0.2726, + "num_input_tokens_seen": 27556968, + "step": 36130 + }, + { + "epoch": 75.12474012474013, + "grad_norm": 0.0004342082538641989, + "learning_rate": 0.00686160225787717, + "loss": 0.2434, + "num_input_tokens_seen": 27560744, + "step": 36135 + }, + { + "epoch": 75.13513513513513, + "grad_norm": 0.0001964966650120914, + "learning_rate": 0.006844001281664463, + "loss": 0.2775, + "num_input_tokens_seen": 27564584, + "step": 36140 + }, + { + "epoch": 75.14553014553015, + "grad_norm": 0.0001552668836666271, + "learning_rate": 0.006826422381905789, + "loss": 0.2856, + "num_input_tokens_seen": 27568424, + "step": 36145 + }, + { + "epoch": 75.15592515592516, + "grad_norm": 0.0002786397817544639, + "learning_rate": 0.006808865561311994, + "loss": 0.2756, + "num_input_tokens_seen": 27572424, + "step": 36150 + }, + { + "epoch": 75.16632016632016, + "grad_norm": 0.0004274522070772946, + "learning_rate": 0.00679133082259058, + "loss": 0.2648, + "num_input_tokens_seen": 27576136, + "step": 36155 + }, + { + "epoch": 75.17671517671518, + "grad_norm": 0.00017544864385854453, + "learning_rate": 0.00677381816844565, + "loss": 0.256, + "num_input_tokens_seen": 27580008, + "step": 36160 + }, + { + "epoch": 75.18711018711019, + "grad_norm": 0.0001505874388385564, + "learning_rate": 0.0067563276015778434, + "loss": 0.2581, + "num_input_tokens_seen": 27583688, + "step": 36165 + }, + { + "epoch": 75.1975051975052, + "grad_norm": 0.0006480725714936852, + "learning_rate": 0.006738859124684437, + "loss": 0.2923, + "num_input_tokens_seen": 27587592, + "step": 36170 + }, + { + "epoch": 75.20790020790021, + "grad_norm": 0.00019662445993162692, + "learning_rate": 0.006721412740459259, + "loss": 0.269, + "num_input_tokens_seen": 27591464, + "step": 36175 + }, + { + "epoch": 75.21829521829522, + "grad_norm": 0.0014681684551760554, + "learning_rate": 0.006703988451592824, + "loss": 0.271, + "num_input_tokens_seen": 27595304, + "step": 36180 + }, + { + "epoch": 75.22869022869023, + "grad_norm": 0.000167108591995202, + "learning_rate": 0.006686586260772114, + "loss": 0.2595, + "num_input_tokens_seen": 27599112, + "step": 36185 + }, + { + "epoch": 75.23908523908524, + "grad_norm": 6.415171083062887e-05, + "learning_rate": 0.006669206170680819, + "loss": 0.2561, + "num_input_tokens_seen": 27602824, + "step": 36190 + }, + { + "epoch": 75.24948024948024, + "grad_norm": 0.00022838514996692538, + "learning_rate": 0.0066518481839991095, + "loss": 0.2681, + "num_input_tokens_seen": 27606632, + "step": 36195 + }, + { + "epoch": 75.25987525987526, + "grad_norm": 0.00026006813277490437, + "learning_rate": 0.006634512303403861, + "loss": 0.2815, + "num_input_tokens_seen": 27610376, + "step": 36200 + }, + { + "epoch": 75.25987525987526, + "eval_loss": 0.25133857131004333, + "eval_runtime": 13.4173, + "eval_samples_per_second": 63.798, + "eval_steps_per_second": 15.95, + "num_input_tokens_seen": 27610376, + "step": 36200 + }, + { + "epoch": 75.27027027027027, + "grad_norm": 0.00018262698722537607, + "learning_rate": 0.0066171985315684355, + "loss": 0.2776, + "num_input_tokens_seen": 27613992, + "step": 36205 + }, + { + "epoch": 75.28066528066527, + "grad_norm": 0.00017371807189192623, + "learning_rate": 0.0065999068711628806, + "loss": 0.2578, + "num_input_tokens_seen": 27617832, + "step": 36210 + }, + { + "epoch": 75.29106029106029, + "grad_norm": 0.00036904754233546555, + "learning_rate": 0.0065826373248537295, + "loss": 0.2678, + "num_input_tokens_seen": 27621672, + "step": 36215 + }, + { + "epoch": 75.3014553014553, + "grad_norm": 0.00015767717559356242, + "learning_rate": 0.006565389895304218, + "loss": 0.2624, + "num_input_tokens_seen": 27625576, + "step": 36220 + }, + { + "epoch": 75.3118503118503, + "grad_norm": 0.0011044838465750217, + "learning_rate": 0.006548164585174104, + "loss": 0.2579, + "num_input_tokens_seen": 27629544, + "step": 36225 + }, + { + "epoch": 75.32224532224532, + "grad_norm": 0.0007148599252104759, + "learning_rate": 0.006530961397119728, + "loss": 0.2892, + "num_input_tokens_seen": 27633480, + "step": 36230 + }, + { + "epoch": 75.33264033264034, + "grad_norm": 0.0002838864747900516, + "learning_rate": 0.00651378033379405, + "loss": 0.2627, + "num_input_tokens_seen": 27637256, + "step": 36235 + }, + { + "epoch": 75.34303534303534, + "grad_norm": 0.00021961372112855315, + "learning_rate": 0.006496621397846619, + "loss": 0.2529, + "num_input_tokens_seen": 27641000, + "step": 36240 + }, + { + "epoch": 75.35343035343035, + "grad_norm": 0.0006204022793099284, + "learning_rate": 0.006479484591923518, + "loss": 0.2718, + "num_input_tokens_seen": 27644904, + "step": 36245 + }, + { + "epoch": 75.36382536382537, + "grad_norm": 0.00027774160844273865, + "learning_rate": 0.006462369918667515, + "loss": 0.2632, + "num_input_tokens_seen": 27648808, + "step": 36250 + }, + { + "epoch": 75.37422037422037, + "grad_norm": 0.0003736469952855259, + "learning_rate": 0.006445277380717851, + "loss": 0.2736, + "num_input_tokens_seen": 27652552, + "step": 36255 + }, + { + "epoch": 75.38461538461539, + "grad_norm": 0.00017816714535001665, + "learning_rate": 0.006428206980710466, + "loss": 0.2652, + "num_input_tokens_seen": 27656264, + "step": 36260 + }, + { + "epoch": 75.39501039501039, + "grad_norm": 0.0005649793311022222, + "learning_rate": 0.006411158721277788, + "loss": 0.2754, + "num_input_tokens_seen": 27659976, + "step": 36265 + }, + { + "epoch": 75.4054054054054, + "grad_norm": 0.0002725072263274342, + "learning_rate": 0.00639413260504888, + "loss": 0.25, + "num_input_tokens_seen": 27663784, + "step": 36270 + }, + { + "epoch": 75.41580041580042, + "grad_norm": 0.00035435458994470537, + "learning_rate": 0.006377128634649376, + "loss": 0.2588, + "num_input_tokens_seen": 27667560, + "step": 36275 + }, + { + "epoch": 75.42619542619542, + "grad_norm": 0.000293812743620947, + "learning_rate": 0.006360146812701528, + "loss": 0.2655, + "num_input_tokens_seen": 27671176, + "step": 36280 + }, + { + "epoch": 75.43659043659044, + "grad_norm": 0.0002869760792236775, + "learning_rate": 0.006343187141824125, + "loss": 0.2601, + "num_input_tokens_seen": 27675080, + "step": 36285 + }, + { + "epoch": 75.44698544698545, + "grad_norm": 0.00024739388027228415, + "learning_rate": 0.00632624962463259, + "loss": 0.2588, + "num_input_tokens_seen": 27678888, + "step": 36290 + }, + { + "epoch": 75.45738045738045, + "grad_norm": 5.570205394178629e-05, + "learning_rate": 0.006309334263738853, + "loss": 0.251, + "num_input_tokens_seen": 27682760, + "step": 36295 + }, + { + "epoch": 75.46777546777547, + "grad_norm": 0.00031986256362870336, + "learning_rate": 0.006292441061751508, + "loss": 0.2666, + "num_input_tokens_seen": 27686728, + "step": 36300 + }, + { + "epoch": 75.47817047817048, + "grad_norm": 0.0002962035941891372, + "learning_rate": 0.0062755700212757054, + "loss": 0.2541, + "num_input_tokens_seen": 27690568, + "step": 36305 + }, + { + "epoch": 75.48856548856548, + "grad_norm": 0.0005598647985607386, + "learning_rate": 0.006258721144913148, + "loss": 0.2476, + "num_input_tokens_seen": 27694344, + "step": 36310 + }, + { + "epoch": 75.4989604989605, + "grad_norm": 0.00040435505798086524, + "learning_rate": 0.0062418944352621575, + "loss": 0.2528, + "num_input_tokens_seen": 27698280, + "step": 36315 + }, + { + "epoch": 75.50935550935552, + "grad_norm": 0.0006754234782420099, + "learning_rate": 0.0062250898949176405, + "loss": 0.2559, + "num_input_tokens_seen": 27702152, + "step": 36320 + }, + { + "epoch": 75.51975051975052, + "grad_norm": 7.313093374250457e-05, + "learning_rate": 0.006208307526471041, + "loss": 0.2943, + "num_input_tokens_seen": 27706056, + "step": 36325 + }, + { + "epoch": 75.53014553014553, + "grad_norm": 0.00019297453400213271, + "learning_rate": 0.006191547332510405, + "loss": 0.2818, + "num_input_tokens_seen": 27709896, + "step": 36330 + }, + { + "epoch": 75.54054054054055, + "grad_norm": 0.0007400016766041517, + "learning_rate": 0.006174809315620416, + "loss": 0.2651, + "num_input_tokens_seen": 27713864, + "step": 36335 + }, + { + "epoch": 75.55093555093555, + "grad_norm": 0.0002630755479913205, + "learning_rate": 0.00615809347838221, + "loss": 0.2815, + "num_input_tokens_seen": 27717768, + "step": 36340 + }, + { + "epoch": 75.56133056133056, + "grad_norm": 0.00019092790898866951, + "learning_rate": 0.006141399823373655, + "loss": 0.2672, + "num_input_tokens_seen": 27721544, + "step": 36345 + }, + { + "epoch": 75.57172557172557, + "grad_norm": 0.00023187852639239281, + "learning_rate": 0.0061247283531690455, + "loss": 0.2683, + "num_input_tokens_seen": 27725160, + "step": 36350 + }, + { + "epoch": 75.58212058212058, + "grad_norm": 0.00012423927546478808, + "learning_rate": 0.0061080790703393895, + "loss": 0.2723, + "num_input_tokens_seen": 27728968, + "step": 36355 + }, + { + "epoch": 75.5925155925156, + "grad_norm": 0.00019474186410661787, + "learning_rate": 0.006091451977452217, + "loss": 0.2723, + "num_input_tokens_seen": 27732840, + "step": 36360 + }, + { + "epoch": 75.6029106029106, + "grad_norm": 0.0007714293315075338, + "learning_rate": 0.00607484707707161, + "loss": 0.2618, + "num_input_tokens_seen": 27736520, + "step": 36365 + }, + { + "epoch": 75.61330561330561, + "grad_norm": 0.00013165008567739278, + "learning_rate": 0.006058264371758254, + "loss": 0.2649, + "num_input_tokens_seen": 27740456, + "step": 36370 + }, + { + "epoch": 75.62370062370063, + "grad_norm": 0.0002822806127369404, + "learning_rate": 0.00604170386406942, + "loss": 0.2546, + "num_input_tokens_seen": 27744232, + "step": 36375 + }, + { + "epoch": 75.63409563409563, + "grad_norm": 0.00011422712850617245, + "learning_rate": 0.006025165556558931, + "loss": 0.2567, + "num_input_tokens_seen": 27748072, + "step": 36380 + }, + { + "epoch": 75.64449064449065, + "grad_norm": 0.00016472676361445338, + "learning_rate": 0.006008649451777248, + "loss": 0.2568, + "num_input_tokens_seen": 27751816, + "step": 36385 + }, + { + "epoch": 75.65488565488566, + "grad_norm": 0.0005685825599357486, + "learning_rate": 0.005992155552271283, + "loss": 0.2592, + "num_input_tokens_seen": 27755464, + "step": 36390 + }, + { + "epoch": 75.66528066528066, + "grad_norm": 0.00030456818058155477, + "learning_rate": 0.005975683860584685, + "loss": 0.2653, + "num_input_tokens_seen": 27759272, + "step": 36395 + }, + { + "epoch": 75.67567567567568, + "grad_norm": 0.00042223164928145707, + "learning_rate": 0.0059592343792575385, + "loss": 0.2609, + "num_input_tokens_seen": 27762984, + "step": 36400 + }, + { + "epoch": 75.67567567567568, + "eval_loss": 0.24879048764705658, + "eval_runtime": 13.3964, + "eval_samples_per_second": 63.898, + "eval_steps_per_second": 15.974, + "num_input_tokens_seen": 27762984, + "step": 36400 + }, + { + "epoch": 75.68607068607069, + "grad_norm": 0.0002706471714191139, + "learning_rate": 0.0059428071108265975, + "loss": 0.2297, + "num_input_tokens_seen": 27766728, + "step": 36405 + }, + { + "epoch": 75.6964656964657, + "grad_norm": 0.00012597993190865964, + "learning_rate": 0.005926402057825136, + "loss": 0.2712, + "num_input_tokens_seen": 27770568, + "step": 36410 + }, + { + "epoch": 75.70686070686071, + "grad_norm": 0.0002018495579250157, + "learning_rate": 0.005910019222782997, + "loss": 0.2606, + "num_input_tokens_seen": 27774344, + "step": 36415 + }, + { + "epoch": 75.71725571725571, + "grad_norm": 0.0010055502643808722, + "learning_rate": 0.005893658608226643, + "loss": 0.2451, + "num_input_tokens_seen": 27778216, + "step": 36420 + }, + { + "epoch": 75.72765072765073, + "grad_norm": 0.0005572371883317828, + "learning_rate": 0.0058773202166791045, + "loss": 0.2745, + "num_input_tokens_seen": 27782056, + "step": 36425 + }, + { + "epoch": 75.73804573804574, + "grad_norm": 0.00024897028924897313, + "learning_rate": 0.005861004050659918, + "loss": 0.2695, + "num_input_tokens_seen": 27785992, + "step": 36430 + }, + { + "epoch": 75.74844074844074, + "grad_norm": 0.00016537195187993348, + "learning_rate": 0.005844710112685286, + "loss": 0.2418, + "num_input_tokens_seen": 27789800, + "step": 36435 + }, + { + "epoch": 75.75883575883576, + "grad_norm": 0.00029151991475373507, + "learning_rate": 0.005828438405267933, + "loss": 0.2698, + "num_input_tokens_seen": 27793544, + "step": 36440 + }, + { + "epoch": 75.76923076923077, + "grad_norm": 0.00041944990516640246, + "learning_rate": 0.00581218893091715, + "loss": 0.2549, + "num_input_tokens_seen": 27797416, + "step": 36445 + }, + { + "epoch": 75.77962577962577, + "grad_norm": 0.0006753249326720834, + "learning_rate": 0.005795961692138801, + "loss": 0.2765, + "num_input_tokens_seen": 27801416, + "step": 36450 + }, + { + "epoch": 75.79002079002079, + "grad_norm": 9.621229401091114e-05, + "learning_rate": 0.00577975669143535, + "loss": 0.2506, + "num_input_tokens_seen": 27805288, + "step": 36455 + }, + { + "epoch": 75.8004158004158, + "grad_norm": 0.0003091284306719899, + "learning_rate": 0.005763573931305782, + "loss": 0.2711, + "num_input_tokens_seen": 27809096, + "step": 36460 + }, + { + "epoch": 75.8108108108108, + "grad_norm": 0.00020964314171578735, + "learning_rate": 0.005747413414245733, + "loss": 0.2481, + "num_input_tokens_seen": 27812776, + "step": 36465 + }, + { + "epoch": 75.82120582120582, + "grad_norm": 0.0003540139296092093, + "learning_rate": 0.005731275142747294, + "loss": 0.2791, + "num_input_tokens_seen": 27816552, + "step": 36470 + }, + { + "epoch": 75.83160083160084, + "grad_norm": 0.0004429105611052364, + "learning_rate": 0.005715159119299256, + "loss": 0.2436, + "num_input_tokens_seen": 27820488, + "step": 36475 + }, + { + "epoch": 75.84199584199584, + "grad_norm": 0.00015174214786384255, + "learning_rate": 0.005699065346386867, + "loss": 0.2398, + "num_input_tokens_seen": 27824328, + "step": 36480 + }, + { + "epoch": 75.85239085239085, + "grad_norm": 0.0002160083968192339, + "learning_rate": 0.0056829938264919885, + "loss": 0.2443, + "num_input_tokens_seen": 27828008, + "step": 36485 + }, + { + "epoch": 75.86278586278586, + "grad_norm": 0.00024172312987502664, + "learning_rate": 0.005666944562093074, + "loss": 0.252, + "num_input_tokens_seen": 27831976, + "step": 36490 + }, + { + "epoch": 75.87318087318087, + "grad_norm": 6.766623846488073e-05, + "learning_rate": 0.005650917555665108, + "loss": 0.2602, + "num_input_tokens_seen": 27835656, + "step": 36495 + }, + { + "epoch": 75.88357588357589, + "grad_norm": 0.00022125743271317333, + "learning_rate": 0.005634912809679632, + "loss": 0.2396, + "num_input_tokens_seen": 27839304, + "step": 36500 + }, + { + "epoch": 75.89397089397089, + "grad_norm": 0.00020492971816565841, + "learning_rate": 0.005618930326604854, + "loss": 0.2702, + "num_input_tokens_seen": 27843112, + "step": 36505 + }, + { + "epoch": 75.9043659043659, + "grad_norm": 0.0004243365256115794, + "learning_rate": 0.005602970108905386, + "loss": 0.2401, + "num_input_tokens_seen": 27846760, + "step": 36510 + }, + { + "epoch": 75.91476091476092, + "grad_norm": 0.00023723937920294702, + "learning_rate": 0.005587032159042543, + "loss": 0.2826, + "num_input_tokens_seen": 27850568, + "step": 36515 + }, + { + "epoch": 75.92515592515592, + "grad_norm": 0.00017458625370636582, + "learning_rate": 0.005571116479474158, + "loss": 0.2568, + "num_input_tokens_seen": 27854280, + "step": 36520 + }, + { + "epoch": 75.93555093555094, + "grad_norm": 8.346794493263587e-05, + "learning_rate": 0.005555223072654619, + "loss": 0.2406, + "num_input_tokens_seen": 27858184, + "step": 36525 + }, + { + "epoch": 75.94594594594595, + "grad_norm": 0.00020504332496784627, + "learning_rate": 0.005539351941034881, + "loss": 0.2794, + "num_input_tokens_seen": 27861928, + "step": 36530 + }, + { + "epoch": 75.95634095634095, + "grad_norm": 0.0004236344830133021, + "learning_rate": 0.0055235030870624865, + "loss": 0.2405, + "num_input_tokens_seen": 27865864, + "step": 36535 + }, + { + "epoch": 75.96673596673597, + "grad_norm": 0.0004484095552470535, + "learning_rate": 0.005507676513181514, + "loss": 0.2711, + "num_input_tokens_seen": 27869768, + "step": 36540 + }, + { + "epoch": 75.97713097713098, + "grad_norm": 0.0002650415408425033, + "learning_rate": 0.005491872221832628, + "loss": 0.2589, + "num_input_tokens_seen": 27873608, + "step": 36545 + }, + { + "epoch": 75.98752598752598, + "grad_norm": 0.00033283577067777514, + "learning_rate": 0.005476090215453061, + "loss": 0.2771, + "num_input_tokens_seen": 27877352, + "step": 36550 + }, + { + "epoch": 75.997920997921, + "grad_norm": 0.0003396625688765198, + "learning_rate": 0.0054603304964765675, + "loss": 0.2373, + "num_input_tokens_seen": 27881128, + "step": 36555 + }, + { + "epoch": 76.00831600831602, + "grad_norm": 0.00016951891302596778, + "learning_rate": 0.005444593067333519, + "loss": 0.2989, + "num_input_tokens_seen": 27884944, + "step": 36560 + }, + { + "epoch": 76.01871101871102, + "grad_norm": 0.0002527138276491314, + "learning_rate": 0.00542887793045081, + "loss": 0.2461, + "num_input_tokens_seen": 27888784, + "step": 36565 + }, + { + "epoch": 76.02910602910603, + "grad_norm": 0.0002157313865609467, + "learning_rate": 0.005413185088251932, + "loss": 0.2476, + "num_input_tokens_seen": 27892752, + "step": 36570 + }, + { + "epoch": 76.03950103950103, + "grad_norm": 0.0003957070875912905, + "learning_rate": 0.005397514543156884, + "loss": 0.249, + "num_input_tokens_seen": 27896496, + "step": 36575 + }, + { + "epoch": 76.04989604989605, + "grad_norm": 9.086212230613455e-05, + "learning_rate": 0.0053818662975822825, + "loss": 0.2527, + "num_input_tokens_seen": 27900208, + "step": 36580 + }, + { + "epoch": 76.06029106029106, + "grad_norm": 0.00037114444421604276, + "learning_rate": 0.005366240353941315, + "loss": 0.2681, + "num_input_tokens_seen": 27903952, + "step": 36585 + }, + { + "epoch": 76.07068607068607, + "grad_norm": 0.0007193362107500434, + "learning_rate": 0.005350636714643636, + "loss": 0.2842, + "num_input_tokens_seen": 27907760, + "step": 36590 + }, + { + "epoch": 76.08108108108108, + "grad_norm": 0.0010902669746428728, + "learning_rate": 0.005335055382095555, + "loss": 0.2543, + "num_input_tokens_seen": 27911568, + "step": 36595 + }, + { + "epoch": 76.0914760914761, + "grad_norm": 9.507292998023331e-05, + "learning_rate": 0.005319496358699915, + "loss": 0.2621, + "num_input_tokens_seen": 27915504, + "step": 36600 + }, + { + "epoch": 76.0914760914761, + "eval_loss": 0.2486761212348938, + "eval_runtime": 13.3627, + "eval_samples_per_second": 64.059, + "eval_steps_per_second": 16.015, + "num_input_tokens_seen": 27915504, + "step": 36600 + }, + { + "epoch": 76.1018711018711, + "grad_norm": 0.000569100200664252, + "learning_rate": 0.005303959646856099, + "loss": 0.2455, + "num_input_tokens_seen": 27919312, + "step": 36605 + }, + { + "epoch": 76.11226611226611, + "grad_norm": 0.0001948667486431077, + "learning_rate": 0.005288445248960089, + "loss": 0.2289, + "num_input_tokens_seen": 27923152, + "step": 36610 + }, + { + "epoch": 76.12266112266113, + "grad_norm": 0.0006906175403855741, + "learning_rate": 0.005272953167404354, + "loss": 0.2763, + "num_input_tokens_seen": 27926960, + "step": 36615 + }, + { + "epoch": 76.13305613305613, + "grad_norm": 0.0004480722709558904, + "learning_rate": 0.005257483404578017, + "loss": 0.2694, + "num_input_tokens_seen": 27930960, + "step": 36620 + }, + { + "epoch": 76.14345114345114, + "grad_norm": 8.20136847323738e-05, + "learning_rate": 0.0052420359628666865, + "loss": 0.2696, + "num_input_tokens_seen": 27934896, + "step": 36625 + }, + { + "epoch": 76.15384615384616, + "grad_norm": 0.00019306126341689378, + "learning_rate": 0.00522661084465254, + "loss": 0.2547, + "num_input_tokens_seen": 27938672, + "step": 36630 + }, + { + "epoch": 76.16424116424116, + "grad_norm": 0.0002487867895979434, + "learning_rate": 0.005211208052314326, + "loss": 0.2666, + "num_input_tokens_seen": 27942512, + "step": 36635 + }, + { + "epoch": 76.17463617463618, + "grad_norm": 0.0005444425041787326, + "learning_rate": 0.005195827588227391, + "loss": 0.2573, + "num_input_tokens_seen": 27946448, + "step": 36640 + }, + { + "epoch": 76.18503118503118, + "grad_norm": 0.0005136593244969845, + "learning_rate": 0.0051804694547635255, + "loss": 0.2542, + "num_input_tokens_seen": 27950192, + "step": 36645 + }, + { + "epoch": 76.1954261954262, + "grad_norm": 0.00027354041230864823, + "learning_rate": 0.005165133654291232, + "loss": 0.2826, + "num_input_tokens_seen": 27953872, + "step": 36650 + }, + { + "epoch": 76.20582120582121, + "grad_norm": 0.0004595712525770068, + "learning_rate": 0.005149820189175402, + "loss": 0.2903, + "num_input_tokens_seen": 27957584, + "step": 36655 + }, + { + "epoch": 76.21621621621621, + "grad_norm": 0.00021985406056046486, + "learning_rate": 0.005134529061777598, + "loss": 0.2653, + "num_input_tokens_seen": 27961424, + "step": 36660 + }, + { + "epoch": 76.22661122661123, + "grad_norm": 0.00019717963004950434, + "learning_rate": 0.005119260274455933, + "loss": 0.2714, + "num_input_tokens_seen": 27965264, + "step": 36665 + }, + { + "epoch": 76.23700623700624, + "grad_norm": 0.00010775313421618193, + "learning_rate": 0.005104013829565007, + "loss": 0.2554, + "num_input_tokens_seen": 27969104, + "step": 36670 + }, + { + "epoch": 76.24740124740124, + "grad_norm": 0.00012310956662986428, + "learning_rate": 0.005088789729456006, + "loss": 0.276, + "num_input_tokens_seen": 27973040, + "step": 36675 + }, + { + "epoch": 76.25779625779626, + "grad_norm": 0.0009582063648849726, + "learning_rate": 0.005073587976476735, + "loss": 0.2711, + "num_input_tokens_seen": 27976848, + "step": 36680 + }, + { + "epoch": 76.26819126819127, + "grad_norm": 0.00040995379094965756, + "learning_rate": 0.005058408572971418, + "loss": 0.2473, + "num_input_tokens_seen": 27980688, + "step": 36685 + }, + { + "epoch": 76.27858627858627, + "grad_norm": 0.00039800367085263133, + "learning_rate": 0.005043251521280983, + "loss": 0.2542, + "num_input_tokens_seen": 27984400, + "step": 36690 + }, + { + "epoch": 76.28898128898129, + "grad_norm": 0.000407436047680676, + "learning_rate": 0.005028116823742795, + "loss": 0.2267, + "num_input_tokens_seen": 27988080, + "step": 36695 + }, + { + "epoch": 76.2993762993763, + "grad_norm": 0.00042980213765986264, + "learning_rate": 0.005013004482690819, + "loss": 0.2489, + "num_input_tokens_seen": 27991856, + "step": 36700 + }, + { + "epoch": 76.3097713097713, + "grad_norm": 0.00021001654386054724, + "learning_rate": 0.0049979145004555746, + "loss": 0.2494, + "num_input_tokens_seen": 27995696, + "step": 36705 + }, + { + "epoch": 76.32016632016632, + "grad_norm": 0.00014379557978827506, + "learning_rate": 0.004982846879364116, + "loss": 0.2516, + "num_input_tokens_seen": 27999344, + "step": 36710 + }, + { + "epoch": 76.33056133056132, + "grad_norm": 0.0003672557068057358, + "learning_rate": 0.0049678016217400535, + "loss": 0.2348, + "num_input_tokens_seen": 28003056, + "step": 36715 + }, + { + "epoch": 76.34095634095634, + "grad_norm": 0.00014944524446036667, + "learning_rate": 0.004952778729903595, + "loss": 0.27, + "num_input_tokens_seen": 28006928, + "step": 36720 + }, + { + "epoch": 76.35135135135135, + "grad_norm": 9.978645539376885e-05, + "learning_rate": 0.004937778206171422, + "loss": 0.2554, + "num_input_tokens_seen": 28010768, + "step": 36725 + }, + { + "epoch": 76.36174636174636, + "grad_norm": 0.0001249907654710114, + "learning_rate": 0.004922800052856835, + "loss": 0.2548, + "num_input_tokens_seen": 28014576, + "step": 36730 + }, + { + "epoch": 76.37214137214137, + "grad_norm": 0.0001946751435752958, + "learning_rate": 0.004907844272269602, + "loss": 0.2485, + "num_input_tokens_seen": 28018448, + "step": 36735 + }, + { + "epoch": 76.38253638253639, + "grad_norm": 0.0003651128790806979, + "learning_rate": 0.004892910866716144, + "loss": 0.2482, + "num_input_tokens_seen": 28022192, + "step": 36740 + }, + { + "epoch": 76.39293139293139, + "grad_norm": 0.0003809960908256471, + "learning_rate": 0.004877999838499369, + "loss": 0.2512, + "num_input_tokens_seen": 28025968, + "step": 36745 + }, + { + "epoch": 76.4033264033264, + "grad_norm": 0.0003875236143358052, + "learning_rate": 0.0048631111899187065, + "loss": 0.294, + "num_input_tokens_seen": 28029808, + "step": 36750 + }, + { + "epoch": 76.41372141372142, + "grad_norm": 0.0003525142674334347, + "learning_rate": 0.0048482449232702335, + "loss": 0.254, + "num_input_tokens_seen": 28033488, + "step": 36755 + }, + { + "epoch": 76.42411642411642, + "grad_norm": 0.0003585628292057663, + "learning_rate": 0.004833401040846469, + "loss": 0.2684, + "num_input_tokens_seen": 28037264, + "step": 36760 + }, + { + "epoch": 76.43451143451144, + "grad_norm": 0.00010031870624516159, + "learning_rate": 0.004818579544936546, + "loss": 0.284, + "num_input_tokens_seen": 28041232, + "step": 36765 + }, + { + "epoch": 76.44490644490645, + "grad_norm": 0.0005903986748307943, + "learning_rate": 0.004803780437826121, + "loss": 0.2611, + "num_input_tokens_seen": 28045200, + "step": 36770 + }, + { + "epoch": 76.45530145530145, + "grad_norm": 8.834667096380144e-05, + "learning_rate": 0.004789003721797402, + "loss": 0.2575, + "num_input_tokens_seen": 28049232, + "step": 36775 + }, + { + "epoch": 76.46569646569647, + "grad_norm": 0.00025207214639522135, + "learning_rate": 0.004774249399129132, + "loss": 0.26, + "num_input_tokens_seen": 28053008, + "step": 36780 + }, + { + "epoch": 76.47609147609148, + "grad_norm": 0.0005553350783884525, + "learning_rate": 0.004759517472096642, + "loss": 0.2383, + "num_input_tokens_seen": 28056944, + "step": 36785 + }, + { + "epoch": 76.48648648648648, + "grad_norm": 0.00025554600870236754, + "learning_rate": 0.004744807942971746, + "loss": 0.2795, + "num_input_tokens_seen": 28060688, + "step": 36790 + }, + { + "epoch": 76.4968814968815, + "grad_norm": 0.0004216076049488038, + "learning_rate": 0.004730120814022881, + "loss": 0.2541, + "num_input_tokens_seen": 28064496, + "step": 36795 + }, + { + "epoch": 76.5072765072765, + "grad_norm": 0.0004004836082458496, + "learning_rate": 0.004715456087514935, + "loss": 0.2488, + "num_input_tokens_seen": 28068432, + "step": 36800 + }, + { + "epoch": 76.5072765072765, + "eval_loss": 0.24843348562717438, + "eval_runtime": 13.3776, + "eval_samples_per_second": 63.987, + "eval_steps_per_second": 15.997, + "num_input_tokens_seen": 28068432, + "step": 36800 + }, + { + "epoch": 76.51767151767152, + "grad_norm": 0.0002491026825737208, + "learning_rate": 0.004700813765709432, + "loss": 0.2713, + "num_input_tokens_seen": 28072144, + "step": 36805 + }, + { + "epoch": 76.52806652806653, + "grad_norm": 0.00027622279594652355, + "learning_rate": 0.004686193850864401, + "loss": 0.2463, + "num_input_tokens_seen": 28075856, + "step": 36810 + }, + { + "epoch": 76.53846153846153, + "grad_norm": 0.00043762658606283367, + "learning_rate": 0.004671596345234385, + "loss": 0.263, + "num_input_tokens_seen": 28079600, + "step": 36815 + }, + { + "epoch": 76.54885654885655, + "grad_norm": 0.00012006593169644475, + "learning_rate": 0.00465702125107052, + "loss": 0.2516, + "num_input_tokens_seen": 28083472, + "step": 36820 + }, + { + "epoch": 76.55925155925156, + "grad_norm": 0.00032866952824406326, + "learning_rate": 0.004642468570620506, + "loss": 0.262, + "num_input_tokens_seen": 28087216, + "step": 36825 + }, + { + "epoch": 76.56964656964657, + "grad_norm": 0.00034902579500339925, + "learning_rate": 0.004627938306128482, + "loss": 0.2713, + "num_input_tokens_seen": 28090992, + "step": 36830 + }, + { + "epoch": 76.58004158004158, + "grad_norm": 0.00010219719115411863, + "learning_rate": 0.004613430459835255, + "loss": 0.2569, + "num_input_tokens_seen": 28094640, + "step": 36835 + }, + { + "epoch": 76.5904365904366, + "grad_norm": 0.00034654364571906626, + "learning_rate": 0.004598945033978085, + "loss": 0.2772, + "num_input_tokens_seen": 28098544, + "step": 36840 + }, + { + "epoch": 76.6008316008316, + "grad_norm": 0.00037653459003195167, + "learning_rate": 0.004584482030790804, + "loss": 0.2788, + "num_input_tokens_seen": 28102224, + "step": 36845 + }, + { + "epoch": 76.61122661122661, + "grad_norm": 0.0005071199266240001, + "learning_rate": 0.004570041452503826, + "loss": 0.2719, + "num_input_tokens_seen": 28105872, + "step": 36850 + }, + { + "epoch": 76.62162162162163, + "grad_norm": 0.0001203383071697317, + "learning_rate": 0.004555623301344003, + "loss": 0.2493, + "num_input_tokens_seen": 28109648, + "step": 36855 + }, + { + "epoch": 76.63201663201663, + "grad_norm": 0.0002816346532199532, + "learning_rate": 0.004541227579534857, + "loss": 0.2732, + "num_input_tokens_seen": 28113456, + "step": 36860 + }, + { + "epoch": 76.64241164241164, + "grad_norm": 0.00053248624317348, + "learning_rate": 0.004526854289296378, + "loss": 0.2771, + "num_input_tokens_seen": 28117008, + "step": 36865 + }, + { + "epoch": 76.65280665280665, + "grad_norm": 0.0005299804033711553, + "learning_rate": 0.004512503432845078, + "loss": 0.2663, + "num_input_tokens_seen": 28120816, + "step": 36870 + }, + { + "epoch": 76.66320166320166, + "grad_norm": 4.515681939665228e-05, + "learning_rate": 0.004498175012394068, + "loss": 0.2712, + "num_input_tokens_seen": 28124688, + "step": 36875 + }, + { + "epoch": 76.67359667359668, + "grad_norm": 0.000797775574028492, + "learning_rate": 0.004483869030152965, + "loss": 0.2637, + "num_input_tokens_seen": 28128528, + "step": 36880 + }, + { + "epoch": 76.68399168399168, + "grad_norm": 0.0002325286332052201, + "learning_rate": 0.004469585488327904, + "loss": 0.2837, + "num_input_tokens_seen": 28132240, + "step": 36885 + }, + { + "epoch": 76.6943866943867, + "grad_norm": 0.0002579780702944845, + "learning_rate": 0.0044553243891216395, + "loss": 0.2657, + "num_input_tokens_seen": 28136112, + "step": 36890 + }, + { + "epoch": 76.70478170478171, + "grad_norm": 0.0003651200095191598, + "learning_rate": 0.004441085734733363, + "loss": 0.265, + "num_input_tokens_seen": 28139824, + "step": 36895 + }, + { + "epoch": 76.71517671517671, + "grad_norm": 0.00021340936655178666, + "learning_rate": 0.004426869527358884, + "loss": 0.2681, + "num_input_tokens_seen": 28143600, + "step": 36900 + }, + { + "epoch": 76.72557172557173, + "grad_norm": 0.0002148688945453614, + "learning_rate": 0.0044126757691905156, + "loss": 0.2748, + "num_input_tokens_seen": 28147472, + "step": 36905 + }, + { + "epoch": 76.73596673596674, + "grad_norm": 0.0008631636155769229, + "learning_rate": 0.004398504462417107, + "loss": 0.2535, + "num_input_tokens_seen": 28151280, + "step": 36910 + }, + { + "epoch": 76.74636174636174, + "grad_norm": 0.00011504167196108028, + "learning_rate": 0.0043843556092240605, + "loss": 0.2905, + "num_input_tokens_seen": 28155280, + "step": 36915 + }, + { + "epoch": 76.75675675675676, + "grad_norm": 0.0010818240698426962, + "learning_rate": 0.004370229211793281, + "loss": 0.2575, + "num_input_tokens_seen": 28159088, + "step": 36920 + }, + { + "epoch": 76.76715176715177, + "grad_norm": 0.0008599944994784892, + "learning_rate": 0.0043561252723032405, + "loss": 0.256, + "num_input_tokens_seen": 28162736, + "step": 36925 + }, + { + "epoch": 76.77754677754677, + "grad_norm": 0.0003508412337396294, + "learning_rate": 0.004342043792929001, + "loss": 0.2484, + "num_input_tokens_seen": 28166640, + "step": 36930 + }, + { + "epoch": 76.78794178794179, + "grad_norm": 0.0007711535436101258, + "learning_rate": 0.004327984775842025, + "loss": 0.2775, + "num_input_tokens_seen": 28170576, + "step": 36935 + }, + { + "epoch": 76.7983367983368, + "grad_norm": 0.00013414390559773892, + "learning_rate": 0.004313948223210428, + "loss": 0.2581, + "num_input_tokens_seen": 28174544, + "step": 36940 + }, + { + "epoch": 76.8087318087318, + "grad_norm": 0.00011797674233093858, + "learning_rate": 0.004299934137198846, + "loss": 0.2716, + "num_input_tokens_seen": 28178384, + "step": 36945 + }, + { + "epoch": 76.81912681912682, + "grad_norm": 0.00018428413022775203, + "learning_rate": 0.004285942519968383, + "loss": 0.2682, + "num_input_tokens_seen": 28182192, + "step": 36950 + }, + { + "epoch": 76.82952182952182, + "grad_norm": 0.0003074467822443694, + "learning_rate": 0.004271973373676746, + "loss": 0.2634, + "num_input_tokens_seen": 28186096, + "step": 36955 + }, + { + "epoch": 76.83991683991684, + "grad_norm": 0.00016882973432075232, + "learning_rate": 0.004258026700478146, + "loss": 0.2518, + "num_input_tokens_seen": 28189904, + "step": 36960 + }, + { + "epoch": 76.85031185031185, + "grad_norm": 0.00015716590860392898, + "learning_rate": 0.004244102502523328, + "loss": 0.274, + "num_input_tokens_seen": 28193712, + "step": 36965 + }, + { + "epoch": 76.86070686070686, + "grad_norm": 0.0004181308322586119, + "learning_rate": 0.004230200781959592, + "loss": 0.2696, + "num_input_tokens_seen": 28197584, + "step": 36970 + }, + { + "epoch": 76.87110187110187, + "grad_norm": 0.0003224349347874522, + "learning_rate": 0.004216321540930756, + "loss": 0.2684, + "num_input_tokens_seen": 28201456, + "step": 36975 + }, + { + "epoch": 76.88149688149689, + "grad_norm": 0.00027789699379354715, + "learning_rate": 0.004202464781577175, + "loss": 0.2581, + "num_input_tokens_seen": 28205360, + "step": 36980 + }, + { + "epoch": 76.89189189189189, + "grad_norm": 0.0003180943022016436, + "learning_rate": 0.00418863050603574, + "loss": 0.2567, + "num_input_tokens_seen": 28209264, + "step": 36985 + }, + { + "epoch": 76.9022869022869, + "grad_norm": 0.00042342691449448466, + "learning_rate": 0.004174818716439843, + "loss": 0.228, + "num_input_tokens_seen": 28213072, + "step": 36990 + }, + { + "epoch": 76.91268191268192, + "grad_norm": 0.00031388431671075523, + "learning_rate": 0.004161029414919464, + "loss": 0.2759, + "num_input_tokens_seen": 28216912, + "step": 36995 + }, + { + "epoch": 76.92307692307692, + "grad_norm": 0.00030096486443653703, + "learning_rate": 0.004147262603601071, + "loss": 0.2812, + "num_input_tokens_seen": 28220720, + "step": 37000 + }, + { + "epoch": 76.92307692307692, + "eval_loss": 0.2496126890182495, + "eval_runtime": 13.3892, + "eval_samples_per_second": 63.932, + "eval_steps_per_second": 15.983, + "num_input_tokens_seen": 28220720, + "step": 37000 + }, + { + "epoch": 76.93347193347194, + "grad_norm": 0.00011489955795696005, + "learning_rate": 0.004133518284607679, + "loss": 0.2784, + "num_input_tokens_seen": 28224624, + "step": 37005 + }, + { + "epoch": 76.94386694386695, + "grad_norm": 0.0002891292970161885, + "learning_rate": 0.004119796460058861, + "loss": 0.2608, + "num_input_tokens_seen": 28228496, + "step": 37010 + }, + { + "epoch": 76.95426195426195, + "grad_norm": 0.0001657196698943153, + "learning_rate": 0.00410609713207064, + "loss": 0.2617, + "num_input_tokens_seen": 28232336, + "step": 37015 + }, + { + "epoch": 76.96465696465697, + "grad_norm": 0.0001507043489255011, + "learning_rate": 0.004092420302755678, + "loss": 0.2563, + "num_input_tokens_seen": 28236112, + "step": 37020 + }, + { + "epoch": 76.97505197505197, + "grad_norm": 0.00017236564599443227, + "learning_rate": 0.004078765974223103, + "loss": 0.2594, + "num_input_tokens_seen": 28239856, + "step": 37025 + }, + { + "epoch": 76.98544698544698, + "grad_norm": 0.0004071624716743827, + "learning_rate": 0.004065134148578564, + "loss": 0.2734, + "num_input_tokens_seen": 28243600, + "step": 37030 + }, + { + "epoch": 76.995841995842, + "grad_norm": 0.0004299864231143147, + "learning_rate": 0.004051524827924279, + "loss": 0.2817, + "num_input_tokens_seen": 28247344, + "step": 37035 + }, + { + "epoch": 77.006237006237, + "grad_norm": 0.003630937309935689, + "learning_rate": 0.004037938014358955, + "loss": 0.2382, + "num_input_tokens_seen": 28251072, + "step": 37040 + }, + { + "epoch": 77.01663201663202, + "grad_norm": 0.00022580896620638669, + "learning_rate": 0.004024373709977863, + "loss": 0.2633, + "num_input_tokens_seen": 28255008, + "step": 37045 + }, + { + "epoch": 77.02702702702703, + "grad_norm": 0.0004324744513723999, + "learning_rate": 0.004010831916872814, + "loss": 0.2635, + "num_input_tokens_seen": 28258752, + "step": 37050 + }, + { + "epoch": 77.03742203742203, + "grad_norm": 0.001715255668386817, + "learning_rate": 0.003997312637132089, + "loss": 0.2768, + "num_input_tokens_seen": 28262656, + "step": 37055 + }, + { + "epoch": 77.04781704781705, + "grad_norm": 0.0004185736470390111, + "learning_rate": 0.003983815872840535, + "loss": 0.2486, + "num_input_tokens_seen": 28266688, + "step": 37060 + }, + { + "epoch": 77.05821205821206, + "grad_norm": 0.00022444193018600345, + "learning_rate": 0.003970341626079521, + "loss": 0.2799, + "num_input_tokens_seen": 28270592, + "step": 37065 + }, + { + "epoch": 77.06860706860707, + "grad_norm": 0.00017643517639953643, + "learning_rate": 0.003956889898926952, + "loss": 0.2621, + "num_input_tokens_seen": 28274336, + "step": 37070 + }, + { + "epoch": 77.07900207900208, + "grad_norm": 0.00017763584037311375, + "learning_rate": 0.0039434606934572675, + "loss": 0.255, + "num_input_tokens_seen": 28278432, + "step": 37075 + }, + { + "epoch": 77.0893970893971, + "grad_norm": 0.0003319371317047626, + "learning_rate": 0.003930054011741396, + "loss": 0.2588, + "num_input_tokens_seen": 28282240, + "step": 37080 + }, + { + "epoch": 77.0997920997921, + "grad_norm": 0.0003477666468825191, + "learning_rate": 0.0039166698558468155, + "loss": 0.261, + "num_input_tokens_seen": 28286080, + "step": 37085 + }, + { + "epoch": 77.11018711018711, + "grad_norm": 0.00012088054063497111, + "learning_rate": 0.0039033082278375594, + "loss": 0.2749, + "num_input_tokens_seen": 28289856, + "step": 37090 + }, + { + "epoch": 77.12058212058211, + "grad_norm": 0.0003110474208369851, + "learning_rate": 0.003889969129774112, + "loss": 0.2634, + "num_input_tokens_seen": 28293632, + "step": 37095 + }, + { + "epoch": 77.13097713097713, + "grad_norm": 0.00029557262314483523, + "learning_rate": 0.0038766525637135784, + "loss": 0.2373, + "num_input_tokens_seen": 28297408, + "step": 37100 + }, + { + "epoch": 77.14137214137214, + "grad_norm": 0.00015276744670700282, + "learning_rate": 0.0038633585317095318, + "loss": 0.2638, + "num_input_tokens_seen": 28301312, + "step": 37105 + }, + { + "epoch": 77.15176715176715, + "grad_norm": 0.00018350358004681766, + "learning_rate": 0.00385008703581205, + "loss": 0.2686, + "num_input_tokens_seen": 28305184, + "step": 37110 + }, + { + "epoch": 77.16216216216216, + "grad_norm": 0.000589017232414335, + "learning_rate": 0.0038368380780677944, + "loss": 0.2537, + "num_input_tokens_seen": 28309024, + "step": 37115 + }, + { + "epoch": 77.17255717255718, + "grad_norm": 0.000341988867148757, + "learning_rate": 0.003823611660519882, + "loss": 0.2661, + "num_input_tokens_seen": 28312896, + "step": 37120 + }, + { + "epoch": 77.18295218295218, + "grad_norm": 0.0005191625095903873, + "learning_rate": 0.0038104077852080475, + "loss": 0.2756, + "num_input_tokens_seen": 28316704, + "step": 37125 + }, + { + "epoch": 77.1933471933472, + "grad_norm": 0.0005864623817615211, + "learning_rate": 0.003797226454168462, + "loss": 0.2705, + "num_input_tokens_seen": 28320416, + "step": 37130 + }, + { + "epoch": 77.20374220374221, + "grad_norm": 0.00032458262285217643, + "learning_rate": 0.003784067669433849, + "loss": 0.269, + "num_input_tokens_seen": 28324288, + "step": 37135 + }, + { + "epoch": 77.21413721413721, + "grad_norm": 0.0003057298017665744, + "learning_rate": 0.0037709314330334528, + "loss": 0.2724, + "num_input_tokens_seen": 28327840, + "step": 37140 + }, + { + "epoch": 77.22453222453223, + "grad_norm": 0.00019162641547154635, + "learning_rate": 0.003757817746993086, + "loss": 0.2661, + "num_input_tokens_seen": 28331520, + "step": 37145 + }, + { + "epoch": 77.23492723492724, + "grad_norm": 0.00037404592148959637, + "learning_rate": 0.0037447266133349977, + "loss": 0.2868, + "num_input_tokens_seen": 28335392, + "step": 37150 + }, + { + "epoch": 77.24532224532224, + "grad_norm": 0.0005784329841844738, + "learning_rate": 0.003731658034078039, + "loss": 0.2519, + "num_input_tokens_seen": 28339264, + "step": 37155 + }, + { + "epoch": 77.25571725571726, + "grad_norm": 0.000151947679114528, + "learning_rate": 0.0037186120112375153, + "loss": 0.2671, + "num_input_tokens_seen": 28343136, + "step": 37160 + }, + { + "epoch": 77.26611226611226, + "grad_norm": 5.967227480141446e-05, + "learning_rate": 0.003705588546825317, + "loss": 0.2763, + "num_input_tokens_seen": 28346784, + "step": 37165 + }, + { + "epoch": 77.27650727650727, + "grad_norm": 0.0003354424552526325, + "learning_rate": 0.0036925876428498205, + "loss": 0.2233, + "num_input_tokens_seen": 28350592, + "step": 37170 + }, + { + "epoch": 77.28690228690229, + "grad_norm": 0.00024831629707477987, + "learning_rate": 0.0036796093013159057, + "loss": 0.2538, + "num_input_tokens_seen": 28354272, + "step": 37175 + }, + { + "epoch": 77.29729729729729, + "grad_norm": 0.00013682476128451526, + "learning_rate": 0.0036666535242250217, + "loss": 0.2762, + "num_input_tokens_seen": 28358080, + "step": 37180 + }, + { + "epoch": 77.3076923076923, + "grad_norm": 6.173827569000423e-05, + "learning_rate": 0.003653720313575104, + "loss": 0.2519, + "num_input_tokens_seen": 28362112, + "step": 37185 + }, + { + "epoch": 77.31808731808732, + "grad_norm": 0.00044085155241191387, + "learning_rate": 0.003640809671360623, + "loss": 0.2484, + "num_input_tokens_seen": 28365984, + "step": 37190 + }, + { + "epoch": 77.32848232848232, + "grad_norm": 0.000660480058286339, + "learning_rate": 0.003627921599572553, + "loss": 0.2821, + "num_input_tokens_seen": 28369888, + "step": 37195 + }, + { + "epoch": 77.33887733887734, + "grad_norm": 0.00039032360655255616, + "learning_rate": 0.003615056100198405, + "loss": 0.276, + "num_input_tokens_seen": 28373600, + "step": 37200 + }, + { + "epoch": 77.33887733887734, + "eval_loss": 0.2491244226694107, + "eval_runtime": 13.3625, + "eval_samples_per_second": 64.06, + "eval_steps_per_second": 16.015, + "num_input_tokens_seen": 28373600, + "step": 37200 + }, + { + "epoch": 77.34927234927235, + "grad_norm": 0.00012770255852956325, + "learning_rate": 0.003602213175222174, + "loss": 0.2446, + "num_input_tokens_seen": 28377312, + "step": 37205 + }, + { + "epoch": 77.35966735966736, + "grad_norm": 0.00030073378002271056, + "learning_rate": 0.0035893928266244432, + "loss": 0.2857, + "num_input_tokens_seen": 28381152, + "step": 37210 + }, + { + "epoch": 77.37006237006237, + "grad_norm": 0.0002108486951328814, + "learning_rate": 0.003576595056382248, + "loss": 0.2736, + "num_input_tokens_seen": 28384896, + "step": 37215 + }, + { + "epoch": 77.38045738045739, + "grad_norm": 4.768271537614055e-05, + "learning_rate": 0.0035638198664691423, + "loss": 0.2516, + "num_input_tokens_seen": 28388544, + "step": 37220 + }, + { + "epoch": 77.39085239085239, + "grad_norm": 0.00032628123881295323, + "learning_rate": 0.003551067258855267, + "loss": 0.222, + "num_input_tokens_seen": 28392352, + "step": 37225 + }, + { + "epoch": 77.4012474012474, + "grad_norm": 0.000332905154209584, + "learning_rate": 0.0035383372355071996, + "loss": 0.272, + "num_input_tokens_seen": 28396288, + "step": 37230 + }, + { + "epoch": 77.41164241164242, + "grad_norm": 0.002037876518443227, + "learning_rate": 0.0035256297983881023, + "loss": 0.2695, + "num_input_tokens_seen": 28400000, + "step": 37235 + }, + { + "epoch": 77.42203742203742, + "grad_norm": 0.00029906872077845037, + "learning_rate": 0.0035129449494575747, + "loss": 0.2663, + "num_input_tokens_seen": 28403872, + "step": 37240 + }, + { + "epoch": 77.43243243243244, + "grad_norm": 0.0013210372999310493, + "learning_rate": 0.0035002826906718187, + "loss": 0.2884, + "num_input_tokens_seen": 28407648, + "step": 37245 + }, + { + "epoch": 77.44282744282744, + "grad_norm": 0.000139789714012295, + "learning_rate": 0.003487643023983522, + "loss": 0.2665, + "num_input_tokens_seen": 28411456, + "step": 37250 + }, + { + "epoch": 77.45322245322245, + "grad_norm": 0.00016201545076910406, + "learning_rate": 0.003475025951341842, + "loss": 0.2514, + "num_input_tokens_seen": 28415328, + "step": 37255 + }, + { + "epoch": 77.46361746361747, + "grad_norm": 0.0002804503310471773, + "learning_rate": 0.00346243147469249, + "loss": 0.2702, + "num_input_tokens_seen": 28419072, + "step": 37260 + }, + { + "epoch": 77.47401247401247, + "grad_norm": 0.0002909105096478015, + "learning_rate": 0.0034498595959777446, + "loss": 0.2557, + "num_input_tokens_seen": 28422912, + "step": 37265 + }, + { + "epoch": 77.48440748440748, + "grad_norm": 0.000797420390881598, + "learning_rate": 0.003437310317136305, + "loss": 0.2684, + "num_input_tokens_seen": 28426880, + "step": 37270 + }, + { + "epoch": 77.4948024948025, + "grad_norm": 0.0002620832237880677, + "learning_rate": 0.0034247836401034236, + "loss": 0.2551, + "num_input_tokens_seen": 28430816, + "step": 37275 + }, + { + "epoch": 77.5051975051975, + "grad_norm": 0.0005849107983522117, + "learning_rate": 0.003412279566810905, + "loss": 0.2596, + "num_input_tokens_seen": 28434688, + "step": 37280 + }, + { + "epoch": 77.51559251559252, + "grad_norm": 0.0003156873572152108, + "learning_rate": 0.00339979809918699, + "loss": 0.2698, + "num_input_tokens_seen": 28438688, + "step": 37285 + }, + { + "epoch": 77.52598752598753, + "grad_norm": 0.00011453476327005774, + "learning_rate": 0.0033873392391565228, + "loss": 0.2679, + "num_input_tokens_seen": 28442272, + "step": 37290 + }, + { + "epoch": 77.53638253638253, + "grad_norm": 0.0007239239057525992, + "learning_rate": 0.003374902988640782, + "loss": 0.2658, + "num_input_tokens_seen": 28446144, + "step": 37295 + }, + { + "epoch": 77.54677754677755, + "grad_norm": 0.00017345913511235267, + "learning_rate": 0.0033624893495576014, + "loss": 0.2696, + "num_input_tokens_seen": 28450048, + "step": 37300 + }, + { + "epoch": 77.55717255717256, + "grad_norm": 0.0003316587535664439, + "learning_rate": 0.0033500983238213323, + "loss": 0.2803, + "num_input_tokens_seen": 28453824, + "step": 37305 + }, + { + "epoch": 77.56756756756756, + "grad_norm": 0.0008936517406255007, + "learning_rate": 0.0033377299133428126, + "loss": 0.2469, + "num_input_tokens_seen": 28457536, + "step": 37310 + }, + { + "epoch": 77.57796257796258, + "grad_norm": 0.0002861677494365722, + "learning_rate": 0.003325384120029434, + "loss": 0.2398, + "num_input_tokens_seen": 28461408, + "step": 37315 + }, + { + "epoch": 77.58835758835758, + "grad_norm": 0.0007366678328253329, + "learning_rate": 0.0033130609457850233, + "loss": 0.2603, + "num_input_tokens_seen": 28465152, + "step": 37320 + }, + { + "epoch": 77.5987525987526, + "grad_norm": 8.969330519903451e-05, + "learning_rate": 0.0033007603925100104, + "loss": 0.2632, + "num_input_tokens_seen": 28469024, + "step": 37325 + }, + { + "epoch": 77.60914760914761, + "grad_norm": 0.00018747115973383188, + "learning_rate": 0.003288482462101294, + "loss": 0.2662, + "num_input_tokens_seen": 28472704, + "step": 37330 + }, + { + "epoch": 77.61954261954261, + "grad_norm": 0.0002932991774287075, + "learning_rate": 0.0032762271564522605, + "loss": 0.2481, + "num_input_tokens_seen": 28476576, + "step": 37335 + }, + { + "epoch": 77.62993762993763, + "grad_norm": 0.00024183867208193988, + "learning_rate": 0.003263994477452864, + "loss": 0.2502, + "num_input_tokens_seen": 28480352, + "step": 37340 + }, + { + "epoch": 77.64033264033264, + "grad_norm": 0.00019381643505766988, + "learning_rate": 0.0032517844269895125, + "loss": 0.2863, + "num_input_tokens_seen": 28484064, + "step": 37345 + }, + { + "epoch": 77.65072765072765, + "grad_norm": 0.000633417977951467, + "learning_rate": 0.0032395970069451496, + "loss": 0.2502, + "num_input_tokens_seen": 28487840, + "step": 37350 + }, + { + "epoch": 77.66112266112266, + "grad_norm": 0.00039819898665882647, + "learning_rate": 0.0032274322191992388, + "loss": 0.2377, + "num_input_tokens_seen": 28491648, + "step": 37355 + }, + { + "epoch": 77.67151767151768, + "grad_norm": 0.0002766170946415514, + "learning_rate": 0.0032152900656277294, + "loss": 0.2651, + "num_input_tokens_seen": 28495488, + "step": 37360 + }, + { + "epoch": 77.68191268191268, + "grad_norm": 7.905794336693361e-05, + "learning_rate": 0.0032031705481030902, + "loss": 0.2577, + "num_input_tokens_seen": 28499264, + "step": 37365 + }, + { + "epoch": 77.6923076923077, + "grad_norm": 0.0003408543416298926, + "learning_rate": 0.0031910736684943428, + "loss": 0.259, + "num_input_tokens_seen": 28503136, + "step": 37370 + }, + { + "epoch": 77.70270270270271, + "grad_norm": 0.00024153085541911423, + "learning_rate": 0.0031789994286669453, + "loss": 0.283, + "num_input_tokens_seen": 28506912, + "step": 37375 + }, + { + "epoch": 77.71309771309771, + "grad_norm": 0.00040425642509944737, + "learning_rate": 0.003166947830482908, + "loss": 0.2536, + "num_input_tokens_seen": 28510880, + "step": 37380 + }, + { + "epoch": 77.72349272349273, + "grad_norm": 0.0010615846840664744, + "learning_rate": 0.003154918875800727, + "loss": 0.2614, + "num_input_tokens_seen": 28514816, + "step": 37385 + }, + { + "epoch": 77.73388773388774, + "grad_norm": 0.00023688064538873732, + "learning_rate": 0.00314291256647542, + "loss": 0.2499, + "num_input_tokens_seen": 28518656, + "step": 37390 + }, + { + "epoch": 77.74428274428274, + "grad_norm": 0.00032043945975601673, + "learning_rate": 0.0031309289043585375, + "loss": 0.2769, + "num_input_tokens_seen": 28522528, + "step": 37395 + }, + { + "epoch": 77.75467775467776, + "grad_norm": 0.0004774409462697804, + "learning_rate": 0.003118967891298069, + "loss": 0.2792, + "num_input_tokens_seen": 28526304, + "step": 37400 + }, + { + "epoch": 77.75467775467776, + "eval_loss": 0.2487938404083252, + "eval_runtime": 13.3666, + "eval_samples_per_second": 64.04, + "eval_steps_per_second": 16.01, + "num_input_tokens_seen": 28526304, + "step": 37400 + }, + { + "epoch": 77.76507276507276, + "grad_norm": 0.00025252869818359613, + "learning_rate": 0.003107029529138572, + "loss": 0.2505, + "num_input_tokens_seen": 28530080, + "step": 37405 + }, + { + "epoch": 77.77546777546777, + "grad_norm": 0.00040116385207511485, + "learning_rate": 0.0030951138197211235, + "loss": 0.2462, + "num_input_tokens_seen": 28533920, + "step": 37410 + }, + { + "epoch": 77.78586278586279, + "grad_norm": 0.0005070054321549833, + "learning_rate": 0.0030832207648832377, + "loss": 0.2601, + "num_input_tokens_seen": 28537760, + "step": 37415 + }, + { + "epoch": 77.79625779625779, + "grad_norm": 0.0002695280418265611, + "learning_rate": 0.0030713503664589635, + "loss": 0.2623, + "num_input_tokens_seen": 28541600, + "step": 37420 + }, + { + "epoch": 77.8066528066528, + "grad_norm": 0.0003062169998884201, + "learning_rate": 0.0030595026262788872, + "loss": 0.2683, + "num_input_tokens_seen": 28545344, + "step": 37425 + }, + { + "epoch": 77.81704781704782, + "grad_norm": 0.00029004839598201215, + "learning_rate": 0.00304767754617008, + "loss": 0.2588, + "num_input_tokens_seen": 28549088, + "step": 37430 + }, + { + "epoch": 77.82744282744282, + "grad_norm": 0.000274400896159932, + "learning_rate": 0.003035875127956117, + "loss": 0.2611, + "num_input_tokens_seen": 28552832, + "step": 37435 + }, + { + "epoch": 77.83783783783784, + "grad_norm": 0.00010281155118718743, + "learning_rate": 0.0030240953734570752, + "loss": 0.2521, + "num_input_tokens_seen": 28556608, + "step": 37440 + }, + { + "epoch": 77.84823284823285, + "grad_norm": 9.546616638544947e-05, + "learning_rate": 0.003012338284489535, + "loss": 0.2651, + "num_input_tokens_seen": 28560512, + "step": 37445 + }, + { + "epoch": 77.85862785862786, + "grad_norm": 0.00038319017039611936, + "learning_rate": 0.0030006038628665964, + "loss": 0.2653, + "num_input_tokens_seen": 28564288, + "step": 37450 + }, + { + "epoch": 77.86902286902287, + "grad_norm": 8.163908933056518e-05, + "learning_rate": 0.002988892110397845, + "loss": 0.2765, + "num_input_tokens_seen": 28568032, + "step": 37455 + }, + { + "epoch": 77.87941787941789, + "grad_norm": 0.00032411047141067684, + "learning_rate": 0.0029772030288894025, + "loss": 0.2961, + "num_input_tokens_seen": 28571872, + "step": 37460 + }, + { + "epoch": 77.88981288981289, + "grad_norm": 0.0011404320830479264, + "learning_rate": 0.0029655366201438438, + "loss": 0.2441, + "num_input_tokens_seen": 28575584, + "step": 37465 + }, + { + "epoch": 77.9002079002079, + "grad_norm": 0.00013742709415964782, + "learning_rate": 0.0029538928859602965, + "loss": 0.2816, + "num_input_tokens_seen": 28579296, + "step": 37470 + }, + { + "epoch": 77.9106029106029, + "grad_norm": 0.0002540506247896701, + "learning_rate": 0.002942271828134374, + "loss": 0.2601, + "num_input_tokens_seen": 28583072, + "step": 37475 + }, + { + "epoch": 77.92099792099792, + "grad_norm": 0.00022144940157886595, + "learning_rate": 0.00293067344845816, + "loss": 0.2602, + "num_input_tokens_seen": 28586784, + "step": 37480 + }, + { + "epoch": 77.93139293139293, + "grad_norm": 0.0003978123713750392, + "learning_rate": 0.0029190977487202896, + "loss": 0.273, + "num_input_tokens_seen": 28590592, + "step": 37485 + }, + { + "epoch": 77.94178794178794, + "grad_norm": 0.00014536465459968895, + "learning_rate": 0.0029075447307058853, + "loss": 0.2623, + "num_input_tokens_seen": 28594464, + "step": 37490 + }, + { + "epoch": 77.95218295218295, + "grad_norm": 0.0004084158572368324, + "learning_rate": 0.0028960143961965722, + "loss": 0.2508, + "num_input_tokens_seen": 28598272, + "step": 37495 + }, + { + "epoch": 77.96257796257797, + "grad_norm": 0.0003114208811894059, + "learning_rate": 0.002884506746970461, + "loss": 0.2349, + "num_input_tokens_seen": 28602048, + "step": 37500 + }, + { + "epoch": 77.97297297297297, + "grad_norm": 0.0004258941044099629, + "learning_rate": 0.0028730217848021654, + "loss": 0.2858, + "num_input_tokens_seen": 28605824, + "step": 37505 + }, + { + "epoch": 77.98336798336798, + "grad_norm": 0.0004563272523228079, + "learning_rate": 0.0028615595114628188, + "loss": 0.2592, + "num_input_tokens_seen": 28609664, + "step": 37510 + }, + { + "epoch": 77.993762993763, + "grad_norm": 0.00025951562565751374, + "learning_rate": 0.002850119928720074, + "loss": 0.2627, + "num_input_tokens_seen": 28613472, + "step": 37515 + }, + { + "epoch": 78.004158004158, + "grad_norm": 0.00015031058865133673, + "learning_rate": 0.0028387030383380195, + "loss": 0.2443, + "num_input_tokens_seen": 28617232, + "step": 37520 + }, + { + "epoch": 78.01455301455302, + "grad_norm": 4.4965101551497355e-05, + "learning_rate": 0.0028273088420772974, + "loss": 0.2599, + "num_input_tokens_seen": 28621104, + "step": 37525 + }, + { + "epoch": 78.02494802494803, + "grad_norm": 0.00021094904514029622, + "learning_rate": 0.002815937341695068, + "loss": 0.3042, + "num_input_tokens_seen": 28625008, + "step": 37530 + }, + { + "epoch": 78.03534303534303, + "grad_norm": 0.00020221843442413956, + "learning_rate": 0.0028045885389448963, + "loss": 0.2747, + "num_input_tokens_seen": 28628848, + "step": 37535 + }, + { + "epoch": 78.04573804573805, + "grad_norm": 0.0005335089517757297, + "learning_rate": 0.002793262435576965, + "loss": 0.2728, + "num_input_tokens_seen": 28632688, + "step": 37540 + }, + { + "epoch": 78.05613305613305, + "grad_norm": 0.000485223630676046, + "learning_rate": 0.0027819590333378772, + "loss": 0.2665, + "num_input_tokens_seen": 28636624, + "step": 37545 + }, + { + "epoch": 78.06652806652806, + "grad_norm": 0.0006221400108188391, + "learning_rate": 0.002770678333970755, + "loss": 0.2671, + "num_input_tokens_seen": 28640368, + "step": 37550 + }, + { + "epoch": 78.07692307692308, + "grad_norm": 0.00023803344811312854, + "learning_rate": 0.0027594203392152573, + "loss": 0.2642, + "num_input_tokens_seen": 28644304, + "step": 37555 + }, + { + "epoch": 78.08731808731808, + "grad_norm": 0.00012186486856080592, + "learning_rate": 0.002748185050807478, + "loss": 0.2778, + "num_input_tokens_seen": 28648016, + "step": 37560 + }, + { + "epoch": 78.0977130977131, + "grad_norm": 0.0003063289914280176, + "learning_rate": 0.002736972470480031, + "loss": 0.2571, + "num_input_tokens_seen": 28651824, + "step": 37565 + }, + { + "epoch": 78.10810810810811, + "grad_norm": 0.0003650210564956069, + "learning_rate": 0.002725782599962068, + "loss": 0.2571, + "num_input_tokens_seen": 28655696, + "step": 37570 + }, + { + "epoch": 78.11850311850311, + "grad_norm": 0.0003527977969497442, + "learning_rate": 0.0027146154409791734, + "loss": 0.2614, + "num_input_tokens_seen": 28659504, + "step": 37575 + }, + { + "epoch": 78.12889812889813, + "grad_norm": 0.0002823907125275582, + "learning_rate": 0.002703470995253504, + "loss": 0.2585, + "num_input_tokens_seen": 28663440, + "step": 37580 + }, + { + "epoch": 78.13929313929314, + "grad_norm": 0.0007945362594909966, + "learning_rate": 0.0026923492645036184, + "loss": 0.2714, + "num_input_tokens_seen": 28667216, + "step": 37585 + }, + { + "epoch": 78.14968814968815, + "grad_norm": 0.0001828245585784316, + "learning_rate": 0.0026812502504446776, + "loss": 0.2472, + "num_input_tokens_seen": 28670928, + "step": 37590 + }, + { + "epoch": 78.16008316008316, + "grad_norm": 0.00011388651910237968, + "learning_rate": 0.0026701739547882798, + "loss": 0.2744, + "num_input_tokens_seen": 28674736, + "step": 37595 + }, + { + "epoch": 78.17047817047818, + "grad_norm": 0.00013194308849051595, + "learning_rate": 0.0026591203792425077, + "loss": 0.2478, + "num_input_tokens_seen": 28678672, + "step": 37600 + }, + { + "epoch": 78.17047817047818, + "eval_loss": 0.2498016059398651, + "eval_runtime": 13.3771, + "eval_samples_per_second": 63.99, + "eval_steps_per_second": 15.997, + "num_input_tokens_seen": 28678672, + "step": 37600 + }, + { + "epoch": 78.18087318087318, + "grad_norm": 0.0006715295021422207, + "learning_rate": 0.0026480895255119818, + "loss": 0.2496, + "num_input_tokens_seen": 28682512, + "step": 37605 + }, + { + "epoch": 78.1912681912682, + "grad_norm": 0.000341462146025151, + "learning_rate": 0.002637081395297791, + "loss": 0.259, + "num_input_tokens_seen": 28686352, + "step": 37610 + }, + { + "epoch": 78.20166320166321, + "grad_norm": 0.00013333007518667728, + "learning_rate": 0.0026260959902975113, + "loss": 0.2537, + "num_input_tokens_seen": 28690288, + "step": 37615 + }, + { + "epoch": 78.21205821205821, + "grad_norm": 5.83543733227998e-05, + "learning_rate": 0.00261513331220527, + "loss": 0.281, + "num_input_tokens_seen": 28694096, + "step": 37620 + }, + { + "epoch": 78.22245322245323, + "grad_norm": 0.0006748784799128771, + "learning_rate": 0.0026041933627116154, + "loss": 0.2653, + "num_input_tokens_seen": 28697808, + "step": 37625 + }, + { + "epoch": 78.23284823284823, + "grad_norm": 0.0004393504641484469, + "learning_rate": 0.0025932761435036476, + "loss": 0.2495, + "num_input_tokens_seen": 28701552, + "step": 37630 + }, + { + "epoch": 78.24324324324324, + "grad_norm": 0.0006128849927335978, + "learning_rate": 0.002582381656264904, + "loss": 0.278, + "num_input_tokens_seen": 28705296, + "step": 37635 + }, + { + "epoch": 78.25363825363826, + "grad_norm": 0.00019382509344723076, + "learning_rate": 0.0025715099026754895, + "loss": 0.2672, + "num_input_tokens_seen": 28709040, + "step": 37640 + }, + { + "epoch": 78.26403326403326, + "grad_norm": 0.0005110765341669321, + "learning_rate": 0.002560660884411947, + "loss": 0.2689, + "num_input_tokens_seen": 28712848, + "step": 37645 + }, + { + "epoch": 78.27442827442827, + "grad_norm": 0.00012160575715824962, + "learning_rate": 0.0025498346031473385, + "loss": 0.2483, + "num_input_tokens_seen": 28716720, + "step": 37650 + }, + { + "epoch": 78.28482328482329, + "grad_norm": 6.866976764285937e-05, + "learning_rate": 0.0025390310605511945, + "loss": 0.2578, + "num_input_tokens_seen": 28720592, + "step": 37655 + }, + { + "epoch": 78.29521829521829, + "grad_norm": 0.0005914228968322277, + "learning_rate": 0.0025282502582895995, + "loss": 0.246, + "num_input_tokens_seen": 28724336, + "step": 37660 + }, + { + "epoch": 78.3056133056133, + "grad_norm": 0.0006792324129492044, + "learning_rate": 0.002517492198025023, + "loss": 0.2412, + "num_input_tokens_seen": 28728208, + "step": 37665 + }, + { + "epoch": 78.31600831600832, + "grad_norm": 0.0005745728267356753, + "learning_rate": 0.0025067568814165554, + "loss": 0.2939, + "num_input_tokens_seen": 28731952, + "step": 37670 + }, + { + "epoch": 78.32640332640332, + "grad_norm": 0.00017038217629306018, + "learning_rate": 0.0024960443101196884, + "loss": 0.2717, + "num_input_tokens_seen": 28735792, + "step": 37675 + }, + { + "epoch": 78.33679833679834, + "grad_norm": 0.0009870089124888182, + "learning_rate": 0.002485354485786434, + "loss": 0.2672, + "num_input_tokens_seen": 28739824, + "step": 37680 + }, + { + "epoch": 78.34719334719335, + "grad_norm": 6.939924787729979e-05, + "learning_rate": 0.002474687410065307, + "loss": 0.2605, + "num_input_tokens_seen": 28743728, + "step": 37685 + }, + { + "epoch": 78.35758835758836, + "grad_norm": 0.0002048071037279442, + "learning_rate": 0.002464043084601308, + "loss": 0.2775, + "num_input_tokens_seen": 28747568, + "step": 37690 + }, + { + "epoch": 78.36798336798337, + "grad_norm": 0.00020292414410505444, + "learning_rate": 0.0024534215110358915, + "loss": 0.2634, + "num_input_tokens_seen": 28751408, + "step": 37695 + }, + { + "epoch": 78.37837837837837, + "grad_norm": 0.0010969911236315966, + "learning_rate": 0.002442822691007096, + "loss": 0.2733, + "num_input_tokens_seen": 28755088, + "step": 37700 + }, + { + "epoch": 78.38877338877339, + "grad_norm": 0.0005535022937692702, + "learning_rate": 0.002432246626149348, + "loss": 0.2584, + "num_input_tokens_seen": 28758896, + "step": 37705 + }, + { + "epoch": 78.3991683991684, + "grad_norm": 0.00024046578619163483, + "learning_rate": 0.002421693318093626, + "loss": 0.2622, + "num_input_tokens_seen": 28762736, + "step": 37710 + }, + { + "epoch": 78.4095634095634, + "grad_norm": 0.00024482040316797793, + "learning_rate": 0.0024111627684673784, + "loss": 0.2783, + "num_input_tokens_seen": 28766640, + "step": 37715 + }, + { + "epoch": 78.41995841995842, + "grad_norm": 0.00029359760810621083, + "learning_rate": 0.0024006549788945395, + "loss": 0.2616, + "num_input_tokens_seen": 28770480, + "step": 37720 + }, + { + "epoch": 78.43035343035343, + "grad_norm": 0.0002024842397077009, + "learning_rate": 0.0023901699509955463, + "loss": 0.2518, + "num_input_tokens_seen": 28774256, + "step": 37725 + }, + { + "epoch": 78.44074844074844, + "grad_norm": 0.00023607561888638884, + "learning_rate": 0.0023797076863873554, + "loss": 0.2566, + "num_input_tokens_seen": 28778096, + "step": 37730 + }, + { + "epoch": 78.45114345114345, + "grad_norm": 0.0002189069928135723, + "learning_rate": 0.0023692681866833262, + "loss": 0.2687, + "num_input_tokens_seen": 28781872, + "step": 37735 + }, + { + "epoch": 78.46153846153847, + "grad_norm": 0.00016448796668555588, + "learning_rate": 0.0023588514534934046, + "loss": 0.2666, + "num_input_tokens_seen": 28785648, + "step": 37740 + }, + { + "epoch": 78.47193347193347, + "grad_norm": 0.00045395278721116483, + "learning_rate": 0.002348457488423955, + "loss": 0.2634, + "num_input_tokens_seen": 28789520, + "step": 37745 + }, + { + "epoch": 78.48232848232848, + "grad_norm": 0.00012836877431254834, + "learning_rate": 0.0023380862930778624, + "loss": 0.2821, + "num_input_tokens_seen": 28793328, + "step": 37750 + }, + { + "epoch": 78.4927234927235, + "grad_norm": 0.0006201587384566665, + "learning_rate": 0.0023277378690545135, + "loss": 0.2515, + "num_input_tokens_seen": 28797168, + "step": 37755 + }, + { + "epoch": 78.5031185031185, + "grad_norm": 0.00013783652684651315, + "learning_rate": 0.0023174122179497325, + "loss": 0.2508, + "num_input_tokens_seen": 28800944, + "step": 37760 + }, + { + "epoch": 78.51351351351352, + "grad_norm": 0.0002832898171618581, + "learning_rate": 0.0023071093413558784, + "loss": 0.256, + "num_input_tokens_seen": 28804688, + "step": 37765 + }, + { + "epoch": 78.52390852390852, + "grad_norm": 7.736971747362986e-05, + "learning_rate": 0.002296829240861814, + "loss": 0.2843, + "num_input_tokens_seen": 28808528, + "step": 37770 + }, + { + "epoch": 78.53430353430353, + "grad_norm": 0.00013587898865807801, + "learning_rate": 0.002286571918052821, + "loss": 0.2707, + "num_input_tokens_seen": 28812528, + "step": 37775 + }, + { + "epoch": 78.54469854469855, + "grad_norm": 0.00033153416006825864, + "learning_rate": 0.0022763373745107174, + "loss": 0.2677, + "num_input_tokens_seen": 28816304, + "step": 37780 + }, + { + "epoch": 78.55509355509355, + "grad_norm": 0.0015719928778707981, + "learning_rate": 0.0022661256118138074, + "loss": 0.2429, + "num_input_tokens_seen": 28820336, + "step": 37785 + }, + { + "epoch": 78.56548856548856, + "grad_norm": 0.00014601957809645683, + "learning_rate": 0.0022559366315368645, + "loss": 0.2792, + "num_input_tokens_seen": 28824048, + "step": 37790 + }, + { + "epoch": 78.57588357588358, + "grad_norm": 6.617201870540157e-05, + "learning_rate": 0.002245770435251182, + "loss": 0.257, + "num_input_tokens_seen": 28827792, + "step": 37795 + }, + { + "epoch": 78.58627858627858, + "grad_norm": 8.801106014288962e-05, + "learning_rate": 0.002235627024524456, + "loss": 0.2691, + "num_input_tokens_seen": 28831632, + "step": 37800 + }, + { + "epoch": 78.58627858627858, + "eval_loss": 0.2498367875814438, + "eval_runtime": 13.3721, + "eval_samples_per_second": 64.014, + "eval_steps_per_second": 16.003, + "num_input_tokens_seen": 28831632, + "step": 37800 + }, + { + "epoch": 78.5966735966736, + "grad_norm": 0.00012745764979626983, + "learning_rate": 0.0022255064009209847, + "loss": 0.2632, + "num_input_tokens_seen": 28835376, + "step": 37805 + }, + { + "epoch": 78.60706860706861, + "grad_norm": 0.00037790200440213084, + "learning_rate": 0.0022154085660014864, + "loss": 0.2546, + "num_input_tokens_seen": 28839088, + "step": 37810 + }, + { + "epoch": 78.61746361746361, + "grad_norm": 0.00021950708469375968, + "learning_rate": 0.0022053335213231494, + "loss": 0.25, + "num_input_tokens_seen": 28843024, + "step": 37815 + }, + { + "epoch": 78.62785862785863, + "grad_norm": 0.00014048628509044647, + "learning_rate": 0.002195281268439697, + "loss": 0.2439, + "num_input_tokens_seen": 28846640, + "step": 37820 + }, + { + "epoch": 78.63825363825364, + "grad_norm": 0.00031565650715492666, + "learning_rate": 0.002185251808901306, + "loss": 0.2377, + "num_input_tokens_seen": 28850448, + "step": 37825 + }, + { + "epoch": 78.64864864864865, + "grad_norm": 0.00028254787321202457, + "learning_rate": 0.0021752451442546227, + "loss": 0.2537, + "num_input_tokens_seen": 28854256, + "step": 37830 + }, + { + "epoch": 78.65904365904366, + "grad_norm": 0.0005398524226620793, + "learning_rate": 0.0021652612760428456, + "loss": 0.2391, + "num_input_tokens_seen": 28858032, + "step": 37835 + }, + { + "epoch": 78.66943866943868, + "grad_norm": 0.0002082987775793299, + "learning_rate": 0.0021553002058055603, + "loss": 0.2702, + "num_input_tokens_seen": 28861776, + "step": 37840 + }, + { + "epoch": 78.67983367983368, + "grad_norm": 0.0008171290392056108, + "learning_rate": 0.0021453619350789376, + "loss": 0.2725, + "num_input_tokens_seen": 28865520, + "step": 37845 + }, + { + "epoch": 78.6902286902287, + "grad_norm": 0.00023345788940787315, + "learning_rate": 0.0021354464653955516, + "loss": 0.2352, + "num_input_tokens_seen": 28869264, + "step": 37850 + }, + { + "epoch": 78.7006237006237, + "grad_norm": 0.0004140743112657219, + "learning_rate": 0.002125553798284513, + "loss": 0.2612, + "num_input_tokens_seen": 28873040, + "step": 37855 + }, + { + "epoch": 78.71101871101871, + "grad_norm": 0.00019932870054617524, + "learning_rate": 0.002115683935271384, + "loss": 0.2661, + "num_input_tokens_seen": 28876752, + "step": 37860 + }, + { + "epoch": 78.72141372141373, + "grad_norm": 0.0005794940516352654, + "learning_rate": 0.0021058368778782144, + "loss": 0.2517, + "num_input_tokens_seen": 28880400, + "step": 37865 + }, + { + "epoch": 78.73180873180873, + "grad_norm": 0.00010781781747937202, + "learning_rate": 0.002096012627623539, + "loss": 0.2622, + "num_input_tokens_seen": 28884176, + "step": 37870 + }, + { + "epoch": 78.74220374220374, + "grad_norm": 0.00013670553744304925, + "learning_rate": 0.00208621118602243, + "loss": 0.2588, + "num_input_tokens_seen": 28887888, + "step": 37875 + }, + { + "epoch": 78.75259875259876, + "grad_norm": 0.00015113857807591558, + "learning_rate": 0.002076432554586327, + "loss": 0.2581, + "num_input_tokens_seen": 28891696, + "step": 37880 + }, + { + "epoch": 78.76299376299376, + "grad_norm": 0.00019641057588160038, + "learning_rate": 0.002066676734823258, + "loss": 0.268, + "num_input_tokens_seen": 28895408, + "step": 37885 + }, + { + "epoch": 78.77338877338877, + "grad_norm": 0.00032262125751003623, + "learning_rate": 0.0020569437282376866, + "loss": 0.286, + "num_input_tokens_seen": 28899056, + "step": 37890 + }, + { + "epoch": 78.78378378378379, + "grad_norm": 0.00017668158398009837, + "learning_rate": 0.002047233536330545, + "loss": 0.2706, + "num_input_tokens_seen": 28902704, + "step": 37895 + }, + { + "epoch": 78.79417879417879, + "grad_norm": 0.00016258470714092255, + "learning_rate": 0.0020375461605993015, + "loss": 0.2573, + "num_input_tokens_seen": 28906352, + "step": 37900 + }, + { + "epoch": 78.8045738045738, + "grad_norm": 0.00019959069322794676, + "learning_rate": 0.002027881602537845, + "loss": 0.2839, + "num_input_tokens_seen": 28910160, + "step": 37905 + }, + { + "epoch": 78.81496881496882, + "grad_norm": 0.0014921323163434863, + "learning_rate": 0.002018239863636567, + "loss": 0.2781, + "num_input_tokens_seen": 28914064, + "step": 37910 + }, + { + "epoch": 78.82536382536382, + "grad_norm": 0.0007923303637653589, + "learning_rate": 0.002008620945382378, + "loss": 0.264, + "num_input_tokens_seen": 28917904, + "step": 37915 + }, + { + "epoch": 78.83575883575884, + "grad_norm": 0.002255060477182269, + "learning_rate": 0.001999024849258607, + "loss": 0.2432, + "num_input_tokens_seen": 28921744, + "step": 37920 + }, + { + "epoch": 78.84615384615384, + "grad_norm": 0.0009984582429751754, + "learning_rate": 0.001989451576745105, + "loss": 0.2721, + "num_input_tokens_seen": 28925616, + "step": 37925 + }, + { + "epoch": 78.85654885654886, + "grad_norm": 0.0002737363975029439, + "learning_rate": 0.00197990112931819, + "loss": 0.2308, + "num_input_tokens_seen": 28929488, + "step": 37930 + }, + { + "epoch": 78.86694386694387, + "grad_norm": 0.00023218551359605044, + "learning_rate": 0.0019703735084506345, + "loss": 0.2473, + "num_input_tokens_seen": 28933296, + "step": 37935 + }, + { + "epoch": 78.87733887733887, + "grad_norm": 0.0006973809795454144, + "learning_rate": 0.001960868715611763, + "loss": 0.24, + "num_input_tokens_seen": 28937072, + "step": 37940 + }, + { + "epoch": 78.88773388773389, + "grad_norm": 0.00022784974134992808, + "learning_rate": 0.0019513867522673034, + "loss": 0.2559, + "num_input_tokens_seen": 28940944, + "step": 37945 + }, + { + "epoch": 78.8981288981289, + "grad_norm": 0.0004798508598469198, + "learning_rate": 0.001941927619879502, + "loss": 0.2755, + "num_input_tokens_seen": 28944720, + "step": 37950 + }, + { + "epoch": 78.9085239085239, + "grad_norm": 0.0010758964344859123, + "learning_rate": 0.0019324913199070758, + "loss": 0.2845, + "num_input_tokens_seen": 28948816, + "step": 37955 + }, + { + "epoch": 78.91891891891892, + "grad_norm": 0.00016881241754163057, + "learning_rate": 0.0019230778538052106, + "loss": 0.2549, + "num_input_tokens_seen": 28952848, + "step": 37960 + }, + { + "epoch": 78.92931392931393, + "grad_norm": 0.00033863424323499203, + "learning_rate": 0.0019136872230255952, + "loss": 0.2767, + "num_input_tokens_seen": 28956688, + "step": 37965 + }, + { + "epoch": 78.93970893970894, + "grad_norm": 0.0002336965117137879, + "learning_rate": 0.0019043194290164045, + "loss": 0.265, + "num_input_tokens_seen": 28960400, + "step": 37970 + }, + { + "epoch": 78.95010395010395, + "grad_norm": 0.00016260083066299558, + "learning_rate": 0.0018949744732222162, + "loss": 0.2651, + "num_input_tokens_seen": 28964208, + "step": 37975 + }, + { + "epoch": 78.96049896049897, + "grad_norm": 0.00023800510098226368, + "learning_rate": 0.0018856523570841776, + "loss": 0.2815, + "num_input_tokens_seen": 28968176, + "step": 37980 + }, + { + "epoch": 78.97089397089397, + "grad_norm": 0.000232863865676336, + "learning_rate": 0.0018763530820398555, + "loss": 0.2432, + "num_input_tokens_seen": 28972016, + "step": 37985 + }, + { + "epoch": 78.98128898128898, + "grad_norm": 0.0002703905920498073, + "learning_rate": 0.0018670766495233525, + "loss": 0.2359, + "num_input_tokens_seen": 28975824, + "step": 37990 + }, + { + "epoch": 78.99168399168398, + "grad_norm": 0.0004151446628384292, + "learning_rate": 0.001857823060965158, + "loss": 0.2633, + "num_input_tokens_seen": 28979568, + "step": 37995 + }, + { + "epoch": 79.002079002079, + "grad_norm": 0.00032281759195029736, + "learning_rate": 0.0018485923177923467, + "loss": 0.2608, + "num_input_tokens_seen": 28983144, + "step": 38000 + }, + { + "epoch": 79.002079002079, + "eval_loss": 0.2485750913619995, + "eval_runtime": 13.3689, + "eval_samples_per_second": 64.029, + "eval_steps_per_second": 16.007, + "num_input_tokens_seen": 28983144, + "step": 38000 + }, + { + "epoch": 79.01247401247402, + "grad_norm": 0.00017966370796784759, + "learning_rate": 0.001839384421428364, + "loss": 0.262, + "num_input_tokens_seen": 28986888, + "step": 38005 + }, + { + "epoch": 79.02286902286902, + "grad_norm": 0.00016717262042220682, + "learning_rate": 0.0018301993732932065, + "loss": 0.2477, + "num_input_tokens_seen": 28990696, + "step": 38010 + }, + { + "epoch": 79.03326403326403, + "grad_norm": 0.0004760668089147657, + "learning_rate": 0.0018210371748033248, + "loss": 0.2709, + "num_input_tokens_seen": 28994504, + "step": 38015 + }, + { + "epoch": 79.04365904365905, + "grad_norm": 0.00012353732017800212, + "learning_rate": 0.0018118978273716556, + "loss": 0.2517, + "num_input_tokens_seen": 28998184, + "step": 38020 + }, + { + "epoch": 79.05405405405405, + "grad_norm": 0.0002884168934542686, + "learning_rate": 0.001802781332407588, + "loss": 0.2772, + "num_input_tokens_seen": 29001960, + "step": 38025 + }, + { + "epoch": 79.06444906444906, + "grad_norm": 0.00020809365378227085, + "learning_rate": 0.0017936876913169806, + "loss": 0.2683, + "num_input_tokens_seen": 29005736, + "step": 38030 + }, + { + "epoch": 79.07484407484408, + "grad_norm": 0.0002984378661494702, + "learning_rate": 0.0017846169055022287, + "loss": 0.2628, + "num_input_tokens_seen": 29009576, + "step": 38035 + }, + { + "epoch": 79.08523908523908, + "grad_norm": 0.0005485824658535421, + "learning_rate": 0.0017755689763621295, + "loss": 0.2568, + "num_input_tokens_seen": 29013192, + "step": 38040 + }, + { + "epoch": 79.0956340956341, + "grad_norm": 0.000435478868894279, + "learning_rate": 0.0017665439052920173, + "loss": 0.2855, + "num_input_tokens_seen": 29016872, + "step": 38045 + }, + { + "epoch": 79.10602910602911, + "grad_norm": 0.0005841461825184524, + "learning_rate": 0.0017575416936836286, + "loss": 0.2606, + "num_input_tokens_seen": 29020744, + "step": 38050 + }, + { + "epoch": 79.11642411642411, + "grad_norm": 0.00043549275142140687, + "learning_rate": 0.0017485623429252528, + "loss": 0.2534, + "num_input_tokens_seen": 29024520, + "step": 38055 + }, + { + "epoch": 79.12681912681913, + "grad_norm": 0.0008446744759567082, + "learning_rate": 0.0017396058544016156, + "loss": 0.2825, + "num_input_tokens_seen": 29028456, + "step": 38060 + }, + { + "epoch": 79.13721413721414, + "grad_norm": 0.00025966682005673647, + "learning_rate": 0.0017306722294938958, + "loss": 0.2656, + "num_input_tokens_seen": 29032424, + "step": 38065 + }, + { + "epoch": 79.14760914760915, + "grad_norm": 0.00022661415277980268, + "learning_rate": 0.0017217614695798078, + "loss": 0.2442, + "num_input_tokens_seen": 29036136, + "step": 38070 + }, + { + "epoch": 79.15800415800416, + "grad_norm": 0.00042740910430438817, + "learning_rate": 0.001712873576033469, + "loss": 0.2537, + "num_input_tokens_seen": 29040136, + "step": 38075 + }, + { + "epoch": 79.16839916839916, + "grad_norm": 0.0006336811347864568, + "learning_rate": 0.0017040085502255163, + "loss": 0.2533, + "num_input_tokens_seen": 29043976, + "step": 38080 + }, + { + "epoch": 79.17879417879418, + "grad_norm": 0.00026188933406956494, + "learning_rate": 0.0016951663935230565, + "loss": 0.261, + "num_input_tokens_seen": 29047848, + "step": 38085 + }, + { + "epoch": 79.1891891891892, + "grad_norm": 0.00043104070937260985, + "learning_rate": 0.0016863471072896485, + "loss": 0.2492, + "num_input_tokens_seen": 29051656, + "step": 38090 + }, + { + "epoch": 79.1995841995842, + "grad_norm": 0.0008819380309432745, + "learning_rate": 0.0016775506928853377, + "loss": 0.2843, + "num_input_tokens_seen": 29055560, + "step": 38095 + }, + { + "epoch": 79.20997920997921, + "grad_norm": 0.00024019851116463542, + "learning_rate": 0.001668777151666656, + "loss": 0.2419, + "num_input_tokens_seen": 29059176, + "step": 38100 + }, + { + "epoch": 79.22037422037423, + "grad_norm": 0.0003162742650602013, + "learning_rate": 0.0016600264849865709, + "loss": 0.2706, + "num_input_tokens_seen": 29063048, + "step": 38105 + }, + { + "epoch": 79.23076923076923, + "grad_norm": 0.0008727354579605162, + "learning_rate": 0.0016512986941945695, + "loss": 0.2565, + "num_input_tokens_seen": 29066984, + "step": 38110 + }, + { + "epoch": 79.24116424116424, + "grad_norm": 0.00014136887330096215, + "learning_rate": 0.0016425937806365753, + "loss": 0.2664, + "num_input_tokens_seen": 29070792, + "step": 38115 + }, + { + "epoch": 79.25155925155926, + "grad_norm": 0.00011961320706177503, + "learning_rate": 0.0016339117456549979, + "loss": 0.2415, + "num_input_tokens_seen": 29074376, + "step": 38120 + }, + { + "epoch": 79.26195426195426, + "grad_norm": 0.0005950881168246269, + "learning_rate": 0.0016252525905886995, + "loss": 0.2604, + "num_input_tokens_seen": 29078184, + "step": 38125 + }, + { + "epoch": 79.27234927234927, + "grad_norm": 0.00018996621656697243, + "learning_rate": 0.0016166163167730617, + "loss": 0.274, + "num_input_tokens_seen": 29082184, + "step": 38130 + }, + { + "epoch": 79.28274428274429, + "grad_norm": 0.0004959648940712214, + "learning_rate": 0.0016080029255398864, + "loss": 0.2775, + "num_input_tokens_seen": 29086024, + "step": 38135 + }, + { + "epoch": 79.29313929313929, + "grad_norm": 7.827136141713709e-05, + "learning_rate": 0.0015994124182174606, + "loss": 0.2498, + "num_input_tokens_seen": 29089928, + "step": 38140 + }, + { + "epoch": 79.3035343035343, + "grad_norm": 0.0004906129324808717, + "learning_rate": 0.001590844796130575, + "loss": 0.2543, + "num_input_tokens_seen": 29093800, + "step": 38145 + }, + { + "epoch": 79.31392931392931, + "grad_norm": 0.00034243756090290844, + "learning_rate": 0.001582300060600439, + "loss": 0.2735, + "num_input_tokens_seen": 29097704, + "step": 38150 + }, + { + "epoch": 79.32432432432432, + "grad_norm": 0.0002170712104998529, + "learning_rate": 0.0015737782129447652, + "loss": 0.2558, + "num_input_tokens_seen": 29101672, + "step": 38155 + }, + { + "epoch": 79.33471933471934, + "grad_norm": 0.00041325410711579025, + "learning_rate": 0.0015652792544777361, + "loss": 0.268, + "num_input_tokens_seen": 29105480, + "step": 38160 + }, + { + "epoch": 79.34511434511434, + "grad_norm": 0.00011577887198654935, + "learning_rate": 0.0015568031865099863, + "loss": 0.2682, + "num_input_tokens_seen": 29109352, + "step": 38165 + }, + { + "epoch": 79.35550935550935, + "grad_norm": 0.0023331954143941402, + "learning_rate": 0.0015483500103486369, + "loss": 0.2708, + "num_input_tokens_seen": 29113256, + "step": 38170 + }, + { + "epoch": 79.36590436590437, + "grad_norm": 0.00028183453832753, + "learning_rate": 0.0015399197272972787, + "loss": 0.2552, + "num_input_tokens_seen": 29116968, + "step": 38175 + }, + { + "epoch": 79.37629937629937, + "grad_norm": 0.00011987083416897804, + "learning_rate": 0.0015315123386559714, + "loss": 0.2516, + "num_input_tokens_seen": 29120776, + "step": 38180 + }, + { + "epoch": 79.38669438669439, + "grad_norm": 0.00038736755959689617, + "learning_rate": 0.0015231278457212283, + "loss": 0.2677, + "num_input_tokens_seen": 29124616, + "step": 38185 + }, + { + "epoch": 79.3970893970894, + "grad_norm": 0.00048630518722347915, + "learning_rate": 0.001514766249786048, + "loss": 0.2726, + "num_input_tokens_seen": 29128360, + "step": 38190 + }, + { + "epoch": 79.4074844074844, + "grad_norm": 0.00014829615247435868, + "learning_rate": 0.0015064275521398994, + "loss": 0.2455, + "num_input_tokens_seen": 29132232, + "step": 38195 + }, + { + "epoch": 79.41787941787942, + "grad_norm": 8.91961099114269e-05, + "learning_rate": 0.0014981117540686872, + "loss": 0.2627, + "num_input_tokens_seen": 29136008, + "step": 38200 + }, + { + "epoch": 79.41787941787942, + "eval_loss": 0.24872873723506927, + "eval_runtime": 13.3604, + "eval_samples_per_second": 64.07, + "eval_steps_per_second": 16.017, + "num_input_tokens_seen": 29136008, + "step": 38200 + }, + { + "epoch": 79.42827442827443, + "grad_norm": 0.00026118342066183686, + "learning_rate": 0.0014898188568548687, + "loss": 0.2796, + "num_input_tokens_seen": 29139912, + "step": 38205 + }, + { + "epoch": 79.43866943866944, + "grad_norm": 0.0008638104773126543, + "learning_rate": 0.0014815488617772542, + "loss": 0.2728, + "num_input_tokens_seen": 29143848, + "step": 38210 + }, + { + "epoch": 79.44906444906445, + "grad_norm": 0.00042454988579265773, + "learning_rate": 0.0014733017701112072, + "loss": 0.2578, + "num_input_tokens_seen": 29147656, + "step": 38215 + }, + { + "epoch": 79.45945945945945, + "grad_norm": 0.00028132752049714327, + "learning_rate": 0.0014650775831285435, + "loss": 0.2601, + "num_input_tokens_seen": 29151176, + "step": 38220 + }, + { + "epoch": 79.46985446985447, + "grad_norm": 0.0004521938681136817, + "learning_rate": 0.001456876302097515, + "loss": 0.2728, + "num_input_tokens_seen": 29154856, + "step": 38225 + }, + { + "epoch": 79.48024948024948, + "grad_norm": 0.0001556698844069615, + "learning_rate": 0.0014486979282828604, + "loss": 0.2164, + "num_input_tokens_seen": 29158696, + "step": 38230 + }, + { + "epoch": 79.49064449064448, + "grad_norm": 9.598196629667655e-05, + "learning_rate": 0.001440542462945804, + "loss": 0.245, + "num_input_tokens_seen": 29162568, + "step": 38235 + }, + { + "epoch": 79.5010395010395, + "grad_norm": 0.0004440545162651688, + "learning_rate": 0.0014324099073440232, + "loss": 0.2664, + "num_input_tokens_seen": 29166376, + "step": 38240 + }, + { + "epoch": 79.51143451143452, + "grad_norm": 0.00024112591927405447, + "learning_rate": 0.0014243002627316482, + "loss": 0.2563, + "num_input_tokens_seen": 29170184, + "step": 38245 + }, + { + "epoch": 79.52182952182952, + "grad_norm": 0.0003036188136320561, + "learning_rate": 0.0014162135303592781, + "loss": 0.272, + "num_input_tokens_seen": 29173928, + "step": 38250 + }, + { + "epoch": 79.53222453222453, + "grad_norm": 6.247423152672127e-05, + "learning_rate": 0.001408149711474016, + "loss": 0.2677, + "num_input_tokens_seen": 29177832, + "step": 38255 + }, + { + "epoch": 79.54261954261955, + "grad_norm": 0.0001993149344343692, + "learning_rate": 0.0014001088073193834, + "loss": 0.2667, + "num_input_tokens_seen": 29181608, + "step": 38260 + }, + { + "epoch": 79.55301455301455, + "grad_norm": 0.0005638791481032968, + "learning_rate": 0.0013920908191354052, + "loss": 0.2849, + "num_input_tokens_seen": 29185384, + "step": 38265 + }, + { + "epoch": 79.56340956340956, + "grad_norm": 0.00023856728512328118, + "learning_rate": 0.001384095748158526, + "loss": 0.2511, + "num_input_tokens_seen": 29189160, + "step": 38270 + }, + { + "epoch": 79.57380457380458, + "grad_norm": 0.0006954625132493675, + "learning_rate": 0.0013761235956217255, + "loss": 0.2514, + "num_input_tokens_seen": 29193000, + "step": 38275 + }, + { + "epoch": 79.58419958419958, + "grad_norm": 0.0007552440511062741, + "learning_rate": 0.0013681743627543873, + "loss": 0.2747, + "num_input_tokens_seen": 29196616, + "step": 38280 + }, + { + "epoch": 79.5945945945946, + "grad_norm": 0.0001778706646291539, + "learning_rate": 0.001360248050782381, + "loss": 0.2456, + "num_input_tokens_seen": 29200360, + "step": 38285 + }, + { + "epoch": 79.60498960498961, + "grad_norm": 0.0003644452372100204, + "learning_rate": 0.001352344660928062, + "loss": 0.251, + "num_input_tokens_seen": 29204328, + "step": 38290 + }, + { + "epoch": 79.61538461538461, + "grad_norm": 0.0005265697836875916, + "learning_rate": 0.0013444641944102052, + "loss": 0.268, + "num_input_tokens_seen": 29208136, + "step": 38295 + }, + { + "epoch": 79.62577962577963, + "grad_norm": 0.00045020924881100655, + "learning_rate": 0.0013366066524441056, + "loss": 0.257, + "num_input_tokens_seen": 29212008, + "step": 38300 + }, + { + "epoch": 79.63617463617463, + "grad_norm": 0.00022348729544319212, + "learning_rate": 0.0013287720362414768, + "loss": 0.2642, + "num_input_tokens_seen": 29215880, + "step": 38305 + }, + { + "epoch": 79.64656964656965, + "grad_norm": 0.0001538680080557242, + "learning_rate": 0.0013209603470105025, + "loss": 0.2634, + "num_input_tokens_seen": 29219880, + "step": 38310 + }, + { + "epoch": 79.65696465696466, + "grad_norm": 8.355820318683982e-05, + "learning_rate": 0.0013131715859558857, + "loss": 0.2495, + "num_input_tokens_seen": 29223752, + "step": 38315 + }, + { + "epoch": 79.66735966735966, + "grad_norm": 0.0005104420124553144, + "learning_rate": 0.001305405754278699, + "loss": 0.2582, + "num_input_tokens_seen": 29227592, + "step": 38320 + }, + { + "epoch": 79.67775467775468, + "grad_norm": 0.005619477946311235, + "learning_rate": 0.0012976628531765843, + "loss": 0.2784, + "num_input_tokens_seen": 29231336, + "step": 38325 + }, + { + "epoch": 79.6881496881497, + "grad_norm": 0.0004730597138404846, + "learning_rate": 0.0012899428838435533, + "loss": 0.2602, + "num_input_tokens_seen": 29234984, + "step": 38330 + }, + { + "epoch": 79.6985446985447, + "grad_norm": 0.00043982380884699523, + "learning_rate": 0.001282245847470137, + "loss": 0.2362, + "num_input_tokens_seen": 29238792, + "step": 38335 + }, + { + "epoch": 79.70893970893971, + "grad_norm": 0.0006354756187647581, + "learning_rate": 0.001274571745243319, + "loss": 0.2552, + "num_input_tokens_seen": 29242536, + "step": 38340 + }, + { + "epoch": 79.71933471933473, + "grad_norm": 0.00026771367993205786, + "learning_rate": 0.0012669205783465364, + "loss": 0.2711, + "num_input_tokens_seen": 29246312, + "step": 38345 + }, + { + "epoch": 79.72972972972973, + "grad_norm": 0.0007430287660099566, + "learning_rate": 0.001259292347959695, + "loss": 0.295, + "num_input_tokens_seen": 29250024, + "step": 38350 + }, + { + "epoch": 79.74012474012474, + "grad_norm": 0.00045486559974960983, + "learning_rate": 0.0012516870552591707, + "loss": 0.2725, + "num_input_tokens_seen": 29253832, + "step": 38355 + }, + { + "epoch": 79.75051975051976, + "grad_norm": 0.00026794298901222646, + "learning_rate": 0.001244104701417792, + "loss": 0.2644, + "num_input_tokens_seen": 29257640, + "step": 38360 + }, + { + "epoch": 79.76091476091476, + "grad_norm": 0.00027484007296152413, + "learning_rate": 0.0012365452876048565, + "loss": 0.2727, + "num_input_tokens_seen": 29261448, + "step": 38365 + }, + { + "epoch": 79.77130977130977, + "grad_norm": 0.0002597956918179989, + "learning_rate": 0.001229008814986099, + "loss": 0.2586, + "num_input_tokens_seen": 29265192, + "step": 38370 + }, + { + "epoch": 79.78170478170478, + "grad_norm": 0.00032989526516757905, + "learning_rate": 0.0012214952847237725, + "loss": 0.2875, + "num_input_tokens_seen": 29269032, + "step": 38375 + }, + { + "epoch": 79.79209979209979, + "grad_norm": 0.0001911343861138448, + "learning_rate": 0.0012140046979765339, + "loss": 0.2701, + "num_input_tokens_seen": 29272936, + "step": 38380 + }, + { + "epoch": 79.8024948024948, + "grad_norm": 0.0003251050948165357, + "learning_rate": 0.0012065370558995258, + "loss": 0.2687, + "num_input_tokens_seen": 29276680, + "step": 38385 + }, + { + "epoch": 79.81288981288981, + "grad_norm": 0.0002685713116079569, + "learning_rate": 0.0011990923596443602, + "loss": 0.2483, + "num_input_tokens_seen": 29280392, + "step": 38390 + }, + { + "epoch": 79.82328482328482, + "grad_norm": 0.001317249028943479, + "learning_rate": 0.001191670610359119, + "loss": 0.262, + "num_input_tokens_seen": 29284200, + "step": 38395 + }, + { + "epoch": 79.83367983367984, + "grad_norm": 9.249548020306975e-05, + "learning_rate": 0.0011842718091882865, + "loss": 0.244, + "num_input_tokens_seen": 29288104, + "step": 38400 + }, + { + "epoch": 79.83367983367984, + "eval_loss": 0.24904431402683258, + "eval_runtime": 13.3762, + "eval_samples_per_second": 63.994, + "eval_steps_per_second": 15.999, + "num_input_tokens_seen": 29288104, + "step": 38400 + }, + { + "epoch": 79.84407484407484, + "grad_norm": 0.00012733276525978, + "learning_rate": 0.0011768959572729, + "loss": 0.2582, + "num_input_tokens_seen": 29291816, + "step": 38405 + }, + { + "epoch": 79.85446985446985, + "grad_norm": 0.00010336189006920904, + "learning_rate": 0.001169543055750366, + "loss": 0.2555, + "num_input_tokens_seen": 29295752, + "step": 38410 + }, + { + "epoch": 79.86486486486487, + "grad_norm": 0.00013424715143628418, + "learning_rate": 0.0011622131057546115, + "loss": 0.2844, + "num_input_tokens_seen": 29299528, + "step": 38415 + }, + { + "epoch": 79.87525987525987, + "grad_norm": 0.0008734024595469236, + "learning_rate": 0.0011549061084160316, + "loss": 0.2686, + "num_input_tokens_seen": 29303336, + "step": 38420 + }, + { + "epoch": 79.88565488565489, + "grad_norm": 0.00042071370990015566, + "learning_rate": 0.0011476220648614088, + "loss": 0.2687, + "num_input_tokens_seen": 29307048, + "step": 38425 + }, + { + "epoch": 79.8960498960499, + "grad_norm": 0.00016543905076105148, + "learning_rate": 0.0011403609762140777, + "loss": 0.2599, + "num_input_tokens_seen": 29310984, + "step": 38430 + }, + { + "epoch": 79.9064449064449, + "grad_norm": 0.0020599847193807364, + "learning_rate": 0.0011331228435937756, + "loss": 0.2573, + "num_input_tokens_seen": 29314664, + "step": 38435 + }, + { + "epoch": 79.91683991683992, + "grad_norm": 0.00016036223678383976, + "learning_rate": 0.0011259076681166935, + "loss": 0.2626, + "num_input_tokens_seen": 29318408, + "step": 38440 + }, + { + "epoch": 79.92723492723492, + "grad_norm": 0.0003660772927105427, + "learning_rate": 0.0011187154508955244, + "loss": 0.2559, + "num_input_tokens_seen": 29322280, + "step": 38445 + }, + { + "epoch": 79.93762993762994, + "grad_norm": 0.000240333829424344, + "learning_rate": 0.001111546193039381, + "loss": 0.2542, + "num_input_tokens_seen": 29326056, + "step": 38450 + }, + { + "epoch": 79.94802494802495, + "grad_norm": 0.0001794466224964708, + "learning_rate": 0.0011043998956538792, + "loss": 0.2601, + "num_input_tokens_seen": 29329800, + "step": 38455 + }, + { + "epoch": 79.95841995841995, + "grad_norm": 0.00012207345571368933, + "learning_rate": 0.0010972765598410538, + "loss": 0.2451, + "num_input_tokens_seen": 29333544, + "step": 38460 + }, + { + "epoch": 79.96881496881497, + "grad_norm": 0.000610252667684108, + "learning_rate": 0.0010901761866993931, + "loss": 0.265, + "num_input_tokens_seen": 29337384, + "step": 38465 + }, + { + "epoch": 79.97920997920998, + "grad_norm": 0.0010604034177958965, + "learning_rate": 0.0010830987773238876, + "loss": 0.2758, + "num_input_tokens_seen": 29341320, + "step": 38470 + }, + { + "epoch": 79.98960498960498, + "grad_norm": 0.00017277008737437427, + "learning_rate": 0.0010760443328059644, + "loss": 0.2675, + "num_input_tokens_seen": 29345160, + "step": 38475 + }, + { + "epoch": 80.0, + "grad_norm": 0.0006391305360011756, + "learning_rate": 0.001069012854233503, + "loss": 0.2667, + "num_input_tokens_seen": 29349080, + "step": 38480 + }, + { + "epoch": 80.01039501039502, + "grad_norm": 0.0009808116592466831, + "learning_rate": 0.0010620043426908365, + "loss": 0.2641, + "num_input_tokens_seen": 29352760, + "step": 38485 + }, + { + "epoch": 80.02079002079002, + "grad_norm": 0.00039683966315351427, + "learning_rate": 0.0010550187992587833, + "loss": 0.2348, + "num_input_tokens_seen": 29356696, + "step": 38490 + }, + { + "epoch": 80.03118503118503, + "grad_norm": 0.000833352969493717, + "learning_rate": 0.0010480562250145653, + "loss": 0.2523, + "num_input_tokens_seen": 29360504, + "step": 38495 + }, + { + "epoch": 80.04158004158005, + "grad_norm": 0.00013748662604484707, + "learning_rate": 0.0010411166210319567, + "loss": 0.2602, + "num_input_tokens_seen": 29364184, + "step": 38500 + }, + { + "epoch": 80.05197505197505, + "grad_norm": 0.0008084644796326756, + "learning_rate": 0.0010341999883810848, + "loss": 0.2497, + "num_input_tokens_seen": 29367992, + "step": 38505 + }, + { + "epoch": 80.06237006237006, + "grad_norm": 0.0005259827012196183, + "learning_rate": 0.0010273063281285965, + "loss": 0.2807, + "num_input_tokens_seen": 29371768, + "step": 38510 + }, + { + "epoch": 80.07276507276508, + "grad_norm": 0.000703663972672075, + "learning_rate": 0.0010204356413375747, + "loss": 0.2881, + "num_input_tokens_seen": 29375512, + "step": 38515 + }, + { + "epoch": 80.08316008316008, + "grad_norm": 0.00019635447824839503, + "learning_rate": 0.001013587929067572, + "loss": 0.2596, + "num_input_tokens_seen": 29379416, + "step": 38520 + }, + { + "epoch": 80.0935550935551, + "grad_norm": 0.00042644611676223576, + "learning_rate": 0.00100676319237461, + "loss": 0.2754, + "num_input_tokens_seen": 29383256, + "step": 38525 + }, + { + "epoch": 80.1039501039501, + "grad_norm": 0.00028283815481700003, + "learning_rate": 0.0009999614323110972, + "loss": 0.2654, + "num_input_tokens_seen": 29387032, + "step": 38530 + }, + { + "epoch": 80.11434511434511, + "grad_norm": 0.000180199378519319, + "learning_rate": 0.000993182649926011, + "loss": 0.2602, + "num_input_tokens_seen": 29390808, + "step": 38535 + }, + { + "epoch": 80.12474012474013, + "grad_norm": 9.865043830359355e-05, + "learning_rate": 0.000986426846264682, + "loss": 0.2652, + "num_input_tokens_seen": 29394616, + "step": 38540 + }, + { + "epoch": 80.13513513513513, + "grad_norm": 0.0003328242455609143, + "learning_rate": 0.00097969402236896, + "loss": 0.2579, + "num_input_tokens_seen": 29398424, + "step": 38545 + }, + { + "epoch": 80.14553014553015, + "grad_norm": 0.00047965397243387997, + "learning_rate": 0.0009729841792771143, + "loss": 0.2733, + "num_input_tokens_seen": 29402232, + "step": 38550 + }, + { + "epoch": 80.15592515592516, + "grad_norm": 0.000206598051590845, + "learning_rate": 0.0009662973180239176, + "loss": 0.2662, + "num_input_tokens_seen": 29406136, + "step": 38555 + }, + { + "epoch": 80.16632016632016, + "grad_norm": 0.0001571306202095002, + "learning_rate": 0.0009596334396405448, + "loss": 0.2553, + "num_input_tokens_seen": 29409816, + "step": 38560 + }, + { + "epoch": 80.17671517671518, + "grad_norm": 0.0002151893131667748, + "learning_rate": 0.0009529925451546406, + "loss": 0.2659, + "num_input_tokens_seen": 29413624, + "step": 38565 + }, + { + "epoch": 80.18711018711019, + "grad_norm": 0.000471391569590196, + "learning_rate": 0.0009463746355903357, + "loss": 0.2453, + "num_input_tokens_seen": 29417304, + "step": 38570 + }, + { + "epoch": 80.1975051975052, + "grad_norm": 0.0002690582477953285, + "learning_rate": 0.0009397797119681971, + "loss": 0.2653, + "num_input_tokens_seen": 29420984, + "step": 38575 + }, + { + "epoch": 80.20790020790021, + "grad_norm": 0.00046467757783830166, + "learning_rate": 0.0009332077753052281, + "loss": 0.2603, + "num_input_tokens_seen": 29424824, + "step": 38580 + }, + { + "epoch": 80.21829521829522, + "grad_norm": 0.00018763489788398147, + "learning_rate": 0.0009266588266149011, + "loss": 0.2607, + "num_input_tokens_seen": 29428696, + "step": 38585 + }, + { + "epoch": 80.22869022869023, + "grad_norm": 0.0005945999873802066, + "learning_rate": 0.0009201328669071584, + "loss": 0.2273, + "num_input_tokens_seen": 29432568, + "step": 38590 + }, + { + "epoch": 80.23908523908524, + "grad_norm": 0.0007603820995427668, + "learning_rate": 0.0009136298971883949, + "loss": 0.2823, + "num_input_tokens_seen": 29436472, + "step": 38595 + }, + { + "epoch": 80.24948024948024, + "grad_norm": 0.00013617028889711946, + "learning_rate": 0.0009071499184614251, + "loss": 0.269, + "num_input_tokens_seen": 29440312, + "step": 38600 + }, + { + "epoch": 80.24948024948024, + "eval_loss": 0.2489522099494934, + "eval_runtime": 13.3672, + "eval_samples_per_second": 64.037, + "eval_steps_per_second": 16.009, + "num_input_tokens_seen": 29440312, + "step": 38600 + }, + { + "epoch": 80.25987525987526, + "grad_norm": 0.0003282705438323319, + "learning_rate": 0.0009006929317255663, + "loss": 0.2673, + "num_input_tokens_seen": 29444280, + "step": 38605 + }, + { + "epoch": 80.27027027027027, + "grad_norm": 0.00017945414583664387, + "learning_rate": 0.0008942589379765387, + "loss": 0.2504, + "num_input_tokens_seen": 29448024, + "step": 38610 + }, + { + "epoch": 80.28066528066527, + "grad_norm": 0.00030554665136151016, + "learning_rate": 0.0008878479382065817, + "loss": 0.2689, + "num_input_tokens_seen": 29451864, + "step": 38615 + }, + { + "epoch": 80.29106029106029, + "grad_norm": 0.0008404122199863195, + "learning_rate": 0.0008814599334043215, + "loss": 0.2408, + "num_input_tokens_seen": 29455672, + "step": 38620 + }, + { + "epoch": 80.3014553014553, + "grad_norm": 0.00035232590744271874, + "learning_rate": 0.0008750949245548866, + "loss": 0.2668, + "num_input_tokens_seen": 29459544, + "step": 38625 + }, + { + "epoch": 80.3118503118503, + "grad_norm": 0.00018367475422564894, + "learning_rate": 0.0008687529126398252, + "loss": 0.2557, + "num_input_tokens_seen": 29463256, + "step": 38630 + }, + { + "epoch": 80.32224532224532, + "grad_norm": 8.786677062744275e-05, + "learning_rate": 0.0008624338986371715, + "loss": 0.2512, + "num_input_tokens_seen": 29467064, + "step": 38635 + }, + { + "epoch": 80.33264033264034, + "grad_norm": 0.00043246959103271365, + "learning_rate": 0.0008561378835213962, + "loss": 0.248, + "num_input_tokens_seen": 29470808, + "step": 38640 + }, + { + "epoch": 80.34303534303534, + "grad_norm": 0.002401023404672742, + "learning_rate": 0.0008498648682634058, + "loss": 0.2563, + "num_input_tokens_seen": 29474616, + "step": 38645 + }, + { + "epoch": 80.35343035343035, + "grad_norm": 6.830610072938725e-05, + "learning_rate": 0.0008436148538306099, + "loss": 0.2706, + "num_input_tokens_seen": 29478392, + "step": 38650 + }, + { + "epoch": 80.36382536382537, + "grad_norm": 0.0003837009717244655, + "learning_rate": 0.0008373878411868041, + "loss": 0.2487, + "num_input_tokens_seen": 29482136, + "step": 38655 + }, + { + "epoch": 80.37422037422037, + "grad_norm": 0.00035995221696794033, + "learning_rate": 0.000831183831292287, + "loss": 0.2851, + "num_input_tokens_seen": 29485912, + "step": 38660 + }, + { + "epoch": 80.38461538461539, + "grad_norm": 0.0004193721688352525, + "learning_rate": 0.0008250028251037933, + "loss": 0.2735, + "num_input_tokens_seen": 29489784, + "step": 38665 + }, + { + "epoch": 80.39501039501039, + "grad_norm": 0.00027916085673496127, + "learning_rate": 0.0008188448235745271, + "loss": 0.2552, + "num_input_tokens_seen": 29493528, + "step": 38670 + }, + { + "epoch": 80.4054054054054, + "grad_norm": 0.0002521621354389936, + "learning_rate": 0.0008127098276541122, + "loss": 0.2529, + "num_input_tokens_seen": 29497272, + "step": 38675 + }, + { + "epoch": 80.41580041580042, + "grad_norm": 0.00043592872680164874, + "learning_rate": 0.0008065978382886418, + "loss": 0.2674, + "num_input_tokens_seen": 29501144, + "step": 38680 + }, + { + "epoch": 80.42619542619542, + "grad_norm": 0.00041409360710531473, + "learning_rate": 0.0008005088564206785, + "loss": 0.2435, + "num_input_tokens_seen": 29504888, + "step": 38685 + }, + { + "epoch": 80.43659043659044, + "grad_norm": 0.0005598809220828116, + "learning_rate": 0.0007944428829891881, + "loss": 0.2531, + "num_input_tokens_seen": 29508600, + "step": 38690 + }, + { + "epoch": 80.44698544698545, + "grad_norm": 0.00016065462841652334, + "learning_rate": 0.0007883999189296386, + "loss": 0.2892, + "num_input_tokens_seen": 29512408, + "step": 38695 + }, + { + "epoch": 80.45738045738045, + "grad_norm": 0.00024615335860289633, + "learning_rate": 0.0007823799651739515, + "loss": 0.2691, + "num_input_tokens_seen": 29516216, + "step": 38700 + }, + { + "epoch": 80.46777546777547, + "grad_norm": 0.0003562385099940002, + "learning_rate": 0.0007763830226504509, + "loss": 0.2625, + "num_input_tokens_seen": 29520024, + "step": 38705 + }, + { + "epoch": 80.47817047817048, + "grad_norm": 0.0010006417287513614, + "learning_rate": 0.0007704090922839468, + "loss": 0.2752, + "num_input_tokens_seen": 29523896, + "step": 38710 + }, + { + "epoch": 80.48856548856548, + "grad_norm": 0.0006938848528079689, + "learning_rate": 0.0007644581749957025, + "loss": 0.2708, + "num_input_tokens_seen": 29527800, + "step": 38715 + }, + { + "epoch": 80.4989604989605, + "grad_norm": 0.006440659984946251, + "learning_rate": 0.000758530271703417, + "loss": 0.2778, + "num_input_tokens_seen": 29531544, + "step": 38720 + }, + { + "epoch": 80.50935550935552, + "grad_norm": 0.0003470568044576794, + "learning_rate": 0.0007526253833212426, + "loss": 0.2197, + "num_input_tokens_seen": 29535480, + "step": 38725 + }, + { + "epoch": 80.51975051975052, + "grad_norm": 0.00038155855145305395, + "learning_rate": 0.0007467435107598008, + "loss": 0.2622, + "num_input_tokens_seen": 29539352, + "step": 38730 + }, + { + "epoch": 80.53014553014553, + "grad_norm": 0.00013656876399181783, + "learning_rate": 0.0007408846549261328, + "loss": 0.2601, + "num_input_tokens_seen": 29543000, + "step": 38735 + }, + { + "epoch": 80.54054054054055, + "grad_norm": 0.0001753578835632652, + "learning_rate": 0.0007350488167237656, + "loss": 0.2473, + "num_input_tokens_seen": 29546808, + "step": 38740 + }, + { + "epoch": 80.55093555093555, + "grad_norm": 0.00041927240090444684, + "learning_rate": 0.0007292359970526629, + "loss": 0.27, + "num_input_tokens_seen": 29550680, + "step": 38745 + }, + { + "epoch": 80.56133056133056, + "grad_norm": 0.00014573319640476257, + "learning_rate": 0.0007234461968092076, + "loss": 0.2598, + "num_input_tokens_seen": 29554584, + "step": 38750 + }, + { + "epoch": 80.57172557172557, + "grad_norm": 0.00018213856674265116, + "learning_rate": 0.0007176794168862854, + "loss": 0.2791, + "num_input_tokens_seen": 29558488, + "step": 38755 + }, + { + "epoch": 80.58212058212058, + "grad_norm": 0.0014664718182757497, + "learning_rate": 0.000711935658173185, + "loss": 0.2509, + "num_input_tokens_seen": 29562328, + "step": 38760 + }, + { + "epoch": 80.5925155925156, + "grad_norm": 0.0008225443307310343, + "learning_rate": 0.0007062149215556812, + "loss": 0.282, + "num_input_tokens_seen": 29566168, + "step": 38765 + }, + { + "epoch": 80.6029106029106, + "grad_norm": 9.875069372355938e-05, + "learning_rate": 0.0007005172079159849, + "loss": 0.2358, + "num_input_tokens_seen": 29570040, + "step": 38770 + }, + { + "epoch": 80.61330561330561, + "grad_norm": 0.00019654349307529628, + "learning_rate": 0.0006948425181327267, + "loss": 0.2551, + "num_input_tokens_seen": 29573816, + "step": 38775 + }, + { + "epoch": 80.62370062370063, + "grad_norm": 0.0003359333495609462, + "learning_rate": 0.000689190853081073, + "loss": 0.2876, + "num_input_tokens_seen": 29577688, + "step": 38780 + }, + { + "epoch": 80.63409563409563, + "grad_norm": 0.0001950456207850948, + "learning_rate": 0.000683562213632527, + "loss": 0.2603, + "num_input_tokens_seen": 29581464, + "step": 38785 + }, + { + "epoch": 80.64449064449065, + "grad_norm": 0.00020145326561760157, + "learning_rate": 0.0006779566006551108, + "loss": 0.2641, + "num_input_tokens_seen": 29585240, + "step": 38790 + }, + { + "epoch": 80.65488565488566, + "grad_norm": 0.000315141020109877, + "learning_rate": 0.0006723740150132995, + "loss": 0.2534, + "num_input_tokens_seen": 29589144, + "step": 38795 + }, + { + "epoch": 80.66528066528066, + "grad_norm": 0.0003175473539158702, + "learning_rate": 0.0006668144575679713, + "loss": 0.2627, + "num_input_tokens_seen": 29592888, + "step": 38800 + }, + { + "epoch": 80.66528066528066, + "eval_loss": 0.24905961751937866, + "eval_runtime": 13.37, + "eval_samples_per_second": 64.024, + "eval_steps_per_second": 16.006, + "num_input_tokens_seen": 29592888, + "step": 38800 + }, + { + "epoch": 80.67567567567568, + "grad_norm": 0.0004954852629452944, + "learning_rate": 0.0006612779291765069, + "loss": 0.2892, + "num_input_tokens_seen": 29596696, + "step": 38805 + }, + { + "epoch": 80.68607068607069, + "grad_norm": 0.0008529546903446317, + "learning_rate": 0.0006557644306926736, + "loss": 0.2736, + "num_input_tokens_seen": 29600376, + "step": 38810 + }, + { + "epoch": 80.6964656964657, + "grad_norm": 0.0004440688935574144, + "learning_rate": 0.0006502739629667575, + "loss": 0.2727, + "num_input_tokens_seen": 29604216, + "step": 38815 + }, + { + "epoch": 80.70686070686071, + "grad_norm": 0.0007419568719342351, + "learning_rate": 0.0006448065268454317, + "loss": 0.2757, + "num_input_tokens_seen": 29608056, + "step": 38820 + }, + { + "epoch": 80.71725571725571, + "grad_norm": 9.028362546814606e-05, + "learning_rate": 0.0006393621231718549, + "loss": 0.2662, + "num_input_tokens_seen": 29611896, + "step": 38825 + }, + { + "epoch": 80.72765072765073, + "grad_norm": 0.0003431664081290364, + "learning_rate": 0.0006339407527856389, + "loss": 0.267, + "num_input_tokens_seen": 29615768, + "step": 38830 + }, + { + "epoch": 80.73804573804574, + "grad_norm": 0.0011426162673160434, + "learning_rate": 0.0006285424165227982, + "loss": 0.2522, + "num_input_tokens_seen": 29619768, + "step": 38835 + }, + { + "epoch": 80.74844074844074, + "grad_norm": 0.00010065758397104219, + "learning_rate": 0.0006231671152158169, + "loss": 0.2621, + "num_input_tokens_seen": 29623672, + "step": 38840 + }, + { + "epoch": 80.75883575883576, + "grad_norm": 0.0007315367693081498, + "learning_rate": 0.0006178148496936819, + "loss": 0.2634, + "num_input_tokens_seen": 29627544, + "step": 38845 + }, + { + "epoch": 80.76923076923077, + "grad_norm": 7.239534897962585e-05, + "learning_rate": 0.000612485620781733, + "loss": 0.2523, + "num_input_tokens_seen": 29631416, + "step": 38850 + }, + { + "epoch": 80.77962577962577, + "grad_norm": 0.00030909874476492405, + "learning_rate": 0.0006071794293018296, + "loss": 0.2581, + "num_input_tokens_seen": 29635096, + "step": 38855 + }, + { + "epoch": 80.79002079002079, + "grad_norm": 0.000763608724810183, + "learning_rate": 0.0006018962760722501, + "loss": 0.2755, + "num_input_tokens_seen": 29638936, + "step": 38860 + }, + { + "epoch": 80.8004158004158, + "grad_norm": 0.0003426999901421368, + "learning_rate": 0.0005966361619077098, + "loss": 0.2917, + "num_input_tokens_seen": 29642744, + "step": 38865 + }, + { + "epoch": 80.8108108108108, + "grad_norm": 0.0001761830208124593, + "learning_rate": 0.000591399087619393, + "loss": 0.2714, + "num_input_tokens_seen": 29646552, + "step": 38870 + }, + { + "epoch": 80.82120582120582, + "grad_norm": 8.107997564366087e-05, + "learning_rate": 0.0005861850540149371, + "loss": 0.2613, + "num_input_tokens_seen": 29650456, + "step": 38875 + }, + { + "epoch": 80.83160083160084, + "grad_norm": 0.000688837026245892, + "learning_rate": 0.0005809940618983822, + "loss": 0.2806, + "num_input_tokens_seen": 29654232, + "step": 38880 + }, + { + "epoch": 80.84199584199584, + "grad_norm": 0.0016362975584343076, + "learning_rate": 0.0005758261120702712, + "loss": 0.2579, + "num_input_tokens_seen": 29657976, + "step": 38885 + }, + { + "epoch": 80.85239085239085, + "grad_norm": 6.6026346758008e-05, + "learning_rate": 0.0005706812053275501, + "loss": 0.262, + "num_input_tokens_seen": 29661912, + "step": 38890 + }, + { + "epoch": 80.86278586278586, + "grad_norm": 0.00039615307468920946, + "learning_rate": 0.0005655593424636173, + "loss": 0.2275, + "num_input_tokens_seen": 29665656, + "step": 38895 + }, + { + "epoch": 80.87318087318087, + "grad_norm": 0.000295100238872692, + "learning_rate": 0.0005604605242683746, + "loss": 0.2479, + "num_input_tokens_seen": 29669528, + "step": 38900 + }, + { + "epoch": 80.88357588357589, + "grad_norm": 0.0003159618645440787, + "learning_rate": 0.0005553847515280596, + "loss": 0.2704, + "num_input_tokens_seen": 29673432, + "step": 38905 + }, + { + "epoch": 80.89397089397089, + "grad_norm": 0.00014353074948303401, + "learning_rate": 0.0005503320250254795, + "loss": 0.2586, + "num_input_tokens_seen": 29677464, + "step": 38910 + }, + { + "epoch": 80.9043659043659, + "grad_norm": 0.00014917370572220534, + "learning_rate": 0.0005453023455397943, + "loss": 0.2588, + "num_input_tokens_seen": 29681304, + "step": 38915 + }, + { + "epoch": 80.91476091476092, + "grad_norm": 0.00016284633602481335, + "learning_rate": 0.0005402957138466502, + "loss": 0.2421, + "num_input_tokens_seen": 29685016, + "step": 38920 + }, + { + "epoch": 80.92515592515592, + "grad_norm": 0.0003256556810811162, + "learning_rate": 0.0005353121307181463, + "loss": 0.2558, + "num_input_tokens_seen": 29688696, + "step": 38925 + }, + { + "epoch": 80.93555093555094, + "grad_norm": 0.00019272591453045607, + "learning_rate": 0.0005303515969227845, + "loss": 0.2713, + "num_input_tokens_seen": 29692440, + "step": 38930 + }, + { + "epoch": 80.94594594594595, + "grad_norm": 0.00023034511832520366, + "learning_rate": 0.0005254141132255862, + "loss": 0.28, + "num_input_tokens_seen": 29696216, + "step": 38935 + }, + { + "epoch": 80.95634095634095, + "grad_norm": 0.000148714505485259, + "learning_rate": 0.0005204996803879258, + "loss": 0.2624, + "num_input_tokens_seen": 29700056, + "step": 38940 + }, + { + "epoch": 80.96673596673597, + "grad_norm": 0.0005001542740501463, + "learning_rate": 0.0005156082991676969, + "loss": 0.2642, + "num_input_tokens_seen": 29703832, + "step": 38945 + }, + { + "epoch": 80.97713097713098, + "grad_norm": 0.00046157496399246156, + "learning_rate": 0.0005107399703192127, + "loss": 0.2257, + "num_input_tokens_seen": 29707608, + "step": 38950 + }, + { + "epoch": 80.98752598752598, + "grad_norm": 0.000228060336667113, + "learning_rate": 0.0005058946945932063, + "loss": 0.2626, + "num_input_tokens_seen": 29711352, + "step": 38955 + }, + { + "epoch": 80.997920997921, + "grad_norm": 0.00011820188228739426, + "learning_rate": 0.0005010724727369131, + "loss": 0.2772, + "num_input_tokens_seen": 29715096, + "step": 38960 + }, + { + "epoch": 81.00831600831602, + "grad_norm": 0.0002430666791042313, + "learning_rate": 0.000496273305493955, + "loss": 0.2728, + "num_input_tokens_seen": 29718824, + "step": 38965 + }, + { + "epoch": 81.01871101871102, + "grad_norm": 0.0003784166183322668, + "learning_rate": 0.0004914971936044399, + "loss": 0.237, + "num_input_tokens_seen": 29722632, + "step": 38970 + }, + { + "epoch": 81.02910602910603, + "grad_norm": 0.0003129137621726841, + "learning_rate": 0.00048674413780491196, + "loss": 0.2859, + "num_input_tokens_seen": 29726408, + "step": 38975 + }, + { + "epoch": 81.03950103950103, + "grad_norm": 0.0002342287771170959, + "learning_rate": 0.0004820141388283183, + "loss": 0.2625, + "num_input_tokens_seen": 29730088, + "step": 38980 + }, + { + "epoch": 81.04989604989605, + "grad_norm": 0.0002921112172771245, + "learning_rate": 0.00047730719740410874, + "loss": 0.258, + "num_input_tokens_seen": 29733992, + "step": 38985 + }, + { + "epoch": 81.06029106029106, + "grad_norm": 0.0007997443899512291, + "learning_rate": 0.00047262331425816927, + "loss": 0.2657, + "num_input_tokens_seen": 29737736, + "step": 38990 + }, + { + "epoch": 81.07068607068607, + "grad_norm": 0.0005732898716814816, + "learning_rate": 0.00046796249011277213, + "loss": 0.251, + "num_input_tokens_seen": 29741512, + "step": 38995 + }, + { + "epoch": 81.08108108108108, + "grad_norm": 0.00032948559965007007, + "learning_rate": 0.00046332472568669236, + "loss": 0.2656, + "num_input_tokens_seen": 29745320, + "step": 39000 + }, + { + "epoch": 81.08108108108108, + "eval_loss": 0.24877241253852844, + "eval_runtime": 13.3787, + "eval_samples_per_second": 63.982, + "eval_steps_per_second": 15.996, + "num_input_tokens_seen": 29745320, + "step": 39000 + }, + { + "epoch": 81.0914760914761, + "grad_norm": 6.946392386453226e-05, + "learning_rate": 0.0004587100216951578, + "loss": 0.2713, + "num_input_tokens_seen": 29749032, + "step": 39005 + }, + { + "epoch": 81.1018711018711, + "grad_norm": 0.0005109703633934259, + "learning_rate": 0.00045411837884978265, + "loss": 0.2346, + "num_input_tokens_seen": 29752744, + "step": 39010 + }, + { + "epoch": 81.11226611226611, + "grad_norm": 0.0003009248757734895, + "learning_rate": 0.00044954979785865045, + "loss": 0.2862, + "num_input_tokens_seen": 29756680, + "step": 39015 + }, + { + "epoch": 81.12266112266113, + "grad_norm": 0.00025270512560382485, + "learning_rate": 0.00044500427942631426, + "loss": 0.2781, + "num_input_tokens_seen": 29760520, + "step": 39020 + }, + { + "epoch": 81.13305613305613, + "grad_norm": 0.0008549754857085645, + "learning_rate": 0.0004404818242537467, + "loss": 0.2735, + "num_input_tokens_seen": 29764232, + "step": 39025 + }, + { + "epoch": 81.14345114345114, + "grad_norm": 0.0007490808493457735, + "learning_rate": 0.00043598243303837324, + "loss": 0.2438, + "num_input_tokens_seen": 29767880, + "step": 39030 + }, + { + "epoch": 81.15384615384616, + "grad_norm": 0.0004002944042440504, + "learning_rate": 0.00043150610647403885, + "loss": 0.2641, + "num_input_tokens_seen": 29771720, + "step": 39035 + }, + { + "epoch": 81.16424116424116, + "grad_norm": 0.00028097417089156806, + "learning_rate": 0.00042705284525104134, + "loss": 0.2787, + "num_input_tokens_seen": 29775624, + "step": 39040 + }, + { + "epoch": 81.17463617463618, + "grad_norm": 0.00032751832623034716, + "learning_rate": 0.0004226226500561647, + "loss": 0.2444, + "num_input_tokens_seen": 29779400, + "step": 39045 + }, + { + "epoch": 81.18503118503118, + "grad_norm": 0.000718900584615767, + "learning_rate": 0.0004182155215725791, + "loss": 0.2363, + "num_input_tokens_seen": 29783368, + "step": 39050 + }, + { + "epoch": 81.1954261954262, + "grad_norm": 0.0004228925099596381, + "learning_rate": 0.00041383146047992424, + "loss": 0.2594, + "num_input_tokens_seen": 29787144, + "step": 39055 + }, + { + "epoch": 81.20582120582121, + "grad_norm": 0.0006319344392977655, + "learning_rate": 0.00040947046745427597, + "loss": 0.2745, + "num_input_tokens_seen": 29790984, + "step": 39060 + }, + { + "epoch": 81.21621621621621, + "grad_norm": 0.0003901736927218735, + "learning_rate": 0.00040513254316814625, + "loss": 0.2524, + "num_input_tokens_seen": 29794856, + "step": 39065 + }, + { + "epoch": 81.22661122661123, + "grad_norm": 0.00023066597350407392, + "learning_rate": 0.0004008176882905168, + "loss": 0.247, + "num_input_tokens_seen": 29798664, + "step": 39070 + }, + { + "epoch": 81.23700623700624, + "grad_norm": 0.00028419942827895284, + "learning_rate": 0.00039652590348677184, + "loss": 0.2797, + "num_input_tokens_seen": 29802344, + "step": 39075 + }, + { + "epoch": 81.24740124740124, + "grad_norm": 0.0009782575070858002, + "learning_rate": 0.00039225718941878206, + "loss": 0.2681, + "num_input_tokens_seen": 29806120, + "step": 39080 + }, + { + "epoch": 81.25779625779626, + "grad_norm": 0.00029170283232815564, + "learning_rate": 0.00038801154674480417, + "loss": 0.2804, + "num_input_tokens_seen": 29810024, + "step": 39085 + }, + { + "epoch": 81.26819126819127, + "grad_norm": 0.00010834466229425743, + "learning_rate": 0.00038378897611959784, + "loss": 0.2646, + "num_input_tokens_seen": 29813832, + "step": 39090 + }, + { + "epoch": 81.27858627858627, + "grad_norm": 0.00031477221637032926, + "learning_rate": 0.00037958947819430875, + "loss": 0.2546, + "num_input_tokens_seen": 29817800, + "step": 39095 + }, + { + "epoch": 81.28898128898129, + "grad_norm": 0.00011645103222690523, + "learning_rate": 0.0003754130536165856, + "loss": 0.2602, + "num_input_tokens_seen": 29821576, + "step": 39100 + }, + { + "epoch": 81.2993762993763, + "grad_norm": 0.00046451727394014597, + "learning_rate": 0.0003712597030304632, + "loss": 0.2324, + "num_input_tokens_seen": 29825512, + "step": 39105 + }, + { + "epoch": 81.3097713097713, + "grad_norm": 0.00027248976402916014, + "learning_rate": 0.00036712942707646247, + "loss": 0.2426, + "num_input_tokens_seen": 29829288, + "step": 39110 + }, + { + "epoch": 81.32016632016632, + "grad_norm": 0.00030916763353161514, + "learning_rate": 0.00036302222639149063, + "loss": 0.251, + "num_input_tokens_seen": 29833192, + "step": 39115 + }, + { + "epoch": 81.33056133056132, + "grad_norm": 0.00022757383703719825, + "learning_rate": 0.000358938101608941, + "loss": 0.2704, + "num_input_tokens_seen": 29836904, + "step": 39120 + }, + { + "epoch": 81.34095634095634, + "grad_norm": 0.00039262103382498026, + "learning_rate": 0.0003548770533586598, + "loss": 0.2542, + "num_input_tokens_seen": 29840744, + "step": 39125 + }, + { + "epoch": 81.35135135135135, + "grad_norm": 0.0001572659966768697, + "learning_rate": 0.0003508390822668961, + "loss": 0.2876, + "num_input_tokens_seen": 29844552, + "step": 39130 + }, + { + "epoch": 81.36174636174636, + "grad_norm": 0.0003366860037203878, + "learning_rate": 0.00034682418895633503, + "loss": 0.2551, + "num_input_tokens_seen": 29848424, + "step": 39135 + }, + { + "epoch": 81.37214137214137, + "grad_norm": 0.00012265951954759657, + "learning_rate": 0.0003428323740461647, + "loss": 0.2631, + "num_input_tokens_seen": 29852488, + "step": 39140 + }, + { + "epoch": 81.38253638253639, + "grad_norm": 0.00023194070672616363, + "learning_rate": 0.00033886363815194276, + "loss": 0.2586, + "num_input_tokens_seen": 29856328, + "step": 39145 + }, + { + "epoch": 81.39293139293139, + "grad_norm": 0.00025766942417249084, + "learning_rate": 0.0003349179818857129, + "loss": 0.2609, + "num_input_tokens_seen": 29860232, + "step": 39150 + }, + { + "epoch": 81.4033264033264, + "grad_norm": 0.0005640723975375295, + "learning_rate": 0.0003309954058559383, + "loss": 0.2825, + "num_input_tokens_seen": 29864200, + "step": 39155 + }, + { + "epoch": 81.41372141372142, + "grad_norm": 0.00011428180005168542, + "learning_rate": 0.0003270959106675186, + "loss": 0.2418, + "num_input_tokens_seen": 29868072, + "step": 39160 + }, + { + "epoch": 81.42411642411642, + "grad_norm": 0.0005130151403136551, + "learning_rate": 0.0003232194969218227, + "loss": 0.2886, + "num_input_tokens_seen": 29871976, + "step": 39165 + }, + { + "epoch": 81.43451143451144, + "grad_norm": 0.0002611126401461661, + "learning_rate": 0.00031936616521663905, + "loss": 0.2711, + "num_input_tokens_seen": 29875816, + "step": 39170 + }, + { + "epoch": 81.44490644490645, + "grad_norm": 0.0007460659253410995, + "learning_rate": 0.00031553591614619236, + "loss": 0.2595, + "num_input_tokens_seen": 29879560, + "step": 39175 + }, + { + "epoch": 81.45530145530145, + "grad_norm": 0.00036192481638863683, + "learning_rate": 0.00031172875030117676, + "loss": 0.2557, + "num_input_tokens_seen": 29883176, + "step": 39180 + }, + { + "epoch": 81.46569646569647, + "grad_norm": 0.00012861716095358133, + "learning_rate": 0.0003079446682686726, + "loss": 0.2327, + "num_input_tokens_seen": 29887176, + "step": 39185 + }, + { + "epoch": 81.47609147609148, + "grad_norm": 0.00029900582740083337, + "learning_rate": 0.0003041836706322465, + "loss": 0.2636, + "num_input_tokens_seen": 29890952, + "step": 39190 + }, + { + "epoch": 81.48648648648648, + "grad_norm": 0.0001206764645758085, + "learning_rate": 0.0003004457579719011, + "loss": 0.2593, + "num_input_tokens_seen": 29894792, + "step": 39195 + }, + { + "epoch": 81.4968814968815, + "grad_norm": 0.0017916184151545167, + "learning_rate": 0.00029673093086405867, + "loss": 0.2409, + "num_input_tokens_seen": 29898600, + "step": 39200 + }, + { + "epoch": 81.4968814968815, + "eval_loss": 0.24881578981876373, + "eval_runtime": 13.3805, + "eval_samples_per_second": 63.974, + "eval_steps_per_second": 15.993, + "num_input_tokens_seen": 29898600, + "step": 39200 + }, + { + "epoch": 81.5072765072765, + "grad_norm": 0.0003137363528367132, + "learning_rate": 0.00029303918988159426, + "loss": 0.2501, + "num_input_tokens_seen": 29902440, + "step": 39205 + }, + { + "epoch": 81.51767151767152, + "grad_norm": 0.0006595077575184405, + "learning_rate": 0.0002893705355938192, + "loss": 0.2673, + "num_input_tokens_seen": 29906376, + "step": 39210 + }, + { + "epoch": 81.52806652806653, + "grad_norm": 0.00019991387671325356, + "learning_rate": 0.0002857249685664975, + "loss": 0.263, + "num_input_tokens_seen": 29910152, + "step": 39215 + }, + { + "epoch": 81.53846153846153, + "grad_norm": 0.0005949345068074763, + "learning_rate": 0.0002821024893618129, + "loss": 0.2453, + "num_input_tokens_seen": 29913896, + "step": 39220 + }, + { + "epoch": 81.54885654885655, + "grad_norm": 0.0002964503364637494, + "learning_rate": 0.0002785030985383852, + "loss": 0.2553, + "num_input_tokens_seen": 29917672, + "step": 39225 + }, + { + "epoch": 81.55925155925156, + "grad_norm": 0.0003674213367048651, + "learning_rate": 0.00027492679665130356, + "loss": 0.2748, + "num_input_tokens_seen": 29921448, + "step": 39230 + }, + { + "epoch": 81.56964656964657, + "grad_norm": 0.0002499967231415212, + "learning_rate": 0.000271373584252077, + "loss": 0.2627, + "num_input_tokens_seen": 29925288, + "step": 39235 + }, + { + "epoch": 81.58004158004158, + "grad_norm": 0.000379693548893556, + "learning_rate": 0.00026784346188865046, + "loss": 0.2592, + "num_input_tokens_seen": 29929192, + "step": 39240 + }, + { + "epoch": 81.5904365904366, + "grad_norm": 0.0003246490960009396, + "learning_rate": 0.0002643364301054218, + "loss": 0.2814, + "num_input_tokens_seen": 29932872, + "step": 39245 + }, + { + "epoch": 81.6008316008316, + "grad_norm": 0.0006923344335518777, + "learning_rate": 0.0002608524894431918, + "loss": 0.2485, + "num_input_tokens_seen": 29936872, + "step": 39250 + }, + { + "epoch": 81.61122661122661, + "grad_norm": 0.00026829156558960676, + "learning_rate": 0.000257391640439264, + "loss": 0.2476, + "num_input_tokens_seen": 29940552, + "step": 39255 + }, + { + "epoch": 81.62162162162163, + "grad_norm": 3.881260272464715e-05, + "learning_rate": 0.00025395388362732806, + "loss": 0.2798, + "num_input_tokens_seen": 29944488, + "step": 39260 + }, + { + "epoch": 81.63201663201663, + "grad_norm": 0.0017335637239739299, + "learning_rate": 0.00025053921953751, + "loss": 0.2657, + "num_input_tokens_seen": 29948296, + "step": 39265 + }, + { + "epoch": 81.64241164241164, + "grad_norm": 0.0003677213389892131, + "learning_rate": 0.00024714764869643855, + "loss": 0.2709, + "num_input_tokens_seen": 29951912, + "step": 39270 + }, + { + "epoch": 81.65280665280665, + "grad_norm": 0.0005958377732895315, + "learning_rate": 0.0002437791716270954, + "loss": 0.2657, + "num_input_tokens_seen": 29955592, + "step": 39275 + }, + { + "epoch": 81.66320166320166, + "grad_norm": 0.00014409460709430277, + "learning_rate": 0.00024043378884896493, + "loss": 0.2596, + "num_input_tokens_seen": 29959528, + "step": 39280 + }, + { + "epoch": 81.67359667359668, + "grad_norm": 0.0011817305348813534, + "learning_rate": 0.00023711150087793453, + "loss": 0.2554, + "num_input_tokens_seen": 29963304, + "step": 39285 + }, + { + "epoch": 81.68399168399168, + "grad_norm": 0.0005550046917051077, + "learning_rate": 0.000233812308226361, + "loss": 0.2582, + "num_input_tokens_seen": 29967176, + "step": 39290 + }, + { + "epoch": 81.6943866943867, + "grad_norm": 0.001390982884913683, + "learning_rate": 0.00023053621140300406, + "loss": 0.2811, + "num_input_tokens_seen": 29970824, + "step": 39295 + }, + { + "epoch": 81.70478170478171, + "grad_norm": 0.00029024569084867835, + "learning_rate": 0.00022728321091307623, + "loss": 0.234, + "num_input_tokens_seen": 29974568, + "step": 39300 + }, + { + "epoch": 81.71517671517671, + "grad_norm": 0.00040963542414829135, + "learning_rate": 0.0002240533072582429, + "loss": 0.246, + "num_input_tokens_seen": 29978536, + "step": 39305 + }, + { + "epoch": 81.72557172557173, + "grad_norm": 0.00022900060866959393, + "learning_rate": 0.00022084650093658897, + "loss": 0.2893, + "num_input_tokens_seen": 29982344, + "step": 39310 + }, + { + "epoch": 81.73596673596674, + "grad_norm": 0.0003072502149734646, + "learning_rate": 0.0002176627924426522, + "loss": 0.2633, + "num_input_tokens_seen": 29986184, + "step": 39315 + }, + { + "epoch": 81.74636174636174, + "grad_norm": 0.0001275045215152204, + "learning_rate": 0.0002145021822673898, + "loss": 0.2463, + "num_input_tokens_seen": 29990056, + "step": 39320 + }, + { + "epoch": 81.75675675675676, + "grad_norm": 0.0003510946989990771, + "learning_rate": 0.00021136467089822862, + "loss": 0.2557, + "num_input_tokens_seen": 29993768, + "step": 39325 + }, + { + "epoch": 81.76715176715177, + "grad_norm": 0.0006697809440083802, + "learning_rate": 0.00020825025881898162, + "loss": 0.2446, + "num_input_tokens_seen": 29997768, + "step": 39330 + }, + { + "epoch": 81.77754677754677, + "grad_norm": 0.0002201414608862251, + "learning_rate": 0.0002051589465099479, + "loss": 0.2779, + "num_input_tokens_seen": 30001416, + "step": 39335 + }, + { + "epoch": 81.78794178794179, + "grad_norm": 0.0005062674172222614, + "learning_rate": 0.0002020907344478462, + "loss": 0.2636, + "num_input_tokens_seen": 30005288, + "step": 39340 + }, + { + "epoch": 81.7983367983368, + "grad_norm": 0.00018497595738153905, + "learning_rate": 0.0001990456231058313, + "loss": 0.2494, + "num_input_tokens_seen": 30009064, + "step": 39345 + }, + { + "epoch": 81.8087318087318, + "grad_norm": 0.00185938342474401, + "learning_rate": 0.00019602361295349423, + "loss": 0.2766, + "num_input_tokens_seen": 30012904, + "step": 39350 + }, + { + "epoch": 81.81912681912682, + "grad_norm": 0.00033537307172082365, + "learning_rate": 0.0001930247044568789, + "loss": 0.2776, + "num_input_tokens_seen": 30016648, + "step": 39355 + }, + { + "epoch": 81.82952182952182, + "grad_norm": 0.0001350029488094151, + "learning_rate": 0.00019004889807843205, + "loss": 0.2658, + "num_input_tokens_seen": 30020456, + "step": 39360 + }, + { + "epoch": 81.83991683991684, + "grad_norm": 0.00029065311537124217, + "learning_rate": 0.00018709619427708656, + "loss": 0.2825, + "num_input_tokens_seen": 30024296, + "step": 39365 + }, + { + "epoch": 81.85031185031185, + "grad_norm": 0.0003413350787013769, + "learning_rate": 0.00018416659350817822, + "loss": 0.2782, + "num_input_tokens_seen": 30028136, + "step": 39370 + }, + { + "epoch": 81.86070686070686, + "grad_norm": 8.406373672187328e-05, + "learning_rate": 0.00018126009622346229, + "loss": 0.2638, + "num_input_tokens_seen": 30031944, + "step": 39375 + }, + { + "epoch": 81.87110187110187, + "grad_norm": 0.0005150201031938195, + "learning_rate": 0.00017837670287119687, + "loss": 0.2387, + "num_input_tokens_seen": 30035720, + "step": 39380 + }, + { + "epoch": 81.88149688149689, + "grad_norm": 0.0007320127333514392, + "learning_rate": 0.00017551641389602633, + "loss": 0.2398, + "num_input_tokens_seen": 30039400, + "step": 39385 + }, + { + "epoch": 81.89189189189189, + "grad_norm": 0.0003280183009337634, + "learning_rate": 0.00017267922973903115, + "loss": 0.2665, + "num_input_tokens_seen": 30043176, + "step": 39390 + }, + { + "epoch": 81.9022869022869, + "grad_norm": 0.00015193450963124633, + "learning_rate": 0.00016986515083774467, + "loss": 0.2815, + "num_input_tokens_seen": 30046824, + "step": 39395 + }, + { + "epoch": 81.91268191268192, + "grad_norm": 0.0002811957092490047, + "learning_rate": 0.00016707417762611975, + "loss": 0.2619, + "num_input_tokens_seen": 30050504, + "step": 39400 + }, + { + "epoch": 81.91268191268192, + "eval_loss": 0.24916960299015045, + "eval_runtime": 13.3732, + "eval_samples_per_second": 64.008, + "eval_steps_per_second": 16.002, + "num_input_tokens_seen": 30050504, + "step": 39400 + }, + { + "epoch": 81.92307692307692, + "grad_norm": 7.521611405536532e-05, + "learning_rate": 0.00016430631053459543, + "loss": 0.2895, + "num_input_tokens_seen": 30054440, + "step": 39405 + }, + { + "epoch": 81.93347193347194, + "grad_norm": 0.00027287990087643266, + "learning_rate": 0.0001615615499899803, + "loss": 0.2631, + "num_input_tokens_seen": 30058312, + "step": 39410 + }, + { + "epoch": 81.94386694386695, + "grad_norm": 0.002547452226281166, + "learning_rate": 0.00015883989641556905, + "loss": 0.28, + "num_input_tokens_seen": 30062120, + "step": 39415 + }, + { + "epoch": 81.95426195426195, + "grad_norm": 0.00022000692842993885, + "learning_rate": 0.00015614135023105934, + "loss": 0.289, + "num_input_tokens_seen": 30065992, + "step": 39420 + }, + { + "epoch": 81.96465696465697, + "grad_norm": 0.0005345044191926718, + "learning_rate": 0.00015346591185261827, + "loss": 0.2729, + "num_input_tokens_seen": 30069896, + "step": 39425 + }, + { + "epoch": 81.97505197505197, + "grad_norm": 0.0002813942264765501, + "learning_rate": 0.00015081358169281576, + "loss": 0.2506, + "num_input_tokens_seen": 30073672, + "step": 39430 + }, + { + "epoch": 81.98544698544698, + "grad_norm": 0.00011182101297890767, + "learning_rate": 0.00014818436016069135, + "loss": 0.2769, + "num_input_tokens_seen": 30077512, + "step": 39435 + }, + { + "epoch": 81.995841995842, + "grad_norm": 0.0004064147942699492, + "learning_rate": 0.00014557824766168735, + "loss": 0.2416, + "num_input_tokens_seen": 30081416, + "step": 39440 + }, + { + "epoch": 82.006237006237, + "grad_norm": 0.001103613874875009, + "learning_rate": 0.00014299524459769896, + "loss": 0.2466, + "num_input_tokens_seen": 30085304, + "step": 39445 + }, + { + "epoch": 82.01663201663202, + "grad_norm": 0.0007096627959981561, + "learning_rate": 0.0001404353513670742, + "loss": 0.2684, + "num_input_tokens_seen": 30089080, + "step": 39450 + }, + { + "epoch": 82.02702702702703, + "grad_norm": 0.00032989992178045213, + "learning_rate": 0.0001378985683645806, + "loss": 0.2713, + "num_input_tokens_seen": 30092920, + "step": 39455 + }, + { + "epoch": 82.03742203742203, + "grad_norm": 0.0001851690758485347, + "learning_rate": 0.0001353848959813886, + "loss": 0.2549, + "num_input_tokens_seen": 30096728, + "step": 39460 + }, + { + "epoch": 82.04781704781705, + "grad_norm": 0.0006134851137176156, + "learning_rate": 0.00013289433460517142, + "loss": 0.2332, + "num_input_tokens_seen": 30100568, + "step": 39465 + }, + { + "epoch": 82.05821205821206, + "grad_norm": 0.0002560316352173686, + "learning_rate": 0.00013042688462000518, + "loss": 0.2816, + "num_input_tokens_seen": 30104376, + "step": 39470 + }, + { + "epoch": 82.06860706860707, + "grad_norm": 0.0002895807265304029, + "learning_rate": 0.0001279825464063855, + "loss": 0.2663, + "num_input_tokens_seen": 30108216, + "step": 39475 + }, + { + "epoch": 82.07900207900208, + "grad_norm": 0.0010379767045378685, + "learning_rate": 0.00012556132034126087, + "loss": 0.239, + "num_input_tokens_seen": 30112088, + "step": 39480 + }, + { + "epoch": 82.0893970893971, + "grad_norm": 0.00022599281510338187, + "learning_rate": 0.0001231632067980326, + "loss": 0.2569, + "num_input_tokens_seen": 30115896, + "step": 39485 + }, + { + "epoch": 82.0997920997921, + "grad_norm": 5.704965224140324e-05, + "learning_rate": 0.00012078820614650486, + "loss": 0.2572, + "num_input_tokens_seen": 30119576, + "step": 39490 + }, + { + "epoch": 82.11018711018711, + "grad_norm": 0.00031116203172132373, + "learning_rate": 0.00011843631875291804, + "loss": 0.2626, + "num_input_tokens_seen": 30123352, + "step": 39495 + }, + { + "epoch": 82.12058212058211, + "grad_norm": 0.00028915132861584425, + "learning_rate": 0.00011610754497999863, + "loss": 0.2918, + "num_input_tokens_seen": 30127192, + "step": 39500 + }, + { + "epoch": 82.13097713097713, + "grad_norm": 0.00018232085858471692, + "learning_rate": 0.0001138018851868594, + "loss": 0.276, + "num_input_tokens_seen": 30131096, + "step": 39505 + }, + { + "epoch": 82.14137214137214, + "grad_norm": 0.0005327480612322688, + "learning_rate": 0.0001115193397290326, + "loss": 0.2566, + "num_input_tokens_seen": 30134968, + "step": 39510 + }, + { + "epoch": 82.15176715176715, + "grad_norm": 0.00025095490855164826, + "learning_rate": 0.00010925990895856996, + "loss": 0.2756, + "num_input_tokens_seen": 30138808, + "step": 39515 + }, + { + "epoch": 82.16216216216216, + "grad_norm": 0.0001093059909180738, + "learning_rate": 0.00010702359322385946, + "loss": 0.268, + "num_input_tokens_seen": 30142552, + "step": 39520 + }, + { + "epoch": 82.17255717255718, + "grad_norm": 0.0013536772457882762, + "learning_rate": 0.00010481039286977523, + "loss": 0.2702, + "num_input_tokens_seen": 30146360, + "step": 39525 + }, + { + "epoch": 82.18295218295218, + "grad_norm": 0.0005254528950899839, + "learning_rate": 0.00010262030823764423, + "loss": 0.2632, + "num_input_tokens_seen": 30150008, + "step": 39530 + }, + { + "epoch": 82.1933471933472, + "grad_norm": 0.0001025453457259573, + "learning_rate": 0.00010045333966517966, + "loss": 0.2697, + "num_input_tokens_seen": 30153752, + "step": 39535 + }, + { + "epoch": 82.20374220374221, + "grad_norm": 0.0001426387025276199, + "learning_rate": 9.83094874865642e-05, + "loss": 0.2635, + "num_input_tokens_seen": 30157624, + "step": 39540 + }, + { + "epoch": 82.21413721413721, + "grad_norm": 0.00043559836922213435, + "learning_rate": 9.618875203241672e-05, + "loss": 0.2667, + "num_input_tokens_seen": 30161496, + "step": 39545 + }, + { + "epoch": 82.22453222453223, + "grad_norm": 0.00016597587091382593, + "learning_rate": 9.409113362977561e-05, + "loss": 0.2545, + "num_input_tokens_seen": 30165368, + "step": 39550 + }, + { + "epoch": 82.23492723492724, + "grad_norm": 0.00024538402794860303, + "learning_rate": 9.20166326020988e-05, + "loss": 0.2807, + "num_input_tokens_seen": 30169112, + "step": 39555 + }, + { + "epoch": 82.24532224532224, + "grad_norm": 0.00018476434343028814, + "learning_rate": 8.996524926933035e-05, + "loss": 0.2729, + "num_input_tokens_seen": 30172888, + "step": 39560 + }, + { + "epoch": 82.25571725571726, + "grad_norm": 8.792390144662932e-05, + "learning_rate": 8.793698394781723e-05, + "loss": 0.2726, + "num_input_tokens_seen": 30176792, + "step": 39565 + }, + { + "epoch": 82.26611226611226, + "grad_norm": 0.00031042980845086277, + "learning_rate": 8.593183695030926e-05, + "loss": 0.2762, + "num_input_tokens_seen": 30180664, + "step": 39570 + }, + { + "epoch": 82.27650727650727, + "grad_norm": 0.0011632252717390656, + "learning_rate": 8.39498085860757e-05, + "loss": 0.2649, + "num_input_tokens_seen": 30184536, + "step": 39575 + }, + { + "epoch": 82.28690228690229, + "grad_norm": 0.00014371753786690533, + "learning_rate": 8.199089916072211e-05, + "loss": 0.2467, + "num_input_tokens_seen": 30188504, + "step": 39580 + }, + { + "epoch": 82.29729729729729, + "grad_norm": 0.0008486981969326735, + "learning_rate": 8.005510897637346e-05, + "loss": 0.2244, + "num_input_tokens_seen": 30192280, + "step": 39585 + }, + { + "epoch": 82.3076923076923, + "grad_norm": 0.0001485886168666184, + "learning_rate": 7.8142438331541e-05, + "loss": 0.2695, + "num_input_tokens_seen": 30196056, + "step": 39590 + }, + { + "epoch": 82.31808731808732, + "grad_norm": 0.00016745635366532952, + "learning_rate": 7.625288752117209e-05, + "loss": 0.2598, + "num_input_tokens_seen": 30199768, + "step": 39595 + }, + { + "epoch": 82.32848232848232, + "grad_norm": 0.00016985958791337907, + "learning_rate": 7.4386456836667e-05, + "loss": 0.2635, + "num_input_tokens_seen": 30203576, + "step": 39600 + }, + { + "epoch": 82.32848232848232, + "eval_loss": 0.2492460459470749, + "eval_runtime": 13.3665, + "eval_samples_per_second": 64.041, + "eval_steps_per_second": 16.01, + "num_input_tokens_seen": 30203576, + "step": 39600 + }, + { + "epoch": 82.33887733887734, + "grad_norm": 0.0007666294113732874, + "learning_rate": 7.254314656586214e-05, + "loss": 0.272, + "num_input_tokens_seen": 30207512, + "step": 39605 + }, + { + "epoch": 82.34927234927235, + "grad_norm": 0.0005155249964445829, + "learning_rate": 7.07229569929968e-05, + "loss": 0.3037, + "num_input_tokens_seen": 30211384, + "step": 39610 + }, + { + "epoch": 82.35966735966736, + "grad_norm": 0.0007861669873818755, + "learning_rate": 6.892588839879643e-05, + "loss": 0.2513, + "num_input_tokens_seen": 30215320, + "step": 39615 + }, + { + "epoch": 82.37006237006237, + "grad_norm": 0.00015402685676235706, + "learning_rate": 6.71519410603727e-05, + "loss": 0.2784, + "num_input_tokens_seen": 30219096, + "step": 39620 + }, + { + "epoch": 82.38045738045739, + "grad_norm": 0.002041741507127881, + "learning_rate": 6.540111525129011e-05, + "loss": 0.2416, + "num_input_tokens_seen": 30222872, + "step": 39625 + }, + { + "epoch": 82.39085239085239, + "grad_norm": 0.0005694655701518059, + "learning_rate": 6.367341124154934e-05, + "loss": 0.2776, + "num_input_tokens_seen": 30226680, + "step": 39630 + }, + { + "epoch": 82.4012474012474, + "grad_norm": 0.0002953947114292532, + "learning_rate": 6.19688292975873e-05, + "loss": 0.2416, + "num_input_tokens_seen": 30230616, + "step": 39635 + }, + { + "epoch": 82.41164241164242, + "grad_norm": 0.0005577355623245239, + "learning_rate": 6.0287369682260336e-05, + "loss": 0.2419, + "num_input_tokens_seen": 30234360, + "step": 39640 + }, + { + "epoch": 82.42203742203742, + "grad_norm": 0.00024115278210956603, + "learning_rate": 5.8629032654894384e-05, + "loss": 0.2723, + "num_input_tokens_seen": 30238040, + "step": 39645 + }, + { + "epoch": 82.43243243243244, + "grad_norm": 0.00030726828845217824, + "learning_rate": 5.699381847120155e-05, + "loss": 0.2567, + "num_input_tokens_seen": 30242008, + "step": 39650 + }, + { + "epoch": 82.44282744282744, + "grad_norm": 0.00019583117682486773, + "learning_rate": 5.5381727383380094e-05, + "loss": 0.2591, + "num_input_tokens_seen": 30245912, + "step": 39655 + }, + { + "epoch": 82.45322245322245, + "grad_norm": 0.0010089607676491141, + "learning_rate": 5.379275964001451e-05, + "loss": 0.27, + "num_input_tokens_seen": 30249752, + "step": 39660 + }, + { + "epoch": 82.46361746361747, + "grad_norm": 5.344822056940757e-05, + "learning_rate": 5.222691548614211e-05, + "loss": 0.2403, + "num_input_tokens_seen": 30253496, + "step": 39665 + }, + { + "epoch": 82.47401247401247, + "grad_norm": 0.0002855885250028223, + "learning_rate": 5.068419516323641e-05, + "loss": 0.2419, + "num_input_tokens_seen": 30257560, + "step": 39670 + }, + { + "epoch": 82.48440748440748, + "grad_norm": 0.0008575962274335325, + "learning_rate": 4.91645989092071e-05, + "loss": 0.2806, + "num_input_tokens_seen": 30261432, + "step": 39675 + }, + { + "epoch": 82.4948024948025, + "grad_norm": 0.00026951474137604237, + "learning_rate": 4.7668126958400056e-05, + "loss": 0.277, + "num_input_tokens_seen": 30265144, + "step": 39680 + }, + { + "epoch": 82.5051975051975, + "grad_norm": 0.00010369114897912368, + "learning_rate": 4.619477954159734e-05, + "loss": 0.2432, + "num_input_tokens_seen": 30269016, + "step": 39685 + }, + { + "epoch": 82.51559251559252, + "grad_norm": 0.001255469978787005, + "learning_rate": 4.4744556885983884e-05, + "loss": 0.2662, + "num_input_tokens_seen": 30272728, + "step": 39690 + }, + { + "epoch": 82.52598752598753, + "grad_norm": 0.0003735451609827578, + "learning_rate": 4.331745921523078e-05, + "loss": 0.2748, + "num_input_tokens_seen": 30276504, + "step": 39695 + }, + { + "epoch": 82.53638253638253, + "grad_norm": 0.0009013859089463949, + "learning_rate": 4.191348674937867e-05, + "loss": 0.2641, + "num_input_tokens_seen": 30280120, + "step": 39700 + }, + { + "epoch": 82.54677754677755, + "grad_norm": 0.00038640134152956307, + "learning_rate": 4.0532639704971006e-05, + "loss": 0.2635, + "num_input_tokens_seen": 30283832, + "step": 39705 + }, + { + "epoch": 82.55717255717256, + "grad_norm": 0.0008163599413819611, + "learning_rate": 3.917491829493747e-05, + "loss": 0.2658, + "num_input_tokens_seen": 30287736, + "step": 39710 + }, + { + "epoch": 82.56756756756756, + "grad_norm": 0.00023090663307812065, + "learning_rate": 3.78403227286439e-05, + "loss": 0.2508, + "num_input_tokens_seen": 30291576, + "step": 39715 + }, + { + "epoch": 82.57796257796258, + "grad_norm": 0.0002808485005516559, + "learning_rate": 3.652885321192567e-05, + "loss": 0.2736, + "num_input_tokens_seen": 30295416, + "step": 39720 + }, + { + "epoch": 82.58835758835758, + "grad_norm": 0.00019189031445421278, + "learning_rate": 3.524050994702099e-05, + "loss": 0.2641, + "num_input_tokens_seen": 30299352, + "step": 39725 + }, + { + "epoch": 82.5987525987526, + "grad_norm": 0.0009198287734761834, + "learning_rate": 3.3975293132604276e-05, + "loss": 0.272, + "num_input_tokens_seen": 30303192, + "step": 39730 + }, + { + "epoch": 82.60914760914761, + "grad_norm": 0.00016505520034115762, + "learning_rate": 3.2733202963786125e-05, + "loss": 0.2671, + "num_input_tokens_seen": 30306904, + "step": 39735 + }, + { + "epoch": 82.61954261954261, + "grad_norm": 0.0004372768453322351, + "learning_rate": 3.15142396321133e-05, + "loss": 0.2594, + "num_input_tokens_seen": 30310648, + "step": 39740 + }, + { + "epoch": 82.62993762993763, + "grad_norm": 0.0002436877111904323, + "learning_rate": 3.0318403325552132e-05, + "loss": 0.2457, + "num_input_tokens_seen": 30314488, + "step": 39745 + }, + { + "epoch": 82.64033264033264, + "grad_norm": 0.000996051006950438, + "learning_rate": 2.914569422855506e-05, + "loss": 0.2755, + "num_input_tokens_seen": 30318296, + "step": 39750 + }, + { + "epoch": 82.65072765072765, + "grad_norm": 0.0002123875601682812, + "learning_rate": 2.7996112521927462e-05, + "loss": 0.2716, + "num_input_tokens_seen": 30322232, + "step": 39755 + }, + { + "epoch": 82.66112266112266, + "grad_norm": 0.0004449710249900818, + "learning_rate": 2.68696583829775e-05, + "loss": 0.2587, + "num_input_tokens_seen": 30326040, + "step": 39760 + }, + { + "epoch": 82.67151767151768, + "grad_norm": 0.0004485077515710145, + "learning_rate": 2.576633198539957e-05, + "loss": 0.268, + "num_input_tokens_seen": 30329784, + "step": 39765 + }, + { + "epoch": 82.68191268191268, + "grad_norm": 0.00030305111431516707, + "learning_rate": 2.46861334993409e-05, + "loss": 0.2656, + "num_input_tokens_seen": 30333624, + "step": 39770 + }, + { + "epoch": 82.6923076923077, + "grad_norm": 6.662662781309336e-05, + "learning_rate": 2.3629063091384903e-05, + "loss": 0.2624, + "num_input_tokens_seen": 30337464, + "step": 39775 + }, + { + "epoch": 82.70270270270271, + "grad_norm": 0.000767660269048065, + "learning_rate": 2.2595120924567834e-05, + "loss": 0.2537, + "num_input_tokens_seen": 30341336, + "step": 39780 + }, + { + "epoch": 82.71309771309771, + "grad_norm": 0.0005912585183978081, + "learning_rate": 2.158430715829551e-05, + "loss": 0.2268, + "num_input_tokens_seen": 30345144, + "step": 39785 + }, + { + "epoch": 82.72349272349273, + "grad_norm": 0.0004818747693207115, + "learning_rate": 2.059662194849321e-05, + "loss": 0.2772, + "num_input_tokens_seen": 30348888, + "step": 39790 + }, + { + "epoch": 82.73388773388774, + "grad_norm": 0.00025212852051481605, + "learning_rate": 1.9632065447422463e-05, + "loss": 0.2722, + "num_input_tokens_seen": 30352568, + "step": 39795 + }, + { + "epoch": 82.74428274428274, + "grad_norm": 0.0006203068769536912, + "learning_rate": 1.8690637803880916e-05, + "loss": 0.2449, + "num_input_tokens_seen": 30356408, + "step": 39800 + }, + { + "epoch": 82.74428274428274, + "eval_loss": 0.2485324889421463, + "eval_runtime": 13.3747, + "eval_samples_per_second": 64.001, + "eval_steps_per_second": 16.0, + "num_input_tokens_seen": 30356408, + "step": 39800 + }, + { + "epoch": 82.75467775467776, + "grad_norm": 0.0005188753711991012, + "learning_rate": 1.7772339163019123e-05, + "loss": 0.2541, + "num_input_tokens_seen": 30360216, + "step": 39805 + }, + { + "epoch": 82.76507276507276, + "grad_norm": 0.0003956534492317587, + "learning_rate": 1.6877169666457138e-05, + "loss": 0.2616, + "num_input_tokens_seen": 30363896, + "step": 39810 + }, + { + "epoch": 82.77546777546777, + "grad_norm": 0.000282161810901016, + "learning_rate": 1.6005129452234532e-05, + "loss": 0.2592, + "num_input_tokens_seen": 30367672, + "step": 39815 + }, + { + "epoch": 82.78586278586279, + "grad_norm": 0.0009957605507224798, + "learning_rate": 1.5156218654843733e-05, + "loss": 0.2667, + "num_input_tokens_seen": 30371480, + "step": 39820 + }, + { + "epoch": 82.79625779625779, + "grad_norm": 0.00037115352461114526, + "learning_rate": 1.4330437405196683e-05, + "loss": 0.2775, + "num_input_tokens_seen": 30375256, + "step": 39825 + }, + { + "epoch": 82.8066528066528, + "grad_norm": 0.00042891796329058707, + "learning_rate": 1.352778583062486e-05, + "loss": 0.2972, + "num_input_tokens_seen": 30379160, + "step": 39830 + }, + { + "epoch": 82.81704781704782, + "grad_norm": 0.00018706402624957263, + "learning_rate": 1.2748264054929237e-05, + "loss": 0.2532, + "num_input_tokens_seen": 30383064, + "step": 39835 + }, + { + "epoch": 82.82744282744282, + "grad_norm": 0.0004231737111695111, + "learning_rate": 1.1991872198297004e-05, + "loss": 0.2205, + "num_input_tokens_seen": 30386872, + "step": 39840 + }, + { + "epoch": 82.83783783783784, + "grad_norm": 0.00023152027279138565, + "learning_rate": 1.1258610377384847e-05, + "loss": 0.2508, + "num_input_tokens_seen": 30390808, + "step": 39845 + }, + { + "epoch": 82.84823284823285, + "grad_norm": 0.00031264254357665777, + "learning_rate": 1.0548478705268982e-05, + "loss": 0.2721, + "num_input_tokens_seen": 30394488, + "step": 39850 + }, + { + "epoch": 82.85862785862786, + "grad_norm": 0.0008656340651214123, + "learning_rate": 9.86147729147846e-06, + "loss": 0.2369, + "num_input_tokens_seen": 30398168, + "step": 39855 + }, + { + "epoch": 82.86902286902287, + "grad_norm": 0.0005971190403215587, + "learning_rate": 9.197606241928557e-06, + "loss": 0.2426, + "num_input_tokens_seen": 30402072, + "step": 39860 + }, + { + "epoch": 82.87941787941789, + "grad_norm": 0.0005209440132603049, + "learning_rate": 8.556865659004042e-06, + "loss": 0.2922, + "num_input_tokens_seen": 30405944, + "step": 39865 + }, + { + "epoch": 82.88981288981289, + "grad_norm": 0.001681764144450426, + "learning_rate": 7.939255641525867e-06, + "loss": 0.2534, + "num_input_tokens_seen": 30409592, + "step": 39870 + }, + { + "epoch": 82.9002079002079, + "grad_norm": 0.00010485886741662398, + "learning_rate": 7.344776284751164e-06, + "loss": 0.2646, + "num_input_tokens_seen": 30413272, + "step": 39875 + }, + { + "epoch": 82.9106029106029, + "grad_norm": 0.00022654587519355118, + "learning_rate": 6.773427680323296e-06, + "loss": 0.2522, + "num_input_tokens_seen": 30417016, + "step": 39880 + }, + { + "epoch": 82.92099792099792, + "grad_norm": 0.0002863763365894556, + "learning_rate": 6.225209916355112e-06, + "loss": 0.2504, + "num_input_tokens_seen": 30420824, + "step": 39885 + }, + { + "epoch": 82.93139293139293, + "grad_norm": 0.0003275486524216831, + "learning_rate": 5.7001230774123e-06, + "loss": 0.274, + "num_input_tokens_seen": 30424664, + "step": 39890 + }, + { + "epoch": 82.94178794178794, + "grad_norm": 0.00010556641063885763, + "learning_rate": 5.198167244446772e-06, + "loss": 0.2699, + "num_input_tokens_seen": 30428632, + "step": 39895 + }, + { + "epoch": 82.95218295218295, + "grad_norm": 0.0003547542146407068, + "learning_rate": 4.71934249487993e-06, + "loss": 0.2715, + "num_input_tokens_seen": 30432344, + "step": 39900 + }, + { + "epoch": 82.96257796257797, + "grad_norm": 0.000539829139597714, + "learning_rate": 4.2636489025527075e-06, + "loss": 0.2503, + "num_input_tokens_seen": 30435928, + "step": 39905 + }, + { + "epoch": 82.97297297297297, + "grad_norm": 0.00030733918538317084, + "learning_rate": 3.831086537742223e-06, + "loss": 0.27, + "num_input_tokens_seen": 30439832, + "step": 39910 + }, + { + "epoch": 82.98336798336798, + "grad_norm": 5.470420001074672e-05, + "learning_rate": 3.4216554671451236e-06, + "loss": 0.2518, + "num_input_tokens_seen": 30443736, + "step": 39915 + }, + { + "epoch": 82.993762993763, + "grad_norm": 0.00018296035705134273, + "learning_rate": 3.035355753894242e-06, + "loss": 0.2552, + "num_input_tokens_seen": 30447576, + "step": 39920 + }, + { + "epoch": 83.004158004158, + "grad_norm": 0.00014197989366948605, + "learning_rate": 2.6721874575752477e-06, + "loss": 0.2584, + "num_input_tokens_seen": 30451280, + "step": 39925 + }, + { + "epoch": 83.01455301455302, + "grad_norm": 0.00011711655679391697, + "learning_rate": 2.3321506341933418e-06, + "loss": 0.2492, + "num_input_tokens_seen": 30455088, + "step": 39930 + }, + { + "epoch": 83.02494802494803, + "grad_norm": 0.0005199620500206947, + "learning_rate": 2.0152453361732546e-06, + "loss": 0.2897, + "num_input_tokens_seen": 30458960, + "step": 39935 + }, + { + "epoch": 83.03534303534303, + "grad_norm": 0.001643101335503161, + "learning_rate": 1.7214716123925554e-06, + "loss": 0.2632, + "num_input_tokens_seen": 30462640, + "step": 39940 + }, + { + "epoch": 83.04573804573805, + "grad_norm": 0.0007351182284764946, + "learning_rate": 1.4508295081649968e-06, + "loss": 0.2362, + "num_input_tokens_seen": 30466512, + "step": 39945 + }, + { + "epoch": 83.05613305613305, + "grad_norm": 0.00045212890836410224, + "learning_rate": 1.2033190652238623e-06, + "loss": 0.2562, + "num_input_tokens_seen": 30470224, + "step": 39950 + }, + { + "epoch": 83.06652806652806, + "grad_norm": 0.0005001293611712754, + "learning_rate": 9.78940321721966e-07, + "loss": 0.2591, + "num_input_tokens_seen": 30473872, + "step": 39955 + }, + { + "epoch": 83.07692307692308, + "grad_norm": 0.0013200159883126616, + "learning_rate": 7.776933122816132e-07, + "loss": 0.2608, + "num_input_tokens_seen": 30477680, + "step": 39960 + }, + { + "epoch": 83.08731808731808, + "grad_norm": 0.0002357655466767028, + "learning_rate": 5.99578067927986e-07, + "loss": 0.2651, + "num_input_tokens_seen": 30481456, + "step": 39965 + }, + { + "epoch": 83.0977130977131, + "grad_norm": 0.0002748349797911942, + "learning_rate": 4.445946161224512e-07, + "loss": 0.2543, + "num_input_tokens_seen": 30485360, + "step": 39970 + }, + { + "epoch": 83.10810810810811, + "grad_norm": 0.00012069835065631196, + "learning_rate": 3.127429807792126e-07, + "loss": 0.2502, + "num_input_tokens_seen": 30489200, + "step": 39975 + }, + { + "epoch": 83.11850311850311, + "grad_norm": 0.0007290628273040056, + "learning_rate": 2.040231822320049e-07, + "loss": 0.2554, + "num_input_tokens_seen": 30493136, + "step": 39980 + }, + { + "epoch": 83.12889812889813, + "grad_norm": 0.0005289536202326417, + "learning_rate": 1.1843523723409354e-07, + "loss": 0.2593, + "num_input_tokens_seen": 30496880, + "step": 39985 + }, + { + "epoch": 83.13929313929314, + "grad_norm": 0.000991914188489318, + "learning_rate": 5.597915897492811e-08, + "loss": 0.2623, + "num_input_tokens_seen": 30500656, + "step": 39990 + }, + { + "epoch": 83.14968814968815, + "grad_norm": 0.00026230866205878556, + "learning_rate": 1.6654957113448885e-08, + "loss": 0.2839, + "num_input_tokens_seen": 30504496, + "step": 39995 + }, + { + "epoch": 83.16008316008316, + "grad_norm": 0.00032065261621028185, + "learning_rate": 4.626377114735902e-10, + "loss": 0.2824, + "num_input_tokens_seen": 30508240, + "step": 40000 + }, + { + "epoch": 83.16008316008316, + "eval_loss": 0.2492801398038864, + "eval_runtime": 13.378, + "eval_samples_per_second": 63.986, + "eval_steps_per_second": 15.996, + "num_input_tokens_seen": 30508240, + "step": 40000 + }, + { + "epoch": 83.16008316008316, + "num_input_tokens_seen": 30508240, + "step": 40000, + "total_flos": 1.3737722584517837e+18, + "train_loss": 0.2765025964975357, + "train_runtime": 24417.94, + "train_samples_per_second": 26.21, + "train_steps_per_second": 1.638 + } + ], + "logging_steps": 5, + "max_steps": 40000, + "num_input_tokens_seen": 30508240, + "num_train_epochs": 84, + "save_steps": 200, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.3737722584517837e+18, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}