| { | |
| "best_metric": 1.601110577583313, | |
| "best_model_checkpoint": "outputs/checkpoint-13300", | |
| "epoch": 2.999832822513235, | |
| "eval_steps": 100, | |
| "global_step": 13458, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.022290331568682084, | |
| "grad_norm": 820153.75, | |
| "learning_rate": 6e-06, | |
| "loss": 3.4337, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.022290331568682084, | |
| "eval_loss": 2.219454050064087, | |
| "eval_runtime": 63.1076, | |
| "eval_samples_per_second": 211.829, | |
| "eval_steps_per_second": 13.247, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.04458066313736417, | |
| "grad_norm": 595487.125, | |
| "learning_rate": 1.2e-05, | |
| "loss": 2.4245, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.04458066313736417, | |
| "eval_loss": 1.938194990158081, | |
| "eval_runtime": 63.043, | |
| "eval_samples_per_second": 212.046, | |
| "eval_steps_per_second": 13.261, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.06687099470604625, | |
| "grad_norm": 613520.625, | |
| "learning_rate": 1.8e-05, | |
| "loss": 2.2722, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.06687099470604625, | |
| "eval_loss": 1.8768234252929688, | |
| "eval_runtime": 63.5969, | |
| "eval_samples_per_second": 210.199, | |
| "eval_steps_per_second": 13.145, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.08916132627472834, | |
| "grad_norm": 566655.875, | |
| "learning_rate": 2.4e-05, | |
| "loss": 2.1739, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.08916132627472834, | |
| "eval_loss": 1.8509334325790405, | |
| "eval_runtime": 63.6581, | |
| "eval_samples_per_second": 209.997, | |
| "eval_steps_per_second": 13.133, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.11145165784341042, | |
| "grad_norm": 618861.25, | |
| "learning_rate": 3e-05, | |
| "loss": 2.1388, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.11145165784341042, | |
| "eval_loss": 1.8238554000854492, | |
| "eval_runtime": 63.0746, | |
| "eval_samples_per_second": 211.939, | |
| "eval_steps_per_second": 13.254, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.1337419894120925, | |
| "grad_norm": 594465.5, | |
| "learning_rate": 2.9768482790554097e-05, | |
| "loss": 2.1116, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.1337419894120925, | |
| "eval_loss": 1.813140869140625, | |
| "eval_runtime": 63.2719, | |
| "eval_samples_per_second": 211.279, | |
| "eval_steps_per_second": 13.213, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.1560323209807746, | |
| "grad_norm": 528193.5625, | |
| "learning_rate": 2.9536965581108196e-05, | |
| "loss": 2.0772, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.1560323209807746, | |
| "eval_loss": 1.7901383638381958, | |
| "eval_runtime": 63.0937, | |
| "eval_samples_per_second": 211.876, | |
| "eval_steps_per_second": 13.25, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.17832265254945667, | |
| "grad_norm": 500452.375, | |
| "learning_rate": 2.9305448371662296e-05, | |
| "loss": 2.0526, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.17832265254945667, | |
| "eval_loss": 1.7764878273010254, | |
| "eval_runtime": 63.8628, | |
| "eval_samples_per_second": 209.324, | |
| "eval_steps_per_second": 13.091, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.20061298411813877, | |
| "grad_norm": 492339.875, | |
| "learning_rate": 2.9073931162216392e-05, | |
| "loss": 2.0419, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.20061298411813877, | |
| "eval_loss": 1.7795604467391968, | |
| "eval_runtime": 63.1552, | |
| "eval_samples_per_second": 211.669, | |
| "eval_steps_per_second": 13.237, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.22290331568682084, | |
| "grad_norm": 549590.1875, | |
| "learning_rate": 2.884241395277049e-05, | |
| "loss": 2.0305, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.22290331568682084, | |
| "eval_loss": 1.7810730934143066, | |
| "eval_runtime": 62.8059, | |
| "eval_samples_per_second": 212.846, | |
| "eval_steps_per_second": 13.311, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.24519364725550294, | |
| "grad_norm": 463470.46875, | |
| "learning_rate": 2.8610896743324588e-05, | |
| "loss": 2.0123, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.24519364725550294, | |
| "eval_loss": 1.739020586013794, | |
| "eval_runtime": 63.3157, | |
| "eval_samples_per_second": 211.133, | |
| "eval_steps_per_second": 13.204, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.267483978824185, | |
| "grad_norm": 483440.3125, | |
| "learning_rate": 2.8379379533878684e-05, | |
| "loss": 2.0051, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.267483978824185, | |
| "eval_loss": 1.754070520401001, | |
| "eval_runtime": 63.6661, | |
| "eval_samples_per_second": 209.97, | |
| "eval_steps_per_second": 13.131, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.2897743103928671, | |
| "grad_norm": 515932.6875, | |
| "learning_rate": 2.8147862324432783e-05, | |
| "loss": 2.0018, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.2897743103928671, | |
| "eval_loss": 1.7487035989761353, | |
| "eval_runtime": 63.3358, | |
| "eval_samples_per_second": 211.065, | |
| "eval_steps_per_second": 13.199, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.3120646419615492, | |
| "grad_norm": 510042.875, | |
| "learning_rate": 2.7916345114986883e-05, | |
| "loss": 1.9878, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.3120646419615492, | |
| "eval_loss": 1.7291030883789062, | |
| "eval_runtime": 63.1739, | |
| "eval_samples_per_second": 211.607, | |
| "eval_steps_per_second": 13.233, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.33435497353023125, | |
| "grad_norm": 508744.25, | |
| "learning_rate": 2.768482790554098e-05, | |
| "loss": 1.9885, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.33435497353023125, | |
| "eval_loss": 1.7247428894042969, | |
| "eval_runtime": 63.3733, | |
| "eval_samples_per_second": 210.941, | |
| "eval_steps_per_second": 13.192, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.35664530509891335, | |
| "grad_norm": 512775.4375, | |
| "learning_rate": 2.7453310696095075e-05, | |
| "loss": 1.9864, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.35664530509891335, | |
| "eval_loss": 1.7243812084197998, | |
| "eval_runtime": 63.6899, | |
| "eval_samples_per_second": 209.892, | |
| "eval_steps_per_second": 13.126, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.37893563666759544, | |
| "grad_norm": 480830.6875, | |
| "learning_rate": 2.7221793486649175e-05, | |
| "loss": 1.9735, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.37893563666759544, | |
| "eval_loss": 1.7244142293930054, | |
| "eval_runtime": 63.3501, | |
| "eval_samples_per_second": 211.018, | |
| "eval_steps_per_second": 13.196, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.40122596823627754, | |
| "grad_norm": 554714.125, | |
| "learning_rate": 2.6990276277203274e-05, | |
| "loss": 1.9756, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.40122596823627754, | |
| "eval_loss": 1.7170045375823975, | |
| "eval_runtime": 62.9859, | |
| "eval_samples_per_second": 212.238, | |
| "eval_steps_per_second": 13.273, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.4235162998049596, | |
| "grad_norm": 464572.84375, | |
| "learning_rate": 2.675875906775737e-05, | |
| "loss": 1.9538, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.4235162998049596, | |
| "eval_loss": 1.7204582691192627, | |
| "eval_runtime": 63.2678, | |
| "eval_samples_per_second": 211.292, | |
| "eval_steps_per_second": 13.214, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.4458066313736417, | |
| "grad_norm": 461049.8125, | |
| "learning_rate": 2.652724185831147e-05, | |
| "loss": 1.9519, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.4458066313736417, | |
| "eval_loss": 1.7141714096069336, | |
| "eval_runtime": 63.4578, | |
| "eval_samples_per_second": 210.66, | |
| "eval_steps_per_second": 13.174, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.4680969629423238, | |
| "grad_norm": 470016.21875, | |
| "learning_rate": 2.6295724648865566e-05, | |
| "loss": 1.9391, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.4680969629423238, | |
| "eval_loss": 1.7201862335205078, | |
| "eval_runtime": 63.3953, | |
| "eval_samples_per_second": 210.867, | |
| "eval_steps_per_second": 13.187, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.4903872945110059, | |
| "grad_norm": 485373.15625, | |
| "learning_rate": 2.6064207439419662e-05, | |
| "loss": 1.9352, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.4903872945110059, | |
| "eval_loss": 1.698974370956421, | |
| "eval_runtime": 63.411, | |
| "eval_samples_per_second": 210.815, | |
| "eval_steps_per_second": 13.184, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.5126776260796879, | |
| "grad_norm": 472483.78125, | |
| "learning_rate": 2.5832690229973762e-05, | |
| "loss": 1.9412, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.5126776260796879, | |
| "eval_loss": 1.7102808952331543, | |
| "eval_runtime": 63.3414, | |
| "eval_samples_per_second": 211.047, | |
| "eval_steps_per_second": 13.198, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.53496795764837, | |
| "grad_norm": 444579.25, | |
| "learning_rate": 2.560117302052786e-05, | |
| "loss": 1.9331, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.53496795764837, | |
| "eval_loss": 1.7026287317276, | |
| "eval_runtime": 63.7826, | |
| "eval_samples_per_second": 209.587, | |
| "eval_steps_per_second": 13.107, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.5572582892170521, | |
| "grad_norm": 544911.375, | |
| "learning_rate": 2.5369655811081957e-05, | |
| "loss": 1.933, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.5572582892170521, | |
| "eval_loss": 1.6917425394058228, | |
| "eval_runtime": 63.6992, | |
| "eval_samples_per_second": 209.861, | |
| "eval_steps_per_second": 13.124, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.5795486207857342, | |
| "grad_norm": 442399.9375, | |
| "learning_rate": 2.5138138601636054e-05, | |
| "loss": 1.9209, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.5795486207857342, | |
| "eval_loss": 1.7024257183074951, | |
| "eval_runtime": 63.6754, | |
| "eval_samples_per_second": 209.94, | |
| "eval_steps_per_second": 13.129, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.6018389523544163, | |
| "grad_norm": 478748.46875, | |
| "learning_rate": 2.4906621392190153e-05, | |
| "loss": 1.9193, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.6018389523544163, | |
| "eval_loss": 1.7053234577178955, | |
| "eval_runtime": 63.8191, | |
| "eval_samples_per_second": 209.467, | |
| "eval_steps_per_second": 13.1, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.6241292839230984, | |
| "grad_norm": 468540.28125, | |
| "learning_rate": 2.4675104182744253e-05, | |
| "loss": 1.8972, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.6241292839230984, | |
| "eval_loss": 1.6971023082733154, | |
| "eval_runtime": 63.3867, | |
| "eval_samples_per_second": 210.896, | |
| "eval_steps_per_second": 13.189, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.6464196154917804, | |
| "grad_norm": 447092.15625, | |
| "learning_rate": 2.444358697329835e-05, | |
| "loss": 1.9116, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.6464196154917804, | |
| "eval_loss": 1.6840194463729858, | |
| "eval_runtime": 62.9634, | |
| "eval_samples_per_second": 212.314, | |
| "eval_steps_per_second": 13.278, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.6687099470604625, | |
| "grad_norm": 498904.8125, | |
| "learning_rate": 2.421206976385245e-05, | |
| "loss": 1.908, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.6687099470604625, | |
| "eval_loss": 1.69094979763031, | |
| "eval_runtime": 63.2731, | |
| "eval_samples_per_second": 211.275, | |
| "eval_steps_per_second": 13.213, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.6910002786291446, | |
| "grad_norm": 462770.21875, | |
| "learning_rate": 2.3980552554406545e-05, | |
| "loss": 1.9074, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.6910002786291446, | |
| "eval_loss": 1.6828372478485107, | |
| "eval_runtime": 63.5905, | |
| "eval_samples_per_second": 210.22, | |
| "eval_steps_per_second": 13.147, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.7132906101978267, | |
| "grad_norm": 510012.96875, | |
| "learning_rate": 2.374903534496064e-05, | |
| "loss": 1.8977, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.7132906101978267, | |
| "eval_loss": 1.6863964796066284, | |
| "eval_runtime": 63.6806, | |
| "eval_samples_per_second": 209.923, | |
| "eval_steps_per_second": 13.128, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.7355809417665088, | |
| "grad_norm": 471075.21875, | |
| "learning_rate": 2.351751813551474e-05, | |
| "loss": 1.9154, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.7355809417665088, | |
| "eval_loss": 1.6765246391296387, | |
| "eval_runtime": 63.5227, | |
| "eval_samples_per_second": 210.444, | |
| "eval_steps_per_second": 13.161, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.7578712733351909, | |
| "grad_norm": 459823.4375, | |
| "learning_rate": 2.328600092606884e-05, | |
| "loss": 1.8869, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.7578712733351909, | |
| "eval_loss": 1.6767551898956299, | |
| "eval_runtime": 62.9808, | |
| "eval_samples_per_second": 212.255, | |
| "eval_steps_per_second": 13.274, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.780161604903873, | |
| "grad_norm": 431094.84375, | |
| "learning_rate": 2.3054483716622936e-05, | |
| "loss": 1.8935, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.780161604903873, | |
| "eval_loss": 1.6749335527420044, | |
| "eval_runtime": 62.9334, | |
| "eval_samples_per_second": 212.415, | |
| "eval_steps_per_second": 13.284, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.8024519364725551, | |
| "grad_norm": 444868.96875, | |
| "learning_rate": 2.2822966507177032e-05, | |
| "loss": 1.8913, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.8024519364725551, | |
| "eval_loss": 1.6784425973892212, | |
| "eval_runtime": 63.2429, | |
| "eval_samples_per_second": 211.376, | |
| "eval_steps_per_second": 13.219, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.8247422680412371, | |
| "grad_norm": 451189.40625, | |
| "learning_rate": 2.259144929773113e-05, | |
| "loss": 1.8878, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.8247422680412371, | |
| "eval_loss": 1.6754982471466064, | |
| "eval_runtime": 63.7373, | |
| "eval_samples_per_second": 209.736, | |
| "eval_steps_per_second": 13.116, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.8470325996099192, | |
| "grad_norm": 441200.9375, | |
| "learning_rate": 2.235993208828523e-05, | |
| "loss": 1.8948, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.8470325996099192, | |
| "eval_loss": 1.6711094379425049, | |
| "eval_runtime": 62.7885, | |
| "eval_samples_per_second": 212.905, | |
| "eval_steps_per_second": 13.315, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.8693229311786013, | |
| "grad_norm": 467896.84375, | |
| "learning_rate": 2.2128414878839327e-05, | |
| "loss": 1.8816, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.8693229311786013, | |
| "eval_loss": 1.671681523323059, | |
| "eval_runtime": 63.1597, | |
| "eval_samples_per_second": 211.654, | |
| "eval_steps_per_second": 13.236, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.8916132627472834, | |
| "grad_norm": 456236.3125, | |
| "learning_rate": 2.1896897669393427e-05, | |
| "loss": 1.8864, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.8916132627472834, | |
| "eval_loss": 1.673299789428711, | |
| "eval_runtime": 63.6594, | |
| "eval_samples_per_second": 209.992, | |
| "eval_steps_per_second": 13.132, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.9139035943159655, | |
| "grad_norm": 470346.40625, | |
| "learning_rate": 2.1665380459947523e-05, | |
| "loss": 1.879, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.9139035943159655, | |
| "eval_loss": 1.6596086025238037, | |
| "eval_runtime": 63.3568, | |
| "eval_samples_per_second": 210.995, | |
| "eval_steps_per_second": 13.195, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.9361939258846476, | |
| "grad_norm": 468418.625, | |
| "learning_rate": 2.143386325050162e-05, | |
| "loss": 1.8862, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.9361939258846476, | |
| "eval_loss": 1.6646808385849, | |
| "eval_runtime": 63.5493, | |
| "eval_samples_per_second": 210.356, | |
| "eval_steps_per_second": 13.155, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.9584842574533297, | |
| "grad_norm": 454157.0, | |
| "learning_rate": 2.120234604105572e-05, | |
| "loss": 1.8722, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.9584842574533297, | |
| "eval_loss": 1.667905330657959, | |
| "eval_runtime": 62.7216, | |
| "eval_samples_per_second": 213.132, | |
| "eval_steps_per_second": 13.329, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.9807745890220118, | |
| "grad_norm": 462536.9375, | |
| "learning_rate": 2.0970828831609818e-05, | |
| "loss": 1.8669, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.9807745890220118, | |
| "eval_loss": 1.65095853805542, | |
| "eval_runtime": 62.9484, | |
| "eval_samples_per_second": 212.364, | |
| "eval_steps_per_second": 13.281, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 1.0030649205906939, | |
| "grad_norm": 478586.0, | |
| "learning_rate": 2.0739311622163914e-05, | |
| "loss": 1.8636, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.0030649205906939, | |
| "eval_loss": 1.6621143817901611, | |
| "eval_runtime": 63.2068, | |
| "eval_samples_per_second": 211.496, | |
| "eval_steps_per_second": 13.226, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.0253552521593758, | |
| "grad_norm": 539991.8125, | |
| "learning_rate": 2.050779441271801e-05, | |
| "loss": 1.8262, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 1.0253552521593758, | |
| "eval_loss": 1.6551061868667603, | |
| "eval_runtime": 63.124, | |
| "eval_samples_per_second": 211.774, | |
| "eval_steps_per_second": 13.244, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 1.047645583728058, | |
| "grad_norm": 428922.53125, | |
| "learning_rate": 2.027627720327211e-05, | |
| "loss": 1.8142, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 1.047645583728058, | |
| "eval_loss": 1.6545677185058594, | |
| "eval_runtime": 63.0786, | |
| "eval_samples_per_second": 211.926, | |
| "eval_steps_per_second": 13.253, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 1.06993591529674, | |
| "grad_norm": 438794.46875, | |
| "learning_rate": 2.004475999382621e-05, | |
| "loss": 1.8097, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.06993591529674, | |
| "eval_loss": 1.6556122303009033, | |
| "eval_runtime": 63.6203, | |
| "eval_samples_per_second": 210.122, | |
| "eval_steps_per_second": 13.14, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.092226246865422, | |
| "grad_norm": 468837.34375, | |
| "learning_rate": 1.9813242784380306e-05, | |
| "loss": 1.8242, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 1.092226246865422, | |
| "eval_loss": 1.657508134841919, | |
| "eval_runtime": 63.2746, | |
| "eval_samples_per_second": 211.27, | |
| "eval_steps_per_second": 13.212, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 1.1145165784341042, | |
| "grad_norm": 449878.40625, | |
| "learning_rate": 1.9581725574934405e-05, | |
| "loss": 1.8143, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.1145165784341042, | |
| "eval_loss": 1.6671580076217651, | |
| "eval_runtime": 63.9114, | |
| "eval_samples_per_second": 209.165, | |
| "eval_steps_per_second": 13.081, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.1368069100027862, | |
| "grad_norm": 427682.84375, | |
| "learning_rate": 1.93502083654885e-05, | |
| "loss": 1.8242, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 1.1368069100027862, | |
| "eval_loss": 1.6521689891815186, | |
| "eval_runtime": 62.5274, | |
| "eval_samples_per_second": 213.794, | |
| "eval_steps_per_second": 13.37, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 1.1590972415714684, | |
| "grad_norm": 441866.34375, | |
| "learning_rate": 1.9118691156042597e-05, | |
| "loss": 1.8105, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 1.1590972415714684, | |
| "eval_loss": 1.6530238389968872, | |
| "eval_runtime": 62.6567, | |
| "eval_samples_per_second": 213.353, | |
| "eval_steps_per_second": 13.343, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 1.1813875731401504, | |
| "grad_norm": 455879.25, | |
| "learning_rate": 1.8887173946596697e-05, | |
| "loss": 1.813, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 1.1813875731401504, | |
| "eval_loss": 1.654367208480835, | |
| "eval_runtime": 63.6338, | |
| "eval_samples_per_second": 210.077, | |
| "eval_steps_per_second": 13.138, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 1.2036779047088326, | |
| "grad_norm": 410671.53125, | |
| "learning_rate": 1.8655656737150796e-05, | |
| "loss": 1.8201, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.2036779047088326, | |
| "eval_loss": 1.6491619348526, | |
| "eval_runtime": 63.5953, | |
| "eval_samples_per_second": 210.204, | |
| "eval_steps_per_second": 13.146, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.2259682362775146, | |
| "grad_norm": 439893.84375, | |
| "learning_rate": 1.8424139527704893e-05, | |
| "loss": 1.8078, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 1.2259682362775146, | |
| "eval_loss": 1.6452112197875977, | |
| "eval_runtime": 63.1345, | |
| "eval_samples_per_second": 211.738, | |
| "eval_steps_per_second": 13.242, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 1.2482585678461966, | |
| "grad_norm": 436707.84375, | |
| "learning_rate": 1.819262231825899e-05, | |
| "loss": 1.8081, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 1.2482585678461966, | |
| "eval_loss": 1.63877272605896, | |
| "eval_runtime": 63.3151, | |
| "eval_samples_per_second": 211.134, | |
| "eval_steps_per_second": 13.204, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 1.2705488994148788, | |
| "grad_norm": 453170.34375, | |
| "learning_rate": 1.7961105108813088e-05, | |
| "loss": 1.809, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 1.2705488994148788, | |
| "eval_loss": 1.6391615867614746, | |
| "eval_runtime": 63.0066, | |
| "eval_samples_per_second": 212.168, | |
| "eval_steps_per_second": 13.268, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 1.292839230983561, | |
| "grad_norm": 419482.625, | |
| "learning_rate": 1.7729587899367188e-05, | |
| "loss": 1.7993, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 1.292839230983561, | |
| "eval_loss": 1.6412807703018188, | |
| "eval_runtime": 62.9872, | |
| "eval_samples_per_second": 212.234, | |
| "eval_steps_per_second": 13.273, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 1.315129562552243, | |
| "grad_norm": 439712.59375, | |
| "learning_rate": 1.7498070689921284e-05, | |
| "loss": 1.806, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 1.315129562552243, | |
| "eval_loss": 1.6369620561599731, | |
| "eval_runtime": 62.9462, | |
| "eval_samples_per_second": 212.372, | |
| "eval_steps_per_second": 13.281, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 1.337419894120925, | |
| "grad_norm": 433225.875, | |
| "learning_rate": 1.7266553480475383e-05, | |
| "loss": 1.8054, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.337419894120925, | |
| "eval_loss": 1.6376512050628662, | |
| "eval_runtime": 63.3967, | |
| "eval_samples_per_second": 210.863, | |
| "eval_steps_per_second": 13.187, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.3597102256896072, | |
| "grad_norm": 442755.03125, | |
| "learning_rate": 1.703503627102948e-05, | |
| "loss": 1.7996, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 1.3597102256896072, | |
| "eval_loss": 1.6386842727661133, | |
| "eval_runtime": 64.1806, | |
| "eval_samples_per_second": 208.287, | |
| "eval_steps_per_second": 13.026, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 1.3820005572582892, | |
| "grad_norm": 444388.1875, | |
| "learning_rate": 1.6803519061583576e-05, | |
| "loss": 1.7973, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 1.3820005572582892, | |
| "eval_loss": 1.637714147567749, | |
| "eval_runtime": 62.6809, | |
| "eval_samples_per_second": 213.271, | |
| "eval_steps_per_second": 13.337, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 1.4042908888269712, | |
| "grad_norm": 457163.5625, | |
| "learning_rate": 1.657200185213768e-05, | |
| "loss": 1.8103, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 1.4042908888269712, | |
| "eval_loss": 1.6353328227996826, | |
| "eval_runtime": 63.7218, | |
| "eval_samples_per_second": 209.787, | |
| "eval_steps_per_second": 13.12, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 1.4265812203956534, | |
| "grad_norm": 423063.71875, | |
| "learning_rate": 1.6340484642691775e-05, | |
| "loss": 1.8193, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 1.4265812203956534, | |
| "eval_loss": 1.632241129875183, | |
| "eval_runtime": 62.6795, | |
| "eval_samples_per_second": 213.276, | |
| "eval_steps_per_second": 13.338, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 1.4488715519643356, | |
| "grad_norm": 450858.5625, | |
| "learning_rate": 1.610896743324587e-05, | |
| "loss": 1.8133, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 1.4488715519643356, | |
| "eval_loss": 1.642052412033081, | |
| "eval_runtime": 64.1592, | |
| "eval_samples_per_second": 208.357, | |
| "eval_steps_per_second": 13.03, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 1.4711618835330176, | |
| "grad_norm": 424338.96875, | |
| "learning_rate": 1.5877450223799967e-05, | |
| "loss": 1.8016, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 1.4711618835330176, | |
| "eval_loss": 1.6317675113677979, | |
| "eval_runtime": 63.0019, | |
| "eval_samples_per_second": 212.184, | |
| "eval_steps_per_second": 13.269, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 1.4934522151016996, | |
| "grad_norm": 441573.03125, | |
| "learning_rate": 1.5645933014354067e-05, | |
| "loss": 1.7923, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 1.4934522151016996, | |
| "eval_loss": 1.6302847862243652, | |
| "eval_runtime": 63.6702, | |
| "eval_samples_per_second": 209.957, | |
| "eval_steps_per_second": 13.13, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 1.5157425466703818, | |
| "grad_norm": 460539.0625, | |
| "learning_rate": 1.5414415804908166e-05, | |
| "loss": 1.7969, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 1.5157425466703818, | |
| "eval_loss": 1.6323357820510864, | |
| "eval_runtime": 62.8247, | |
| "eval_samples_per_second": 212.783, | |
| "eval_steps_per_second": 13.307, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 1.5380328782390638, | |
| "grad_norm": 432685.96875, | |
| "learning_rate": 1.5182898595462264e-05, | |
| "loss": 1.7955, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 1.5380328782390638, | |
| "eval_loss": 1.6306127309799194, | |
| "eval_runtime": 63.1442, | |
| "eval_samples_per_second": 211.706, | |
| "eval_steps_per_second": 13.24, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 1.5603232098077457, | |
| "grad_norm": 414684.125, | |
| "learning_rate": 1.495138138601636e-05, | |
| "loss": 1.7933, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 1.5603232098077457, | |
| "eval_loss": 1.6255934238433838, | |
| "eval_runtime": 63.4943, | |
| "eval_samples_per_second": 210.539, | |
| "eval_steps_per_second": 13.167, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 1.582613541376428, | |
| "grad_norm": 423529.65625, | |
| "learning_rate": 1.471986417657046e-05, | |
| "loss": 1.7951, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 1.582613541376428, | |
| "eval_loss": 1.6387838125228882, | |
| "eval_runtime": 63.598, | |
| "eval_samples_per_second": 210.195, | |
| "eval_steps_per_second": 13.145, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 1.6049038729451102, | |
| "grad_norm": 421001.21875, | |
| "learning_rate": 1.4488346967124556e-05, | |
| "loss": 1.8165, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 1.6049038729451102, | |
| "eval_loss": 1.6351381540298462, | |
| "eval_runtime": 63.005, | |
| "eval_samples_per_second": 212.174, | |
| "eval_steps_per_second": 13.269, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 1.6271942045137922, | |
| "grad_norm": 401336.375, | |
| "learning_rate": 1.4256829757678654e-05, | |
| "loss": 1.7934, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 1.6271942045137922, | |
| "eval_loss": 1.630112886428833, | |
| "eval_runtime": 63.0665, | |
| "eval_samples_per_second": 211.967, | |
| "eval_steps_per_second": 13.256, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 1.6494845360824741, | |
| "grad_norm": 439442.84375, | |
| "learning_rate": 1.4025312548232753e-05, | |
| "loss": 1.7809, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 1.6494845360824741, | |
| "eval_loss": 1.6299716234207153, | |
| "eval_runtime": 63.6859, | |
| "eval_samples_per_second": 209.905, | |
| "eval_steps_per_second": 13.127, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 1.6717748676511563, | |
| "grad_norm": 407261.71875, | |
| "learning_rate": 1.379379533878685e-05, | |
| "loss": 1.7899, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.6717748676511563, | |
| "eval_loss": 1.6285357475280762, | |
| "eval_runtime": 63.1642, | |
| "eval_samples_per_second": 211.639, | |
| "eval_steps_per_second": 13.235, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.6940651992198386, | |
| "grad_norm": 403575.21875, | |
| "learning_rate": 1.3562278129340949e-05, | |
| "loss": 1.7887, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 1.6940651992198386, | |
| "eval_loss": 1.6282232999801636, | |
| "eval_runtime": 62.5905, | |
| "eval_samples_per_second": 213.579, | |
| "eval_steps_per_second": 13.357, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 1.7163555307885203, | |
| "grad_norm": 440494.6875, | |
| "learning_rate": 1.3330760919895045e-05, | |
| "loss": 1.7859, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 1.7163555307885203, | |
| "eval_loss": 1.622281551361084, | |
| "eval_runtime": 63.4527, | |
| "eval_samples_per_second": 210.677, | |
| "eval_steps_per_second": 13.175, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 1.7386458623572025, | |
| "grad_norm": 414236.25, | |
| "learning_rate": 1.3099243710449143e-05, | |
| "loss": 1.8, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 1.7386458623572025, | |
| "eval_loss": 1.623448371887207, | |
| "eval_runtime": 62.7927, | |
| "eval_samples_per_second": 212.891, | |
| "eval_steps_per_second": 13.314, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 1.7609361939258847, | |
| "grad_norm": 438500.71875, | |
| "learning_rate": 1.2867726501003242e-05, | |
| "loss": 1.7907, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 1.7609361939258847, | |
| "eval_loss": 1.6255710124969482, | |
| "eval_runtime": 63.6158, | |
| "eval_samples_per_second": 210.137, | |
| "eval_steps_per_second": 13.141, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 1.7832265254945667, | |
| "grad_norm": 413099.34375, | |
| "learning_rate": 1.2636209291557339e-05, | |
| "loss": 1.8052, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.7832265254945667, | |
| "eval_loss": 1.6219525337219238, | |
| "eval_runtime": 63.4015, | |
| "eval_samples_per_second": 210.847, | |
| "eval_steps_per_second": 13.186, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.8055168570632487, | |
| "grad_norm": 442402.25, | |
| "learning_rate": 1.2404692082111438e-05, | |
| "loss": 1.7881, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 1.8055168570632487, | |
| "eval_loss": 1.6209757328033447, | |
| "eval_runtime": 62.8346, | |
| "eval_samples_per_second": 212.749, | |
| "eval_steps_per_second": 13.305, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 1.827807188631931, | |
| "grad_norm": 482508.3125, | |
| "learning_rate": 1.2173174872665536e-05, | |
| "loss": 1.7967, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 1.827807188631931, | |
| "eval_loss": 1.6282365322113037, | |
| "eval_runtime": 63.2359, | |
| "eval_samples_per_second": 211.399, | |
| "eval_steps_per_second": 13.22, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 1.8500975202006131, | |
| "grad_norm": 443377.25, | |
| "learning_rate": 1.1941657663219632e-05, | |
| "loss": 1.7921, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 1.8500975202006131, | |
| "eval_loss": 1.624315619468689, | |
| "eval_runtime": 63.0489, | |
| "eval_samples_per_second": 212.026, | |
| "eval_steps_per_second": 13.26, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 1.8723878517692951, | |
| "grad_norm": 397804.8125, | |
| "learning_rate": 1.1710140453773732e-05, | |
| "loss": 1.7898, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 1.8723878517692951, | |
| "eval_loss": 1.6195310354232788, | |
| "eval_runtime": 63.4127, | |
| "eval_samples_per_second": 210.81, | |
| "eval_steps_per_second": 13.183, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 1.894678183337977, | |
| "grad_norm": 422386.96875, | |
| "learning_rate": 1.1478623244327828e-05, | |
| "loss": 1.779, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 1.894678183337977, | |
| "eval_loss": 1.6192779541015625, | |
| "eval_runtime": 63.476, | |
| "eval_samples_per_second": 210.599, | |
| "eval_steps_per_second": 13.17, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 1.9169685149066593, | |
| "grad_norm": 411348.03125, | |
| "learning_rate": 1.1247106034881927e-05, | |
| "loss": 1.7852, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 1.9169685149066593, | |
| "eval_loss": 1.6169335842132568, | |
| "eval_runtime": 63.0679, | |
| "eval_samples_per_second": 211.962, | |
| "eval_steps_per_second": 13.256, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 1.9392588464753413, | |
| "grad_norm": 444615.65625, | |
| "learning_rate": 1.1015588825436025e-05, | |
| "loss": 1.7813, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 1.9392588464753413, | |
| "eval_loss": 1.6124924421310425, | |
| "eval_runtime": 63.9864, | |
| "eval_samples_per_second": 208.919, | |
| "eval_steps_per_second": 13.065, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 1.9615491780440233, | |
| "grad_norm": 414404.125, | |
| "learning_rate": 1.0784071615990121e-05, | |
| "loss": 1.7893, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 1.9615491780440233, | |
| "eval_loss": 1.6191661357879639, | |
| "eval_runtime": 63.6221, | |
| "eval_samples_per_second": 210.116, | |
| "eval_steps_per_second": 13.14, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 1.9838395096127055, | |
| "grad_norm": 445457.1875, | |
| "learning_rate": 1.055255440654422e-05, | |
| "loss": 1.7905, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 1.9838395096127055, | |
| "eval_loss": 1.6134953498840332, | |
| "eval_runtime": 63.3933, | |
| "eval_samples_per_second": 210.874, | |
| "eval_steps_per_second": 13.188, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 2.0061298411813877, | |
| "grad_norm": 439959.25, | |
| "learning_rate": 1.0321037197098317e-05, | |
| "loss": 1.7873, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 2.0061298411813877, | |
| "eval_loss": 1.6150258779525757, | |
| "eval_runtime": 62.9808, | |
| "eval_samples_per_second": 212.255, | |
| "eval_steps_per_second": 13.274, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 2.0284201727500695, | |
| "grad_norm": 399886.46875, | |
| "learning_rate": 1.0089519987652416e-05, | |
| "loss": 1.7304, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 2.0284201727500695, | |
| "eval_loss": 1.615242600440979, | |
| "eval_runtime": 63.4005, | |
| "eval_samples_per_second": 210.85, | |
| "eval_steps_per_second": 13.186, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 2.0507105043187517, | |
| "grad_norm": 432171.59375, | |
| "learning_rate": 9.858002778206514e-06, | |
| "loss": 1.748, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 2.0507105043187517, | |
| "eval_loss": 1.6178884506225586, | |
| "eval_runtime": 63.1372, | |
| "eval_samples_per_second": 211.729, | |
| "eval_steps_per_second": 13.241, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 2.073000835887434, | |
| "grad_norm": 438172.875, | |
| "learning_rate": 9.62648556876061e-06, | |
| "loss": 1.7488, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 2.073000835887434, | |
| "eval_loss": 1.615029215812683, | |
| "eval_runtime": 63.7948, | |
| "eval_samples_per_second": 209.547, | |
| "eval_steps_per_second": 13.105, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 2.095291167456116, | |
| "grad_norm": 422492.28125, | |
| "learning_rate": 9.39496835931471e-06, | |
| "loss": 1.7482, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 2.095291167456116, | |
| "eval_loss": 1.6207051277160645, | |
| "eval_runtime": 63.458, | |
| "eval_samples_per_second": 210.659, | |
| "eval_steps_per_second": 13.174, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 2.117581499024798, | |
| "grad_norm": 438706.75, | |
| "learning_rate": 9.163451149868806e-06, | |
| "loss": 1.739, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 2.117581499024798, | |
| "eval_loss": 1.6097335815429688, | |
| "eval_runtime": 63.1577, | |
| "eval_samples_per_second": 211.66, | |
| "eval_steps_per_second": 13.237, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 2.13987183059348, | |
| "grad_norm": 418132.5625, | |
| "learning_rate": 8.931933940422906e-06, | |
| "loss": 1.7599, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 2.13987183059348, | |
| "eval_loss": 1.6131938695907593, | |
| "eval_runtime": 63.4093, | |
| "eval_samples_per_second": 210.821, | |
| "eval_steps_per_second": 13.184, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 2.1621621621621623, | |
| "grad_norm": 430660.1875, | |
| "learning_rate": 8.700416730977003e-06, | |
| "loss": 1.7473, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 2.1621621621621623, | |
| "eval_loss": 1.6116857528686523, | |
| "eval_runtime": 63.5141, | |
| "eval_samples_per_second": 210.473, | |
| "eval_steps_per_second": 13.162, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 2.184452493730844, | |
| "grad_norm": 410735.0625, | |
| "learning_rate": 8.4688995215311e-06, | |
| "loss": 1.75, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 2.184452493730844, | |
| "eval_loss": 1.611021876335144, | |
| "eval_runtime": 63.3479, | |
| "eval_samples_per_second": 211.025, | |
| "eval_steps_per_second": 13.197, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 2.2067428252995263, | |
| "grad_norm": 437123.8125, | |
| "learning_rate": 8.237382312085199e-06, | |
| "loss": 1.7432, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 2.2067428252995263, | |
| "eval_loss": 1.614385962486267, | |
| "eval_runtime": 63.4099, | |
| "eval_samples_per_second": 210.819, | |
| "eval_steps_per_second": 13.184, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 2.2290331568682085, | |
| "grad_norm": 434528.1875, | |
| "learning_rate": 8.005865102639295e-06, | |
| "loss": 1.7491, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 2.2290331568682085, | |
| "eval_loss": 1.6123193502426147, | |
| "eval_runtime": 63.2705, | |
| "eval_samples_per_second": 211.283, | |
| "eval_steps_per_second": 13.213, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 2.2513234884368907, | |
| "grad_norm": 394168.875, | |
| "learning_rate": 7.774347893193395e-06, | |
| "loss": 1.7377, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 2.2513234884368907, | |
| "eval_loss": 1.6091859340667725, | |
| "eval_runtime": 63.9924, | |
| "eval_samples_per_second": 208.9, | |
| "eval_steps_per_second": 13.064, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 2.2736138200055724, | |
| "grad_norm": 405977.28125, | |
| "learning_rate": 7.542830683747492e-06, | |
| "loss": 1.7439, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 2.2736138200055724, | |
| "eval_loss": 1.6095374822616577, | |
| "eval_runtime": 63.1561, | |
| "eval_samples_per_second": 211.666, | |
| "eval_steps_per_second": 13.237, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 2.2959041515742546, | |
| "grad_norm": 441095.4375, | |
| "learning_rate": 7.31131347430159e-06, | |
| "loss": 1.7417, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 2.2959041515742546, | |
| "eval_loss": 1.6078873872756958, | |
| "eval_runtime": 62.9452, | |
| "eval_samples_per_second": 212.375, | |
| "eval_steps_per_second": 13.281, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 2.318194483142937, | |
| "grad_norm": 450342.375, | |
| "learning_rate": 7.079796264855688e-06, | |
| "loss": 1.7482, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 2.318194483142937, | |
| "eval_loss": 1.6116305589675903, | |
| "eval_runtime": 63.4969, | |
| "eval_samples_per_second": 210.53, | |
| "eval_steps_per_second": 13.166, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 2.340484814711619, | |
| "grad_norm": 462851.03125, | |
| "learning_rate": 6.848279055409786e-06, | |
| "loss": 1.7391, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 2.340484814711619, | |
| "eval_loss": 1.6078118085861206, | |
| "eval_runtime": 63.3093, | |
| "eval_samples_per_second": 211.154, | |
| "eval_steps_per_second": 13.205, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 2.362775146280301, | |
| "grad_norm": 405519.75, | |
| "learning_rate": 6.616761845963883e-06, | |
| "loss": 1.7345, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 2.362775146280301, | |
| "eval_loss": 1.6122682094573975, | |
| "eval_runtime": 63.3463, | |
| "eval_samples_per_second": 211.031, | |
| "eval_steps_per_second": 13.197, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 2.385065477848983, | |
| "grad_norm": 471771.9375, | |
| "learning_rate": 6.385244636517981e-06, | |
| "loss": 1.7365, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 2.385065477848983, | |
| "eval_loss": 1.6120107173919678, | |
| "eval_runtime": 63.2512, | |
| "eval_samples_per_second": 211.348, | |
| "eval_steps_per_second": 13.217, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 2.4073558094176652, | |
| "grad_norm": 416676.4375, | |
| "learning_rate": 6.153727427072079e-06, | |
| "loss": 1.7446, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 2.4073558094176652, | |
| "eval_loss": 1.6103585958480835, | |
| "eval_runtime": 63.4362, | |
| "eval_samples_per_second": 210.731, | |
| "eval_steps_per_second": 13.179, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 2.429646140986347, | |
| "grad_norm": 409846.5, | |
| "learning_rate": 5.9222102176261775e-06, | |
| "loss": 1.7273, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 2.429646140986347, | |
| "eval_loss": 1.6118780374526978, | |
| "eval_runtime": 63.3423, | |
| "eval_samples_per_second": 211.044, | |
| "eval_steps_per_second": 13.198, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 2.4519364725550292, | |
| "grad_norm": 400530.59375, | |
| "learning_rate": 5.690693008180275e-06, | |
| "loss": 1.7494, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 2.4519364725550292, | |
| "eval_loss": 1.6059410572052002, | |
| "eval_runtime": 63.1379, | |
| "eval_samples_per_second": 211.727, | |
| "eval_steps_per_second": 13.241, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 2.4742268041237114, | |
| "grad_norm": 427720.5625, | |
| "learning_rate": 5.459175798734372e-06, | |
| "loss": 1.7332, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 2.4742268041237114, | |
| "eval_loss": 1.6064603328704834, | |
| "eval_runtime": 63.7921, | |
| "eval_samples_per_second": 209.556, | |
| "eval_steps_per_second": 13.105, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 2.496517135692393, | |
| "grad_norm": 450690.875, | |
| "learning_rate": 5.22765858928847e-06, | |
| "loss": 1.7285, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 2.496517135692393, | |
| "eval_loss": 1.6082170009613037, | |
| "eval_runtime": 62.5222, | |
| "eval_samples_per_second": 213.812, | |
| "eval_steps_per_second": 13.371, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 2.5188074672610754, | |
| "grad_norm": 412636.5, | |
| "learning_rate": 4.996141379842568e-06, | |
| "loss": 1.7319, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 2.5188074672610754, | |
| "eval_loss": 1.6035947799682617, | |
| "eval_runtime": 63.3474, | |
| "eval_samples_per_second": 211.027, | |
| "eval_steps_per_second": 13.197, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 2.5410977988297576, | |
| "grad_norm": 403399.875, | |
| "learning_rate": 4.764624170396667e-06, | |
| "loss": 1.7245, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 2.5410977988297576, | |
| "eval_loss": 1.6081373691558838, | |
| "eval_runtime": 63.6595, | |
| "eval_samples_per_second": 209.992, | |
| "eval_steps_per_second": 13.132, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 2.56338813039844, | |
| "grad_norm": 432179.4375, | |
| "learning_rate": 4.5331069609507645e-06, | |
| "loss": 1.7379, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 2.56338813039844, | |
| "eval_loss": 1.6043556928634644, | |
| "eval_runtime": 63.1427, | |
| "eval_samples_per_second": 211.711, | |
| "eval_steps_per_second": 13.24, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 2.585678461967122, | |
| "grad_norm": 398148.96875, | |
| "learning_rate": 4.3015897515048615e-06, | |
| "loss": 1.7349, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 2.585678461967122, | |
| "eval_loss": 1.6080468893051147, | |
| "eval_runtime": 63.1768, | |
| "eval_samples_per_second": 211.597, | |
| "eval_steps_per_second": 13.233, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 2.607968793535804, | |
| "grad_norm": 406493.65625, | |
| "learning_rate": 4.070072542058959e-06, | |
| "loss": 1.7408, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 2.607968793535804, | |
| "eval_loss": 1.6043606996536255, | |
| "eval_runtime": 63.7306, | |
| "eval_samples_per_second": 209.758, | |
| "eval_steps_per_second": 13.118, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 2.630259125104486, | |
| "grad_norm": 415316.5625, | |
| "learning_rate": 3.838555332613058e-06, | |
| "loss": 1.7317, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 2.630259125104486, | |
| "eval_loss": 1.6045427322387695, | |
| "eval_runtime": 63.3899, | |
| "eval_samples_per_second": 210.885, | |
| "eval_steps_per_second": 13.188, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 2.652549456673168, | |
| "grad_norm": 433577.84375, | |
| "learning_rate": 3.6070381231671554e-06, | |
| "loss": 1.7288, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 2.652549456673168, | |
| "eval_loss": 1.603334903717041, | |
| "eval_runtime": 63.1008, | |
| "eval_samples_per_second": 211.851, | |
| "eval_steps_per_second": 13.249, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 2.67483978824185, | |
| "grad_norm": 408270.8125, | |
| "learning_rate": 3.3755209137212533e-06, | |
| "loss": 1.7327, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 2.67483978824185, | |
| "eval_loss": 1.6032938957214355, | |
| "eval_runtime": 62.9232, | |
| "eval_samples_per_second": 212.45, | |
| "eval_steps_per_second": 13.286, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 2.697130119810532, | |
| "grad_norm": 412520.28125, | |
| "learning_rate": 3.144003704275351e-06, | |
| "loss": 1.7349, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 2.697130119810532, | |
| "eval_loss": 1.6040232181549072, | |
| "eval_runtime": 63.2685, | |
| "eval_samples_per_second": 211.29, | |
| "eval_steps_per_second": 13.214, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 2.7194204513792144, | |
| "grad_norm": 430910.1875, | |
| "learning_rate": 2.9124864948294494e-06, | |
| "loss": 1.7378, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 2.7194204513792144, | |
| "eval_loss": 1.6026825904846191, | |
| "eval_runtime": 63.0275, | |
| "eval_samples_per_second": 212.098, | |
| "eval_steps_per_second": 13.264, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 2.741710782947896, | |
| "grad_norm": 426524.9375, | |
| "learning_rate": 2.680969285383547e-06, | |
| "loss": 1.7438, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 2.741710782947896, | |
| "eval_loss": 1.6039636135101318, | |
| "eval_runtime": 63.8654, | |
| "eval_samples_per_second": 209.315, | |
| "eval_steps_per_second": 13.09, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 2.7640011145165784, | |
| "grad_norm": 439803.8125, | |
| "learning_rate": 2.449452075937645e-06, | |
| "loss": 1.7357, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 2.7640011145165784, | |
| "eval_loss": 1.6037323474884033, | |
| "eval_runtime": 63.1113, | |
| "eval_samples_per_second": 211.816, | |
| "eval_steps_per_second": 13.246, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 2.7862914460852606, | |
| "grad_norm": 442227.34375, | |
| "learning_rate": 2.2179348664917425e-06, | |
| "loss": 1.7333, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 2.7862914460852606, | |
| "eval_loss": 1.6014666557312012, | |
| "eval_runtime": 63.3883, | |
| "eval_samples_per_second": 210.891, | |
| "eval_steps_per_second": 13.189, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 2.8085817776539423, | |
| "grad_norm": 409348.65625, | |
| "learning_rate": 1.9864176570458403e-06, | |
| "loss": 1.7344, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 2.8085817776539423, | |
| "eval_loss": 1.6025654077529907, | |
| "eval_runtime": 63.562, | |
| "eval_samples_per_second": 210.315, | |
| "eval_steps_per_second": 13.153, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 2.8308721092226246, | |
| "grad_norm": 438774.5, | |
| "learning_rate": 1.7549004475999384e-06, | |
| "loss": 1.7356, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 2.8308721092226246, | |
| "eval_loss": 1.6015514135360718, | |
| "eval_runtime": 63.4784, | |
| "eval_samples_per_second": 210.591, | |
| "eval_steps_per_second": 13.17, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 2.8531624407913068, | |
| "grad_norm": 425528.09375, | |
| "learning_rate": 1.523383238154036e-06, | |
| "loss": 1.7494, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 2.8531624407913068, | |
| "eval_loss": 1.602927565574646, | |
| "eval_runtime": 63.9753, | |
| "eval_samples_per_second": 208.956, | |
| "eval_steps_per_second": 13.068, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 2.875452772359989, | |
| "grad_norm": 402082.0625, | |
| "learning_rate": 1.291866028708134e-06, | |
| "loss": 1.7414, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 2.875452772359989, | |
| "eval_loss": 1.6020632982254028, | |
| "eval_runtime": 63.2754, | |
| "eval_samples_per_second": 211.267, | |
| "eval_steps_per_second": 13.212, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 2.897743103928671, | |
| "grad_norm": 409632.125, | |
| "learning_rate": 1.0603488192622319e-06, | |
| "loss": 1.7337, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 2.897743103928671, | |
| "eval_loss": 1.6021629571914673, | |
| "eval_runtime": 63.0484, | |
| "eval_samples_per_second": 212.028, | |
| "eval_steps_per_second": 13.26, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 2.920033435497353, | |
| "grad_norm": 431921.65625, | |
| "learning_rate": 8.288316098163297e-07, | |
| "loss": 1.7239, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 2.920033435497353, | |
| "eval_loss": 1.6020002365112305, | |
| "eval_runtime": 63.0873, | |
| "eval_samples_per_second": 211.897, | |
| "eval_steps_per_second": 13.251, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 2.942323767066035, | |
| "grad_norm": 436490.65625, | |
| "learning_rate": 5.973144003704276e-07, | |
| "loss": 1.7461, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 2.942323767066035, | |
| "eval_loss": 1.6018171310424805, | |
| "eval_runtime": 63.2016, | |
| "eval_samples_per_second": 211.514, | |
| "eval_steps_per_second": 13.228, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 2.9646140986347174, | |
| "grad_norm": 396898.375, | |
| "learning_rate": 3.657971909245254e-07, | |
| "loss": 1.7358, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 2.9646140986347174, | |
| "eval_loss": 1.601110577583313, | |
| "eval_runtime": 63.1463, | |
| "eval_samples_per_second": 211.699, | |
| "eval_steps_per_second": 13.239, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 2.986904430203399, | |
| "grad_norm": 419936.46875, | |
| "learning_rate": 1.3427998147862324e-07, | |
| "loss": 1.7293, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 2.986904430203399, | |
| "eval_loss": 1.6014801263809204, | |
| "eval_runtime": 63.6722, | |
| "eval_samples_per_second": 209.95, | |
| "eval_steps_per_second": 13.13, | |
| "step": 13400 | |
| } | |
| ], | |
| "logging_steps": 100, | |
| "max_steps": 13458, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.251647920091955e+17, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |