| { | |
| "best_metric": 0.7144020795822144, | |
| "best_model_checkpoint": "/users/u2023000898/model/atrain_qc/scenario_train_40K_cot_ratio_3/checkpoint-1220", | |
| "epoch": 2.9986708925467744, | |
| "eval_steps": 20, | |
| "global_step": 1833, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.01635824557816174, | |
| "grad_norm": 5.295344734073509, | |
| "learning_rate": 5.434782608695653e-07, | |
| "loss": 1.4359, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.03271649115632348, | |
| "grad_norm": 3.422164657559387, | |
| "learning_rate": 1.0869565217391306e-06, | |
| "loss": 1.4181, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.03271649115632348, | |
| "eval_loss": 1.3935878276824951, | |
| "eval_runtime": 8.9905, | |
| "eval_samples_per_second": 8.787, | |
| "eval_steps_per_second": 4.449, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.04907473673448522, | |
| "grad_norm": 1.986613395208058, | |
| "learning_rate": 1.6304347826086957e-06, | |
| "loss": 1.303, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.06543298231264696, | |
| "grad_norm": 1.4523886723377106, | |
| "learning_rate": 2.173913043478261e-06, | |
| "loss": 1.1795, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.06543298231264696, | |
| "eval_loss": 1.1402088403701782, | |
| "eval_runtime": 8.8832, | |
| "eval_samples_per_second": 8.893, | |
| "eval_steps_per_second": 4.503, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.08179122789080871, | |
| "grad_norm": 1.1700390466976565, | |
| "learning_rate": 2.7173913043478263e-06, | |
| "loss": 1.0727, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.09814947346897045, | |
| "grad_norm": 0.8879790034956169, | |
| "learning_rate": 3.2608695652173914e-06, | |
| "loss": 0.9978, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.09814947346897045, | |
| "eval_loss": 0.9959300756454468, | |
| "eval_runtime": 8.9186, | |
| "eval_samples_per_second": 8.858, | |
| "eval_steps_per_second": 4.485, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.1145077190471322, | |
| "grad_norm": 0.8251876813224245, | |
| "learning_rate": 3.804347826086957e-06, | |
| "loss": 0.962, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.13086596462529393, | |
| "grad_norm": 0.7597303029800314, | |
| "learning_rate": 4.347826086956522e-06, | |
| "loss": 0.9344, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.13086596462529393, | |
| "eval_loss": 0.9332086443901062, | |
| "eval_runtime": 9.1918, | |
| "eval_samples_per_second": 8.595, | |
| "eval_steps_per_second": 4.352, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.1472242102034557, | |
| "grad_norm": 0.7643744159525921, | |
| "learning_rate": 4.891304347826087e-06, | |
| "loss": 0.9113, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.16358245578161743, | |
| "grad_norm": 0.7298856847403965, | |
| "learning_rate": 5.4347826086956525e-06, | |
| "loss": 0.8836, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.16358245578161743, | |
| "eval_loss": 0.8985484838485718, | |
| "eval_runtime": 9.0729, | |
| "eval_samples_per_second": 8.707, | |
| "eval_steps_per_second": 4.409, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.17994070135977916, | |
| "grad_norm": 0.7853004926976781, | |
| "learning_rate": 5.978260869565218e-06, | |
| "loss": 0.873, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.1962989469379409, | |
| "grad_norm": 0.8086725449370039, | |
| "learning_rate": 6.521739130434783e-06, | |
| "loss": 0.8649, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.1962989469379409, | |
| "eval_loss": 0.8743940591812134, | |
| "eval_runtime": 8.8733, | |
| "eval_samples_per_second": 8.903, | |
| "eval_steps_per_second": 4.508, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.21265719251610266, | |
| "grad_norm": 0.78611945621464, | |
| "learning_rate": 7.065217391304349e-06, | |
| "loss": 0.8502, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2290154380942644, | |
| "grad_norm": 0.8139497555344073, | |
| "learning_rate": 7.608695652173914e-06, | |
| "loss": 0.8424, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.2290154380942644, | |
| "eval_loss": 0.8560407757759094, | |
| "eval_runtime": 12.0219, | |
| "eval_samples_per_second": 6.571, | |
| "eval_steps_per_second": 3.327, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.24537368367242612, | |
| "grad_norm": 0.8656686458129763, | |
| "learning_rate": 8.15217391304348e-06, | |
| "loss": 0.8449, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.26173192925058786, | |
| "grad_norm": 0.9283042437403443, | |
| "learning_rate": 8.695652173913044e-06, | |
| "loss": 0.8265, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.26173192925058786, | |
| "eval_loss": 0.8416913747787476, | |
| "eval_runtime": 9.036, | |
| "eval_samples_per_second": 8.743, | |
| "eval_steps_per_second": 4.427, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.2780901748287496, | |
| "grad_norm": 0.810298408576699, | |
| "learning_rate": 9.23913043478261e-06, | |
| "loss": 0.8293, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.2944484204069114, | |
| "grad_norm": 0.8313112566441393, | |
| "learning_rate": 9.782608695652175e-06, | |
| "loss": 0.8159, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.2944484204069114, | |
| "eval_loss": 0.8295494318008423, | |
| "eval_runtime": 8.8777, | |
| "eval_samples_per_second": 8.899, | |
| "eval_steps_per_second": 4.506, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.3108066659850731, | |
| "grad_norm": 0.8697727174133723, | |
| "learning_rate": 9.999673339806799e-06, | |
| "loss": 0.8191, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.32716491156323485, | |
| "grad_norm": 0.900525558157214, | |
| "learning_rate": 9.997677237639758e-06, | |
| "loss": 0.8024, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.32716491156323485, | |
| "eval_loss": 0.822140097618103, | |
| "eval_runtime": 8.8793, | |
| "eval_samples_per_second": 8.897, | |
| "eval_steps_per_second": 4.505, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.3435231571413966, | |
| "grad_norm": 0.8222174164436653, | |
| "learning_rate": 9.993867234795303e-06, | |
| "loss": 0.7988, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.3598814027195583, | |
| "grad_norm": 0.8209330165865478, | |
| "learning_rate": 9.98824471410918e-06, | |
| "loss": 0.7934, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.3598814027195583, | |
| "eval_loss": 0.8108458518981934, | |
| "eval_runtime": 8.8086, | |
| "eval_samples_per_second": 8.969, | |
| "eval_steps_per_second": 4.541, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.37623964829772005, | |
| "grad_norm": 0.7678941313429579, | |
| "learning_rate": 9.980811716268215e-06, | |
| "loss": 0.7959, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.3925978938758818, | |
| "grad_norm": 0.82512300528877, | |
| "learning_rate": 9.971570939069637e-06, | |
| "loss": 0.7903, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.3925978938758818, | |
| "eval_loss": 0.8030208349227905, | |
| "eval_runtime": 8.8553, | |
| "eval_samples_per_second": 8.921, | |
| "eval_steps_per_second": 4.517, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.4089561394540436, | |
| "grad_norm": 0.7846619918234344, | |
| "learning_rate": 9.96052573644193e-06, | |
| "loss": 0.7893, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.4253143850322053, | |
| "grad_norm": 0.7731531067943592, | |
| "learning_rate": 9.947680117227512e-06, | |
| "loss": 0.7858, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.4253143850322053, | |
| "eval_loss": 0.7969117164611816, | |
| "eval_runtime": 8.8378, | |
| "eval_samples_per_second": 8.939, | |
| "eval_steps_per_second": 4.526, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.44167263061036705, | |
| "grad_norm": 0.8323139648125473, | |
| "learning_rate": 9.933038743727749e-06, | |
| "loss": 0.7882, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.4580308761885288, | |
| "grad_norm": 0.8203697180665835, | |
| "learning_rate": 9.916606930010765e-06, | |
| "loss": 0.7794, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.4580308761885288, | |
| "eval_loss": 0.7901803255081177, | |
| "eval_runtime": 8.873, | |
| "eval_samples_per_second": 8.903, | |
| "eval_steps_per_second": 4.508, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.4743891217666905, | |
| "grad_norm": 0.7847835132710914, | |
| "learning_rate": 9.898390639982715e-06, | |
| "loss": 0.7771, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.49074736734485225, | |
| "grad_norm": 0.899857567905345, | |
| "learning_rate": 9.878396485223187e-06, | |
| "loss": 0.7759, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.49074736734485225, | |
| "eval_loss": 0.7846674919128418, | |
| "eval_runtime": 8.9104, | |
| "eval_samples_per_second": 8.866, | |
| "eval_steps_per_second": 4.489, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.507105612923014, | |
| "grad_norm": 0.8725377228611381, | |
| "learning_rate": 9.856631722585545e-06, | |
| "loss": 0.7694, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.5234638585011757, | |
| "grad_norm": 0.8975303080276704, | |
| "learning_rate": 9.833104251563058e-06, | |
| "loss": 0.7702, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.5234638585011757, | |
| "eval_loss": 0.7819108366966248, | |
| "eval_runtime": 9.1634, | |
| "eval_samples_per_second": 8.621, | |
| "eval_steps_per_second": 4.365, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.5398221040793375, | |
| "grad_norm": 0.8040447653131452, | |
| "learning_rate": 9.807822611421783e-06, | |
| "loss": 0.7737, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.5561803496574992, | |
| "grad_norm": 0.7791606284623682, | |
| "learning_rate": 9.780795978101265e-06, | |
| "loss": 0.7687, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.5561803496574992, | |
| "eval_loss": 0.7773299217224121, | |
| "eval_runtime": 8.8107, | |
| "eval_samples_per_second": 8.966, | |
| "eval_steps_per_second": 4.54, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.572538595235661, | |
| "grad_norm": 0.7841990500752456, | |
| "learning_rate": 9.752034160884126e-06, | |
| "loss": 0.7685, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.5888968408138228, | |
| "grad_norm": 0.8046795669671181, | |
| "learning_rate": 9.721547598835807e-06, | |
| "loss": 0.7481, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.5888968408138228, | |
| "eval_loss": 0.774566113948822, | |
| "eval_runtime": 9.0394, | |
| "eval_samples_per_second": 8.739, | |
| "eval_steps_per_second": 4.425, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.6052550863919844, | |
| "grad_norm": 1.183625080940533, | |
| "learning_rate": 9.689347357015721e-06, | |
| "loss": 0.748, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.6216133319701462, | |
| "grad_norm": 0.8040310134429428, | |
| "learning_rate": 9.6554451224612e-06, | |
| "loss": 0.7548, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.6216133319701462, | |
| "eval_loss": 0.7700055241584778, | |
| "eval_runtime": 8.8572, | |
| "eval_samples_per_second": 8.919, | |
| "eval_steps_per_second": 4.516, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.6379715775483079, | |
| "grad_norm": 0.9470332937426067, | |
| "learning_rate": 9.619853199945699e-06, | |
| "loss": 0.7517, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.6543298231264697, | |
| "grad_norm": 0.8950375364388944, | |
| "learning_rate": 9.5825845075128e-06, | |
| "loss": 0.7502, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.6543298231264697, | |
| "eval_loss": 0.7674439549446106, | |
| "eval_runtime": 8.9218, | |
| "eval_samples_per_second": 8.855, | |
| "eval_steps_per_second": 4.483, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.6706880687046314, | |
| "grad_norm": 0.9169203286801689, | |
| "learning_rate": 9.543652571787623e-06, | |
| "loss": 0.7496, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.6870463142827932, | |
| "grad_norm": 0.8093191265332849, | |
| "learning_rate": 9.503071523067346e-06, | |
| "loss": 0.7544, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.6870463142827932, | |
| "eval_loss": 0.7653623223304749, | |
| "eval_runtime": 8.8891, | |
| "eval_samples_per_second": 8.887, | |
| "eval_steps_per_second": 4.5, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.703404559860955, | |
| "grad_norm": 0.8858043762442335, | |
| "learning_rate": 9.460856090192643e-06, | |
| "loss": 0.7455, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.7197628054391166, | |
| "grad_norm": 0.7155973463395218, | |
| "learning_rate": 9.417021595201864e-06, | |
| "loss": 0.7527, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.7197628054391166, | |
| "eval_loss": 0.7620683908462524, | |
| "eval_runtime": 8.8784, | |
| "eval_samples_per_second": 8.898, | |
| "eval_steps_per_second": 4.505, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.7361210510172784, | |
| "grad_norm": 0.737512769355951, | |
| "learning_rate": 9.371583947769914e-06, | |
| "loss": 0.7457, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.7524792965954401, | |
| "grad_norm": 0.8059791945207864, | |
| "learning_rate": 9.324559639433874e-06, | |
| "loss": 0.7498, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.7524792965954401, | |
| "eval_loss": 0.7595985531806946, | |
| "eval_runtime": 8.8587, | |
| "eval_samples_per_second": 8.918, | |
| "eval_steps_per_second": 4.515, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.7688375421736019, | |
| "grad_norm": 0.8453208453911044, | |
| "learning_rate": 9.2759657376074e-06, | |
| "loss": 0.7399, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.7851957877517636, | |
| "grad_norm": 0.7794080969117457, | |
| "learning_rate": 9.225819879386137e-06, | |
| "loss": 0.7422, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.7851957877517636, | |
| "eval_loss": 0.757483959197998, | |
| "eval_runtime": 8.8082, | |
| "eval_samples_per_second": 8.969, | |
| "eval_steps_per_second": 4.541, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.8015540333299254, | |
| "grad_norm": 0.7170361500782209, | |
| "learning_rate": 9.174140265146356e-06, | |
| "loss": 0.7366, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.8179122789080872, | |
| "grad_norm": 0.886247962125953, | |
| "learning_rate": 9.120945651939146e-06, | |
| "loss": 0.7385, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.8179122789080872, | |
| "eval_loss": 0.7534403800964355, | |
| "eval_runtime": 8.8411, | |
| "eval_samples_per_second": 8.936, | |
| "eval_steps_per_second": 4.524, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.8342705244862488, | |
| "grad_norm": 0.7303115546636174, | |
| "learning_rate": 9.066255346682567e-06, | |
| "loss": 0.7426, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.8506287700644106, | |
| "grad_norm": 0.720904332688552, | |
| "learning_rate": 9.010089199154238e-06, | |
| "loss": 0.7396, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.8506287700644106, | |
| "eval_loss": 0.7510408163070679, | |
| "eval_runtime": 8.8187, | |
| "eval_samples_per_second": 8.958, | |
| "eval_steps_per_second": 4.536, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.8669870156425723, | |
| "grad_norm": 0.7269189612270449, | |
| "learning_rate": 8.952467594786867e-06, | |
| "loss": 0.7372, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.8833452612207341, | |
| "grad_norm": 0.7411165689482159, | |
| "learning_rate": 8.8934114472694e-06, | |
| "loss": 0.7455, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.8833452612207341, | |
| "eval_loss": 0.7494012117385864, | |
| "eval_runtime": 8.8703, | |
| "eval_samples_per_second": 8.906, | |
| "eval_steps_per_second": 4.509, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.8997035067988958, | |
| "grad_norm": 0.7222933451317762, | |
| "learning_rate": 8.83294219095642e-06, | |
| "loss": 0.7396, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.9160617523770576, | |
| "grad_norm": 0.7405492243894416, | |
| "learning_rate": 8.771081773088568e-06, | |
| "loss": 0.7275, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.9160617523770576, | |
| "eval_loss": 0.747172474861145, | |
| "eval_runtime": 8.8684, | |
| "eval_samples_per_second": 8.908, | |
| "eval_steps_per_second": 4.51, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.9324199979552193, | |
| "grad_norm": 0.7343527467764037, | |
| "learning_rate": 8.707852645826824e-06, | |
| "loss": 0.7227, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.948778243533381, | |
| "grad_norm": 0.8149960345815405, | |
| "learning_rate": 8.643277758103527e-06, | |
| "loss": 0.74, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.948778243533381, | |
| "eval_loss": 0.745248019695282, | |
| "eval_runtime": 8.8473, | |
| "eval_samples_per_second": 8.929, | |
| "eval_steps_per_second": 4.521, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.9651364891115428, | |
| "grad_norm": 0.7210978276382376, | |
| "learning_rate": 8.577380547293089e-06, | |
| "loss": 0.7387, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.9814947346897045, | |
| "grad_norm": 0.7095208468673599, | |
| "learning_rate": 8.510184930705413e-06, | |
| "loss": 0.7339, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.9814947346897045, | |
| "eval_loss": 0.7437242269515991, | |
| "eval_runtime": 8.82, | |
| "eval_samples_per_second": 8.957, | |
| "eval_steps_per_second": 4.535, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.9978529802678663, | |
| "grad_norm": 0.7505905361093098, | |
| "learning_rate": 8.44171529690516e-06, | |
| "loss": 0.7248, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.0143134648808916, | |
| "grad_norm": 0.6979646401902959, | |
| "learning_rate": 8.371996496859938e-06, | |
| "loss": 0.7404, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.0143134648808916, | |
| "eval_loss": 0.7440437078475952, | |
| "eval_runtime": 8.7724, | |
| "eval_samples_per_second": 9.005, | |
| "eval_steps_per_second": 4.56, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.0306717104590533, | |
| "grad_norm": 0.7236767629054202, | |
| "learning_rate": 8.30105383492067e-06, | |
| "loss": 0.6848, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.047029956037215, | |
| "grad_norm": 0.6872848093215694, | |
| "learning_rate": 8.228913059637414e-06, | |
| "loss": 0.6877, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.047029956037215, | |
| "eval_loss": 0.7421544790267944, | |
| "eval_runtime": 8.8557, | |
| "eval_samples_per_second": 8.921, | |
| "eval_steps_per_second": 4.517, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.0633882016153768, | |
| "grad_norm": 0.7171174222919368, | |
| "learning_rate": 8.155600354413971e-06, | |
| "loss": 0.6752, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.0797464471935385, | |
| "grad_norm": 0.6937193535340093, | |
| "learning_rate": 8.081142328004638e-06, | |
| "loss": 0.6785, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.0797464471935385, | |
| "eval_loss": 0.7413665056228638, | |
| "eval_runtime": 8.7177, | |
| "eval_samples_per_second": 9.062, | |
| "eval_steps_per_second": 4.588, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.0961046927717002, | |
| "grad_norm": 0.7347355982975387, | |
| "learning_rate": 8.005566004856593e-06, | |
| "loss": 0.6746, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.1124629383498619, | |
| "grad_norm": 0.8465986634067116, | |
| "learning_rate": 7.928898815301418e-06, | |
| "loss": 0.6723, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.1124629383498619, | |
| "eval_loss": 0.7399634122848511, | |
| "eval_runtime": 8.9138, | |
| "eval_samples_per_second": 8.863, | |
| "eval_steps_per_second": 4.487, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.1288211839280238, | |
| "grad_norm": 0.7674101027640446, | |
| "learning_rate": 7.851168585599285e-06, | |
| "loss": 0.6756, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.1451794295061855, | |
| "grad_norm": 0.7525909134460642, | |
| "learning_rate": 7.77240352783945e-06, | |
| "loss": 0.6807, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.1451794295061855, | |
| "eval_loss": 0.7387439608573914, | |
| "eval_runtime": 8.8038, | |
| "eval_samples_per_second": 8.973, | |
| "eval_steps_per_second": 4.544, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.1615376750843471, | |
| "grad_norm": 0.7798312578850733, | |
| "learning_rate": 7.692632229700718e-06, | |
| "loss": 0.6779, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.177895920662509, | |
| "grad_norm": 0.7474295968254813, | |
| "learning_rate": 7.611883644075573e-06, | |
| "loss": 0.6823, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.177895920662509, | |
| "eval_loss": 0.7377986907958984, | |
| "eval_runtime": 8.8162, | |
| "eval_samples_per_second": 8.961, | |
| "eval_steps_per_second": 4.537, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.1942541662406707, | |
| "grad_norm": 0.7587723123453163, | |
| "learning_rate": 7.5301870785617635e-06, | |
| "loss": 0.675, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.2106124118188324, | |
| "grad_norm": 0.7440613541332155, | |
| "learning_rate": 7.447572184825149e-06, | |
| "loss": 0.6792, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.2106124118188324, | |
| "eval_loss": 0.7359431982040405, | |
| "eval_runtime": 8.8106, | |
| "eval_samples_per_second": 8.966, | |
| "eval_steps_per_second": 4.54, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.2269706573969943, | |
| "grad_norm": 0.7463811292658835, | |
| "learning_rate": 7.3640689478376695e-06, | |
| "loss": 0.682, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.243328902975156, | |
| "grad_norm": 0.7709021578765768, | |
| "learning_rate": 7.2797076749943214e-06, | |
| "loss": 0.6799, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.243328902975156, | |
| "eval_loss": 0.7348581552505493, | |
| "eval_runtime": 8.8852, | |
| "eval_samples_per_second": 8.891, | |
| "eval_steps_per_second": 4.502, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.2596871485533176, | |
| "grad_norm": 0.7343520193643218, | |
| "learning_rate": 7.194518985113149e-06, | |
| "loss": 0.6765, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.2760453941314793, | |
| "grad_norm": 0.7190112773845664, | |
| "learning_rate": 7.108533797322166e-06, | |
| "loss": 0.6804, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.2760453941314793, | |
| "eval_loss": 0.7342889904975891, | |
| "eval_runtime": 8.8721, | |
| "eval_samples_per_second": 8.904, | |
| "eval_steps_per_second": 4.508, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.292403639709641, | |
| "grad_norm": 0.7133811365918709, | |
| "learning_rate": 7.021783319837298e-06, | |
| "loss": 0.6767, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.308761885287803, | |
| "grad_norm": 0.7307976869448247, | |
| "learning_rate": 6.934299038635414e-06, | |
| "loss": 0.6812, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.308761885287803, | |
| "eval_loss": 0.7325922846794128, | |
| "eval_runtime": 8.8154, | |
| "eval_samples_per_second": 8.962, | |
| "eval_steps_per_second": 4.538, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.3251201308659646, | |
| "grad_norm": 0.7219505844302012, | |
| "learning_rate": 6.8461127060265135e-06, | |
| "loss": 0.679, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.3414783764441265, | |
| "grad_norm": 0.7057903127014294, | |
| "learning_rate": 6.7572563291292724e-06, | |
| "loss": 0.6714, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.3414783764441265, | |
| "eval_loss": 0.7310522794723511, | |
| "eval_runtime": 8.9723, | |
| "eval_samples_per_second": 8.805, | |
| "eval_steps_per_second": 4.458, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.3578366220222882, | |
| "grad_norm": 0.7481237758190087, | |
| "learning_rate": 6.667762158254104e-06, | |
| "loss": 0.671, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.3741948676004498, | |
| "grad_norm": 0.6771887136482816, | |
| "learning_rate": 6.57766267519794e-06, | |
| "loss": 0.6764, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.3741948676004498, | |
| "eval_loss": 0.7305134534835815, | |
| "eval_runtime": 8.7821, | |
| "eval_samples_per_second": 8.996, | |
| "eval_steps_per_second": 4.555, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.3905531131786115, | |
| "grad_norm": 0.7720857727869926, | |
| "learning_rate": 6.486990581455011e-06, | |
| "loss": 0.6819, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.4069113587567732, | |
| "grad_norm": 0.7488122619323809, | |
| "learning_rate": 6.395778786347878e-06, | |
| "loss": 0.6801, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.4069113587567732, | |
| "eval_loss": 0.7291988134384155, | |
| "eval_runtime": 8.8848, | |
| "eval_samples_per_second": 8.892, | |
| "eval_steps_per_second": 4.502, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.423269604334935, | |
| "grad_norm": 0.7831777981437086, | |
| "learning_rate": 6.304060395083031e-06, | |
| "loss": 0.6687, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.4396278499130968, | |
| "grad_norm": 0.7391424956481155, | |
| "learning_rate": 6.211868696735405e-06, | |
| "loss": 0.6693, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.4396278499130968, | |
| "eval_loss": 0.727660059928894, | |
| "eval_runtime": 8.8922, | |
| "eval_samples_per_second": 8.884, | |
| "eval_steps_per_second": 4.498, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.4559860954912587, | |
| "grad_norm": 0.6633367855169299, | |
| "learning_rate": 6.119237152166139e-06, | |
| "loss": 0.6657, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.4723443410694204, | |
| "grad_norm": 0.6920250318389223, | |
| "learning_rate": 6.026199381878011e-06, | |
| "loss": 0.6726, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.4723443410694204, | |
| "eval_loss": 0.7265572547912598, | |
| "eval_runtime": 8.8512, | |
| "eval_samples_per_second": 8.925, | |
| "eval_steps_per_second": 4.519, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.488702586647582, | |
| "grad_norm": 0.7081219959408172, | |
| "learning_rate": 5.932789153812911e-06, | |
| "loss": 0.6672, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.5050608322257437, | |
| "grad_norm": 0.7707103364304354, | |
| "learning_rate": 5.839040371095814e-06, | |
| "loss": 0.6762, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.5050608322257437, | |
| "eval_loss": 0.7252368330955505, | |
| "eval_runtime": 8.8447, | |
| "eval_samples_per_second": 8.932, | |
| "eval_steps_per_second": 4.522, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.5214190778039054, | |
| "grad_norm": 0.7285700753630275, | |
| "learning_rate": 5.74498705972968e-06, | |
| "loss": 0.6716, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.5377773233820673, | |
| "grad_norm": 0.6719863494102282, | |
| "learning_rate": 5.650663356245756e-06, | |
| "loss": 0.6666, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.5377773233820673, | |
| "eval_loss": 0.7246092557907104, | |
| "eval_runtime": 8.7654, | |
| "eval_samples_per_second": 9.013, | |
| "eval_steps_per_second": 4.563, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.554135568960229, | |
| "grad_norm": 0.697927897632164, | |
| "learning_rate": 5.556103495313774e-06, | |
| "loss": 0.6679, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.5704938145383909, | |
| "grad_norm": 0.6843159800619036, | |
| "learning_rate": 5.46134179731651e-06, | |
| "loss": 0.6713, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.5704938145383909, | |
| "eval_loss": 0.7233731150627136, | |
| "eval_runtime": 8.9372, | |
| "eval_samples_per_second": 8.839, | |
| "eval_steps_per_second": 4.476, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.5868520601165526, | |
| "grad_norm": 0.6766351504145809, | |
| "learning_rate": 5.366412655893256e-06, | |
| "loss": 0.6689, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.6032103056947142, | |
| "grad_norm": 0.7167144088999912, | |
| "learning_rate": 5.271350525456679e-06, | |
| "loss": 0.6659, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.6032103056947142, | |
| "eval_loss": 0.7226489186286926, | |
| "eval_runtime": 8.789, | |
| "eval_samples_per_second": 8.989, | |
| "eval_steps_per_second": 4.551, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.619568551272876, | |
| "grad_norm": 0.7126879315212699, | |
| "learning_rate": 5.176189908687646e-06, | |
| "loss": 0.6695, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.6359267968510376, | |
| "grad_norm": 0.6583586257751949, | |
| "learning_rate": 5.080965344012509e-06, | |
| "loss": 0.6694, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.6359267968510376, | |
| "eval_loss": 0.7216107249259949, | |
| "eval_runtime": 8.8487, | |
| "eval_samples_per_second": 8.928, | |
| "eval_steps_per_second": 4.52, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.6522850424291995, | |
| "grad_norm": 0.6561297313170917, | |
| "learning_rate": 4.985711393067437e-06, | |
| "loss": 0.6674, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.6686432880073612, | |
| "grad_norm": 0.6626437713636341, | |
| "learning_rate": 4.890462628154309e-06, | |
| "loss": 0.6667, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.6686432880073612, | |
| "eval_loss": 0.7210881114006042, | |
| "eval_runtime": 8.7877, | |
| "eval_samples_per_second": 8.99, | |
| "eval_steps_per_second": 4.552, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.685001533585523, | |
| "grad_norm": 0.7166433188372661, | |
| "learning_rate": 4.795253619692737e-06, | |
| "loss": 0.663, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.7013597791636847, | |
| "grad_norm": 0.741234806048152, | |
| "learning_rate": 4.700118923672779e-06, | |
| "loss": 0.6739, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.7013597791636847, | |
| "eval_loss": 0.7206884622573853, | |
| "eval_runtime": 8.8528, | |
| "eval_samples_per_second": 8.924, | |
| "eval_steps_per_second": 4.518, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.7177180247418464, | |
| "grad_norm": 0.7016761274840421, | |
| "learning_rate": 4.605093069112892e-06, | |
| "loss": 0.6711, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.734076270320008, | |
| "grad_norm": 0.6970244053157433, | |
| "learning_rate": 4.5102105455276525e-06, | |
| "loss": 0.6745, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.734076270320008, | |
| "eval_loss": 0.7203969955444336, | |
| "eval_runtime": 8.8182, | |
| "eval_samples_per_second": 8.959, | |
| "eval_steps_per_second": 4.536, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.7504345158981698, | |
| "grad_norm": 0.7335723892483486, | |
| "learning_rate": 4.415505790409847e-06, | |
| "loss": 0.6686, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.7667927614763317, | |
| "grad_norm": 0.6956225175618133, | |
| "learning_rate": 4.321013176731414e-06, | |
| "loss": 0.6665, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.7667927614763317, | |
| "eval_loss": 0.7191005945205688, | |
| "eval_runtime": 8.9226, | |
| "eval_samples_per_second": 8.854, | |
| "eval_steps_per_second": 4.483, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.7831510070544934, | |
| "grad_norm": 0.698587154786482, | |
| "learning_rate": 4.226767000467826e-06, | |
| "loss": 0.6666, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.7995092526326553, | |
| "grad_norm": 0.7438773573796561, | |
| "learning_rate": 4.1328014681504105e-06, | |
| "loss": 0.6626, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.7995092526326553, | |
| "eval_loss": 0.7183343768119812, | |
| "eval_runtime": 8.8586, | |
| "eval_samples_per_second": 8.918, | |
| "eval_steps_per_second": 4.515, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.815867498210817, | |
| "grad_norm": 0.6947292310543136, | |
| "learning_rate": 4.039150684451117e-06, | |
| "loss": 0.66, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.8322257437889786, | |
| "grad_norm": 0.6477593004868593, | |
| "learning_rate": 3.945848639804287e-06, | |
| "loss": 0.6583, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.8322257437889786, | |
| "eval_loss": 0.7178105711936951, | |
| "eval_runtime": 8.8059, | |
| "eval_samples_per_second": 8.971, | |
| "eval_steps_per_second": 4.542, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.8485839893671403, | |
| "grad_norm": 0.6753153859283428, | |
| "learning_rate": 3.852929198069856e-06, | |
| "loss": 0.6551, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.864942234945302, | |
| "grad_norm": 0.734130420389638, | |
| "learning_rate": 3.7604260842425146e-06, | |
| "loss": 0.665, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.864942234945302, | |
| "eval_loss": 0.7165009379386902, | |
| "eval_runtime": 8.8639, | |
| "eval_samples_per_second": 8.913, | |
| "eval_steps_per_second": 4.513, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.8813004805234639, | |
| "grad_norm": 0.6466537300434777, | |
| "learning_rate": 3.668372872211266e-06, | |
| "loss": 0.6601, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.8976587261016256, | |
| "grad_norm": 0.6650377275740064, | |
| "learning_rate": 3.5768029725738157e-06, | |
| "loss": 0.6609, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.8976587261016256, | |
| "eval_loss": 0.7164144515991211, | |
| "eval_runtime": 8.8323, | |
| "eval_samples_per_second": 8.944, | |
| "eval_steps_per_second": 4.529, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.9140169716797875, | |
| "grad_norm": 0.6705397336213909, | |
| "learning_rate": 3.4857496205102475e-06, | |
| "loss": 0.6558, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.9303752172579491, | |
| "grad_norm": 0.6821158056706338, | |
| "learning_rate": 3.3952458637203475e-06, | |
| "loss": 0.6715, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.9303752172579491, | |
| "eval_loss": 0.715703010559082, | |
| "eval_runtime": 8.8492, | |
| "eval_samples_per_second": 8.927, | |
| "eval_steps_per_second": 4.52, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.9467334628361108, | |
| "grad_norm": 0.6630764255722235, | |
| "learning_rate": 3.3053245504289894e-06, | |
| "loss": 0.656, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.9630917084142725, | |
| "grad_norm": 0.669639110840792, | |
| "learning_rate": 3.216018317463914e-06, | |
| "loss": 0.6755, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.9630917084142725, | |
| "eval_loss": 0.7149642109870911, | |
| "eval_runtime": 8.6929, | |
| "eval_samples_per_second": 9.088, | |
| "eval_steps_per_second": 4.601, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.9794499539924342, | |
| "grad_norm": 0.677640160297301, | |
| "learning_rate": 3.1273595784102244e-06, | |
| "loss": 0.6656, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.995808199570596, | |
| "grad_norm": 0.6478481336084418, | |
| "learning_rate": 3.0393805118459257e-06, | |
| "loss": 0.6654, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.995808199570596, | |
| "eval_loss": 0.7144020795822144, | |
| "eval_runtime": 8.7728, | |
| "eval_samples_per_second": 9.005, | |
| "eval_steps_per_second": 4.56, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 2.0122686841836215, | |
| "grad_norm": 0.7160988089736929, | |
| "learning_rate": 2.952113049662744e-06, | |
| "loss": 0.6911, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 2.028626929761783, | |
| "grad_norm": 0.6658394367051226, | |
| "learning_rate": 2.86558886547649e-06, | |
| "loss": 0.6207, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 2.028626929761783, | |
| "eval_loss": 0.7192620038986206, | |
| "eval_runtime": 8.7807, | |
| "eval_samples_per_second": 8.997, | |
| "eval_steps_per_second": 4.555, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 2.044985175339945, | |
| "grad_norm": 0.654677579170524, | |
| "learning_rate": 2.7798393631311565e-06, | |
| "loss": 0.6216, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 2.0613434209181065, | |
| "grad_norm": 0.6845236856939948, | |
| "learning_rate": 2.6948956653009416e-06, | |
| "loss": 0.618, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.0613434209181065, | |
| "eval_loss": 0.7191548347473145, | |
| "eval_runtime": 8.8502, | |
| "eval_samples_per_second": 8.926, | |
| "eval_steps_per_second": 4.52, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.077701666496268, | |
| "grad_norm": 0.6800346983315418, | |
| "learning_rate": 2.6107886021942984e-06, | |
| "loss": 0.6219, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 2.09405991207443, | |
| "grad_norm": 0.6473993386866002, | |
| "learning_rate": 2.527548700364164e-06, | |
| "loss": 0.6182, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.09405991207443, | |
| "eval_loss": 0.7187679409980774, | |
| "eval_runtime": 8.8369, | |
| "eval_samples_per_second": 8.94, | |
| "eval_steps_per_second": 4.526, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.1104181576525916, | |
| "grad_norm": 0.680332157190334, | |
| "learning_rate": 2.4452061716283764e-06, | |
| "loss": 0.6247, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 2.1267764032307537, | |
| "grad_norm": 0.6713647024120337, | |
| "learning_rate": 2.3637909021043332e-06, | |
| "loss": 0.6188, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.1267764032307537, | |
| "eval_loss": 0.7195940017700195, | |
| "eval_runtime": 8.934, | |
| "eval_samples_per_second": 8.843, | |
| "eval_steps_per_second": 4.477, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.1431346488089154, | |
| "grad_norm": 0.6905966915530017, | |
| "learning_rate": 2.2833324413618695e-06, | |
| "loss": 0.6213, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 2.159492894387077, | |
| "grad_norm": 0.6961837258954594, | |
| "learning_rate": 2.2038599916982775e-06, | |
| "loss": 0.6173, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 2.159492894387077, | |
| "eval_loss": 0.718852162361145, | |
| "eval_runtime": 8.7975, | |
| "eval_samples_per_second": 8.98, | |
| "eval_steps_per_second": 4.547, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 2.1758511399652387, | |
| "grad_norm": 0.7109159972241073, | |
| "learning_rate": 2.1254023975393585e-06, | |
| "loss": 0.6193, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 2.1922093855434004, | |
| "grad_norm": 0.6856177780539675, | |
| "learning_rate": 2.0479881349703885e-06, | |
| "loss": 0.6206, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.1922093855434004, | |
| "eval_loss": 0.719401478767395, | |
| "eval_runtime": 9.6891, | |
| "eval_samples_per_second": 8.154, | |
| "eval_steps_per_second": 4.128, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.208567631121562, | |
| "grad_norm": 0.6846382708470592, | |
| "learning_rate": 1.9716453014007493e-06, | |
| "loss": 0.6172, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.2249258766997237, | |
| "grad_norm": 0.6821924271490651, | |
| "learning_rate": 1.8964016053660167e-06, | |
| "loss": 0.6206, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.2249258766997237, | |
| "eval_loss": 0.7185901403427124, | |
| "eval_runtime": 10.538, | |
| "eval_samples_per_second": 7.497, | |
| "eval_steps_per_second": 3.796, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.241284122277886, | |
| "grad_norm": 0.7020054137197129, | |
| "learning_rate": 1.822284356471179e-06, | |
| "loss": 0.628, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 2.2576423678560475, | |
| "grad_norm": 0.6827467082055568, | |
| "learning_rate": 1.7493204554786453e-06, | |
| "loss": 0.6192, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.2576423678560475, | |
| "eval_loss": 0.7185772657394409, | |
| "eval_runtime": 8.7402, | |
| "eval_samples_per_second": 9.039, | |
| "eval_steps_per_second": 4.577, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.2740006134342092, | |
| "grad_norm": 0.6787057813877333, | |
| "learning_rate": 1.6775363845446552e-06, | |
| "loss": 0.6215, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 2.290358859012371, | |
| "grad_norm": 0.6965105305161102, | |
| "learning_rate": 1.6069581976076059e-06, | |
| "loss": 0.6168, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.290358859012371, | |
| "eval_loss": 0.7182484269142151, | |
| "eval_runtime": 8.8061, | |
| "eval_samples_per_second": 8.971, | |
| "eval_steps_per_second": 4.542, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.3067171045905326, | |
| "grad_norm": 0.657899834231394, | |
| "learning_rate": 1.537611510931819e-06, | |
| "loss": 0.622, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 2.3230753501686943, | |
| "grad_norm": 0.6763933278225029, | |
| "learning_rate": 1.4695214938101394e-06, | |
| "loss": 0.6247, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 2.3230753501686943, | |
| "eval_loss": 0.7175112962722778, | |
| "eval_runtime": 8.833, | |
| "eval_samples_per_second": 8.944, | |
| "eval_steps_per_second": 4.528, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 2.339433595746856, | |
| "grad_norm": 0.6445654002256987, | |
| "learning_rate": 1.4027128594287743e-06, | |
| "loss": 0.6124, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 2.355791841325018, | |
| "grad_norm": 0.6483502135699247, | |
| "learning_rate": 1.3372098558976753e-06, | |
| "loss": 0.6152, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 2.355791841325018, | |
| "eval_loss": 0.7176142930984497, | |
| "eval_runtime": 8.7696, | |
| "eval_samples_per_second": 9.008, | |
| "eval_steps_per_second": 4.561, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 2.3721500869031797, | |
| "grad_norm": 0.6748060520638486, | |
| "learning_rate": 1.2730362574497146e-06, | |
| "loss": 0.6191, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 2.3885083324813414, | |
| "grad_norm": 0.6411389753036478, | |
| "learning_rate": 1.2102153558118635e-06, | |
| "loss": 0.6118, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 2.3885083324813414, | |
| "eval_loss": 0.7174015641212463, | |
| "eval_runtime": 8.8824, | |
| "eval_samples_per_second": 8.894, | |
| "eval_steps_per_second": 4.503, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 2.404866578059503, | |
| "grad_norm": 0.6147421703730933, | |
| "learning_rate": 1.1487699517514828e-06, | |
| "loss": 0.6213, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 2.421224823637665, | |
| "grad_norm": 0.6564011388418958, | |
| "learning_rate": 1.088722346800813e-06, | |
| "loss": 0.6192, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 2.421224823637665, | |
| "eval_loss": 0.7174366116523743, | |
| "eval_runtime": 8.8213, | |
| "eval_samples_per_second": 8.956, | |
| "eval_steps_per_second": 4.534, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 2.4375830692158265, | |
| "grad_norm": 0.6637159235119273, | |
| "learning_rate": 1.0300943351626642e-06, | |
| "loss": 0.6172, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 2.4539413147939886, | |
| "grad_norm": 0.6399937055494419, | |
| "learning_rate": 9.729071958002356e-07, | |
| "loss": 0.6125, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 2.4539413147939886, | |
| "eval_loss": 0.7172605395317078, | |
| "eval_runtime": 8.8588, | |
| "eval_samples_per_second": 8.918, | |
| "eval_steps_per_second": 4.515, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 2.4702995603721503, | |
| "grad_norm": 0.6548848015845544, | |
| "learning_rate": 9.171816847139447e-07, | |
| "loss": 0.6197, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 2.486657805950312, | |
| "grad_norm": 0.652681143911981, | |
| "learning_rate": 8.62938027408064e-07, | |
| "loss": 0.6067, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 2.486657805950312, | |
| "eval_loss": 0.7170479893684387, | |
| "eval_runtime": 8.821, | |
| "eval_samples_per_second": 8.956, | |
| "eval_steps_per_second": 4.535, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 2.5030160515284736, | |
| "grad_norm": 0.6223909947895746, | |
| "learning_rate": 8.10195911549892e-07, | |
| "loss": 0.6229, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 2.5193742971066353, | |
| "grad_norm": 0.6675446368924337, | |
| "learning_rate": 7.589744798241472e-07, | |
| "loss": 0.6196, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 2.5193742971066353, | |
| "eval_loss": 0.7164150476455688, | |
| "eval_runtime": 8.8151, | |
| "eval_samples_per_second": 8.962, | |
| "eval_steps_per_second": 4.538, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 2.535732542684797, | |
| "grad_norm": 0.6748594529596444, | |
| "learning_rate": 7.092923229851506e-07, | |
| "loss": 0.6259, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 2.5520907882629587, | |
| "grad_norm": 0.6436382148035624, | |
| "learning_rate": 6.611674731093376e-07, | |
| "loss": 0.6147, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 2.5520907882629587, | |
| "eval_loss": 0.7163916826248169, | |
| "eval_runtime": 8.7789, | |
| "eval_samples_per_second": 8.999, | |
| "eval_steps_per_second": 4.556, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 2.5684490338411203, | |
| "grad_norm": 0.6954982109692219, | |
| "learning_rate": 6.146173970505387e-07, | |
| "loss": 0.6221, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 2.584807279419282, | |
| "grad_norm": 0.6636316314884414, | |
| "learning_rate": 5.696589901004001e-07, | |
| "loss": 0.6166, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 2.584807279419282, | |
| "eval_loss": 0.7162200808525085, | |
| "eval_runtime": 8.8522, | |
| "eval_samples_per_second": 8.924, | |
| "eval_steps_per_second": 4.519, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 2.601165524997444, | |
| "grad_norm": 0.6348189748255966, | |
| "learning_rate": 5.263085698562675e-07, | |
| "loss": 0.6162, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 2.617523770575606, | |
| "grad_norm": 0.6656187164775476, | |
| "learning_rate": 4.845818702987287e-07, | |
| "loss": 0.6235, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 2.617523770575606, | |
| "eval_loss": 0.7161623239517212, | |
| "eval_runtime": 8.9168, | |
| "eval_samples_per_second": 8.86, | |
| "eval_steps_per_second": 4.486, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 2.6338820161537675, | |
| "grad_norm": 0.644537564705373, | |
| "learning_rate": 4.4449403608098965e-07, | |
| "loss": 0.6126, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 2.650240261731929, | |
| "grad_norm": 0.6394762326342863, | |
| "learning_rate": 4.0605961703213845e-07, | |
| "loss": 0.6112, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 2.650240261731929, | |
| "eval_loss": 0.7161468267440796, | |
| "eval_runtime": 8.9454, | |
| "eval_samples_per_second": 8.831, | |
| "eval_steps_per_second": 4.472, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 2.666598507310091, | |
| "grad_norm": 0.6651958412760407, | |
| "learning_rate": 3.692925628763033e-07, | |
| "loss": 0.6218, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 2.682956752888253, | |
| "grad_norm": 0.6547622124607134, | |
| "learning_rate": 3.3420621816961874e-07, | |
| "loss": 0.6186, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 2.682956752888253, | |
| "eval_loss": 0.7159026265144348, | |
| "eval_runtime": 8.8156, | |
| "eval_samples_per_second": 8.961, | |
| "eval_steps_per_second": 4.537, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 2.6993149984664147, | |
| "grad_norm": 0.6391386777884149, | |
| "learning_rate": 3.0081331745683395e-07, | |
| "loss": 0.613, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 2.7156732440445763, | |
| "grad_norm": 0.6368321437722378, | |
| "learning_rate": 2.691259806493296e-07, | |
| "loss": 0.6164, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 2.7156732440445763, | |
| "eval_loss": 0.7158184051513672, | |
| "eval_runtime": 8.7951, | |
| "eval_samples_per_second": 8.982, | |
| "eval_steps_per_second": 4.548, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 2.732031489622738, | |
| "grad_norm": 0.6365243949451423, | |
| "learning_rate": 2.391557086262086e-07, | |
| "loss": 0.6228, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 2.7483897352008997, | |
| "grad_norm": 0.6321404050434235, | |
| "learning_rate": 2.109133790600648e-07, | |
| "loss": 0.6147, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 2.7483897352008997, | |
| "eval_loss": 0.7158446311950684, | |
| "eval_runtime": 8.8513, | |
| "eval_samples_per_second": 8.925, | |
| "eval_steps_per_second": 4.519, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 2.7647479807790614, | |
| "grad_norm": 0.6903153602015651, | |
| "learning_rate": 1.8440924246894753e-07, | |
| "loss": 0.6158, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 2.781106226357223, | |
| "grad_norm": 0.6670191692486903, | |
| "learning_rate": 1.5965291849594944e-07, | |
| "loss": 0.6098, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 2.781106226357223, | |
| "eval_loss": 0.7158127427101135, | |
| "eval_runtime": 8.8596, | |
| "eval_samples_per_second": 8.917, | |
| "eval_steps_per_second": 4.515, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 2.7974644719353847, | |
| "grad_norm": 0.6357999130773923, | |
| "learning_rate": 1.3665339241776765e-07, | |
| "loss": 0.6212, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 2.8138227175135464, | |
| "grad_norm": 0.6630012411471573, | |
| "learning_rate": 1.1541901188351134e-07, | |
| "loss": 0.6126, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 2.8138227175135464, | |
| "eval_loss": 0.7158257365226746, | |
| "eval_runtime": 9.6888, | |
| "eval_samples_per_second": 8.154, | |
| "eval_steps_per_second": 4.128, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 2.8301809630917085, | |
| "grad_norm": 0.6453855721831139, | |
| "learning_rate": 9.59574838849281e-08, | |
| "loss": 0.6162, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 2.84653920866987, | |
| "grad_norm": 0.647793683675027, | |
| "learning_rate": 7.827587195916697e-08, | |
| "loss": 0.6179, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 2.84653920866987, | |
| "eval_loss": 0.7156850695610046, | |
| "eval_runtime": 8.8101, | |
| "eval_samples_per_second": 8.967, | |
| "eval_steps_per_second": 4.54, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 2.862897454248032, | |
| "grad_norm": 0.627906665508174, | |
| "learning_rate": 6.238059362507043e-08, | |
| "loss": 0.6157, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 2.8792556998261936, | |
| "grad_norm": 0.6991824017680971, | |
| "learning_rate": 4.827741805395025e-08, | |
| "loss": 0.6229, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 2.8792556998261936, | |
| "eval_loss": 0.7156457901000977, | |
| "eval_runtime": 8.8917, | |
| "eval_samples_per_second": 8.885, | |
| "eval_steps_per_second": 4.499, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 2.8956139454043552, | |
| "grad_norm": 0.6604597808333147, | |
| "learning_rate": 3.59714639756692e-08, | |
| "loss": 0.6198, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 2.9119721909825174, | |
| "grad_norm": 0.6412019833547566, | |
| "learning_rate": 2.5467197820805977e-08, | |
| "loss": 0.624, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 2.9119721909825174, | |
| "eval_loss": 0.7153981328010559, | |
| "eval_runtime": 8.897, | |
| "eval_samples_per_second": 8.879, | |
| "eval_steps_per_second": 4.496, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 2.928330436560679, | |
| "grad_norm": 0.6539771563156503, | |
| "learning_rate": 1.6768432099570907e-08, | |
| "loss": 0.6181, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 2.9446886821388407, | |
| "grad_norm": 0.6484588822526755, | |
| "learning_rate": 9.878324018058327e-09, | |
| "loss": 0.6155, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 2.9446886821388407, | |
| "eval_loss": 0.7157188653945923, | |
| "eval_runtime": 8.8628, | |
| "eval_samples_per_second": 8.914, | |
| "eval_steps_per_second": 4.513, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 2.9610469277170024, | |
| "grad_norm": 0.6119368873622615, | |
| "learning_rate": 4.799374332344342e-09, | |
| "loss": 0.6189, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 2.977405173295164, | |
| "grad_norm": 0.6333710387576432, | |
| "learning_rate": 1.533426440839536e-09, | |
| "loss": 0.6211, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 2.977405173295164, | |
| "eval_loss": 0.7156268358230591, | |
| "eval_runtime": 8.9113, | |
| "eval_samples_per_second": 8.865, | |
| "eval_steps_per_second": 4.489, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 2.9937634188733258, | |
| "grad_norm": 0.6603951738557682, | |
| "learning_rate": 8.166571522916222e-11, | |
| "loss": 0.6176, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 2.9986708925467744, | |
| "step": 1833, | |
| "total_flos": 531028544847872.0, | |
| "train_loss": 0.7064719163714637, | |
| "train_runtime": 53129.6318, | |
| "train_samples_per_second": 2.209, | |
| "train_steps_per_second": 0.035 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 1833, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 20, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 531028544847872.0, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |