| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0213386832026263, | |
| "eval_steps": 500, | |
| "global_step": 1400, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.007295276308590188, | |
| "grad_norm": 0.9992304444313049, | |
| "learning_rate": 9.999671349822887e-06, | |
| "loss": 0.087, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.014590552617180376, | |
| "grad_norm": 1.0456160306930542, | |
| "learning_rate": 9.998685442495921e-06, | |
| "loss": 0.0757, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.021885828925770564, | |
| "grad_norm": 1.014310598373413, | |
| "learning_rate": 9.99704240762655e-06, | |
| "loss": 0.0901, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.029181105234360752, | |
| "grad_norm": 0.9391605257987976, | |
| "learning_rate": 9.994742461208251e-06, | |
| "loss": 0.0811, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.036476381542950936, | |
| "grad_norm": 0.7982305884361267, | |
| "learning_rate": 9.991785905592149e-06, | |
| "loss": 0.0818, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.04377165785154113, | |
| "grad_norm": 0.9946795701980591, | |
| "learning_rate": 9.988173129447251e-06, | |
| "loss": 0.0905, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.05106693416013131, | |
| "grad_norm": 0.8562415242195129, | |
| "learning_rate": 9.983904607709365e-06, | |
| "loss": 0.0831, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.058362210468721504, | |
| "grad_norm": 1.054521918296814, | |
| "learning_rate": 9.978980901518663e-06, | |
| "loss": 0.083, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.06565748677731169, | |
| "grad_norm": 1.0189253091812134, | |
| "learning_rate": 9.973402658145908e-06, | |
| "loss": 0.0891, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.07295276308590187, | |
| "grad_norm": 0.7745731472969055, | |
| "learning_rate": 9.96717061090737e-06, | |
| "loss": 0.0846, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.08024803939449207, | |
| "grad_norm": 0.8214731812477112, | |
| "learning_rate": 9.960285579068419e-06, | |
| "loss": 0.0895, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.08754331570308226, | |
| "grad_norm": 0.9114850759506226, | |
| "learning_rate": 9.95274846773583e-06, | |
| "loss": 0.0862, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.09483859201167244, | |
| "grad_norm": 1.216098427772522, | |
| "learning_rate": 9.944560267738792e-06, | |
| "loss": 0.0896, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.10213386832026262, | |
| "grad_norm": 0.7587252855300903, | |
| "learning_rate": 9.935722055498655e-06, | |
| "loss": 0.0888, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.10942914462885282, | |
| "grad_norm": 0.8791028261184692, | |
| "learning_rate": 9.92623499288743e-06, | |
| "loss": 0.0832, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.11672442093744301, | |
| "grad_norm": 1.133716344833374, | |
| "learning_rate": 9.916100327075038e-06, | |
| "loss": 0.0854, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.12401969724603319, | |
| "grad_norm": 0.7774285674095154, | |
| "learning_rate": 9.905319390365364e-06, | |
| "loss": 0.0873, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.13131497355462338, | |
| "grad_norm": 0.7029784321784973, | |
| "learning_rate": 9.893893600021112e-06, | |
| "loss": 0.0786, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.13861024986321357, | |
| "grad_norm": 1.0894945859909058, | |
| "learning_rate": 9.881824458077491e-06, | |
| "loss": 0.089, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.14590552617180375, | |
| "grad_norm": 0.7053947448730469, | |
| "learning_rate": 9.869113551144754e-06, | |
| "loss": 0.0843, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.15320080248039394, | |
| "grad_norm": 0.9423328638076782, | |
| "learning_rate": 9.85576255019963e-06, | |
| "loss": 0.0887, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.16049607878898414, | |
| "grad_norm": 1.1778377294540405, | |
| "learning_rate": 9.841773210365646e-06, | |
| "loss": 0.0939, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.1677913550975743, | |
| "grad_norm": 0.8132289052009583, | |
| "learning_rate": 9.82714737068241e-06, | |
| "loss": 0.0812, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.1750866314061645, | |
| "grad_norm": 0.925722062587738, | |
| "learning_rate": 9.811886953863841e-06, | |
| "loss": 0.091, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.1823819077147547, | |
| "grad_norm": 0.8174653649330139, | |
| "learning_rate": 9.795993966045418e-06, | |
| "loss": 0.0868, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.18967718402334488, | |
| "grad_norm": 1.086427092552185, | |
| "learning_rate": 9.779470496520442e-06, | |
| "loss": 0.0882, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.19697246033193508, | |
| "grad_norm": 0.8431033492088318, | |
| "learning_rate": 9.76231871746539e-06, | |
| "loss": 0.0819, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.20426773664052525, | |
| "grad_norm": 0.9793428778648376, | |
| "learning_rate": 9.744540883654348e-06, | |
| "loss": 0.0838, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.21156301294911545, | |
| "grad_norm": 1.1424306631088257, | |
| "learning_rate": 9.726139332162613e-06, | |
| "loss": 0.0923, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.21885828925770565, | |
| "grad_norm": 0.8779186010360718, | |
| "learning_rate": 9.707116482059447e-06, | |
| "loss": 0.0909, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.22615356556629582, | |
| "grad_norm": 0.862575113773346, | |
| "learning_rate": 9.68747483409007e-06, | |
| "loss": 0.094, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.23344884187488602, | |
| "grad_norm": 0.8033697009086609, | |
| "learning_rate": 9.667216970346916e-06, | |
| "loss": 0.0821, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.24074411818347619, | |
| "grad_norm": 1.136168360710144, | |
| "learning_rate": 9.646345553930187e-06, | |
| "loss": 0.084, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.24803939449206638, | |
| "grad_norm": 1.227287769317627, | |
| "learning_rate": 9.624863328597767e-06, | |
| "loss": 0.0867, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.2553346708006566, | |
| "grad_norm": 0.8906837105751038, | |
| "learning_rate": 9.602773118404518e-06, | |
| "loss": 0.0845, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.26262994710924675, | |
| "grad_norm": 0.9557612538337708, | |
| "learning_rate": 9.580077827331038e-06, | |
| "loss": 0.0896, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.2699252234178369, | |
| "grad_norm": 0.8804728984832764, | |
| "learning_rate": 9.556780438901899e-06, | |
| "loss": 0.0799, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.27722049972642715, | |
| "grad_norm": 0.8825680017471313, | |
| "learning_rate": 9.532884015793432e-06, | |
| "loss": 0.0879, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.2845157760350173, | |
| "grad_norm": 1.216217041015625, | |
| "learning_rate": 9.508391699431114e-06, | |
| "loss": 0.0878, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.2918110523436075, | |
| "grad_norm": 1.0918773412704468, | |
| "learning_rate": 9.48330670957659e-06, | |
| "loss": 0.0842, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.2991063286521977, | |
| "grad_norm": 0.8797096014022827, | |
| "learning_rate": 9.457632343904404e-06, | |
| "loss": 0.075, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.3064016049607879, | |
| "grad_norm": 0.9638001322746277, | |
| "learning_rate": 9.431371977568483e-06, | |
| "loss": 0.0834, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.31369688126937806, | |
| "grad_norm": 0.7447642683982849, | |
| "learning_rate": 9.404529062758447e-06, | |
| "loss": 0.0909, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.3209921575779683, | |
| "grad_norm": 1.179291844367981, | |
| "learning_rate": 9.377107128245782e-06, | |
| "loss": 0.0832, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.32828743388655846, | |
| "grad_norm": 0.6798911094665527, | |
| "learning_rate": 9.349109778919938e-06, | |
| "loss": 0.0842, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.3355827101951486, | |
| "grad_norm": 0.9468401670455933, | |
| "learning_rate": 9.32054069531444e-06, | |
| "loss": 0.0937, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.34287798650373885, | |
| "grad_norm": 0.9600223898887634, | |
| "learning_rate": 9.291403633123046e-06, | |
| "loss": 0.0909, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.350173262812329, | |
| "grad_norm": 0.7957492470741272, | |
| "learning_rate": 9.261702422706014e-06, | |
| "loss": 0.0924, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.3574685391209192, | |
| "grad_norm": 0.9197902083396912, | |
| "learning_rate": 9.231440968586572e-06, | |
| "loss": 0.0852, | |
| "step": 490 | |
| }, | |
| { | |
| "execution_accuracy": 37.0, | |
| "selected_difficulty": "challenging", | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.3647638154295094, | |
| "grad_norm": 1.0779783725738525, | |
| "learning_rate": 9.200623248937619e-06, | |
| "loss": 0.0872, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.3720590917380996, | |
| "grad_norm": 1.0330190658569336, | |
| "learning_rate": 9.169253315058764e-06, | |
| "loss": 0.0884, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.37935436804668976, | |
| "grad_norm": 0.8854564428329468, | |
| "learning_rate": 9.13733529084374e-06, | |
| "loss": 0.0888, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.38664964435527993, | |
| "grad_norm": 0.7928789258003235, | |
| "learning_rate": 9.104873372238269e-06, | |
| "loss": 0.0858, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.39394492066387016, | |
| "grad_norm": 1.1074901819229126, | |
| "learning_rate": 9.071871826688472e-06, | |
| "loss": 0.0866, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.40124019697246033, | |
| "grad_norm": 0.9030365347862244, | |
| "learning_rate": 9.038334992579863e-06, | |
| "loss": 0.084, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.4085354732810505, | |
| "grad_norm": 0.9121033549308777, | |
| "learning_rate": 9.004267278667032e-06, | |
| "loss": 0.0841, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.4158307495896407, | |
| "grad_norm": 0.7888039350509644, | |
| "learning_rate": 8.969673163494063e-06, | |
| "loss": 0.0844, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.4231260258982309, | |
| "grad_norm": 0.8982441425323486, | |
| "learning_rate": 8.934557194805787e-06, | |
| "loss": 0.0881, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.43042130220682107, | |
| "grad_norm": 0.9907477498054504, | |
| "learning_rate": 8.898923988949936e-06, | |
| "loss": 0.0829, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.4377165785154113, | |
| "grad_norm": 0.7510169744491577, | |
| "learning_rate": 8.862778230270276e-06, | |
| "loss": 0.0812, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.44501185482400146, | |
| "grad_norm": 0.8274700045585632, | |
| "learning_rate": 8.826124670490804e-06, | |
| "loss": 0.0872, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.45230713113259163, | |
| "grad_norm": 1.118033766746521, | |
| "learning_rate": 8.788968128091084e-06, | |
| "loss": 0.0934, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.45960240744118186, | |
| "grad_norm": 0.7703680992126465, | |
| "learning_rate": 8.751313487672815e-06, | |
| "loss": 0.0891, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.46689768374977203, | |
| "grad_norm": 1.0380890369415283, | |
| "learning_rate": 8.71316569931769e-06, | |
| "loss": 0.0931, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.4741929600583622, | |
| "grad_norm": 0.9548109769821167, | |
| "learning_rate": 8.674529777936674e-06, | |
| "loss": 0.0816, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.48148823636695237, | |
| "grad_norm": 1.2412773370742798, | |
| "learning_rate": 8.635410802610724e-06, | |
| "loss": 0.0872, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.4887835126755426, | |
| "grad_norm": 0.8023186922073364, | |
| "learning_rate": 8.595813915923113e-06, | |
| "loss": 0.0857, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.49607878898413277, | |
| "grad_norm": 0.829302966594696, | |
| "learning_rate": 8.555744323283364e-06, | |
| "loss": 0.0932, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.503374065292723, | |
| "grad_norm": 1.0218778848648071, | |
| "learning_rate": 8.515207292242969e-06, | |
| "loss": 0.0791, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.5106693416013132, | |
| "grad_norm": 0.9110598564147949, | |
| "learning_rate": 8.474208151802898e-06, | |
| "loss": 0.0917, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.5179646179099033, | |
| "grad_norm": 0.8709658980369568, | |
| "learning_rate": 8.432752291713058e-06, | |
| "loss": 0.0805, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.5252598942184935, | |
| "grad_norm": 0.9375218749046326, | |
| "learning_rate": 8.390845161763756e-06, | |
| "loss": 0.0887, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.5325551705270837, | |
| "grad_norm": 0.819020688533783, | |
| "learning_rate": 8.34849227106926e-06, | |
| "loss": 0.0823, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.5398504468356738, | |
| "grad_norm": 0.7377147078514099, | |
| "learning_rate": 8.305699187343586e-06, | |
| "loss": 0.0867, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.5471457231442641, | |
| "grad_norm": 0.9633333086967468, | |
| "learning_rate": 8.262471536168547e-06, | |
| "loss": 0.0893, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.5544409994528543, | |
| "grad_norm": 0.894578218460083, | |
| "learning_rate": 8.218815000254233e-06, | |
| "loss": 0.0874, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.5617362757614445, | |
| "grad_norm": 0.9977262616157532, | |
| "learning_rate": 8.174735318691946e-06, | |
| "loss": 0.0822, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.5690315520700346, | |
| "grad_norm": 0.8501657247543335, | |
| "learning_rate": 8.130238286199747e-06, | |
| "loss": 0.0874, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.5763268283786248, | |
| "grad_norm": 0.7603849172592163, | |
| "learning_rate": 8.085329752360683e-06, | |
| "loss": 0.0784, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.583622104687215, | |
| "grad_norm": 0.8487511277198792, | |
| "learning_rate": 8.04001562085379e-06, | |
| "loss": 0.0878, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.5909173809958053, | |
| "grad_norm": 0.9253877401351929, | |
| "learning_rate": 7.994301848678006e-06, | |
| "loss": 0.0815, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.5982126573043954, | |
| "grad_norm": 0.9946874976158142, | |
| "learning_rate": 7.948194445369065e-06, | |
| "loss": 0.0901, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.6055079336129856, | |
| "grad_norm": 1.006040334701538, | |
| "learning_rate": 7.901699472209467e-06, | |
| "loss": 0.0792, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.6128032099215758, | |
| "grad_norm": 0.8797623515129089, | |
| "learning_rate": 7.85482304143168e-06, | |
| "loss": 0.0816, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.620098486230166, | |
| "grad_norm": 0.9626962542533875, | |
| "learning_rate": 7.807571315414616e-06, | |
| "loss": 0.08, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.6273937625387561, | |
| "grad_norm": 1.075498104095459, | |
| "learning_rate": 7.759950505873523e-06, | |
| "loss": 0.0802, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.6346890388473464, | |
| "grad_norm": 0.90218585729599, | |
| "learning_rate": 7.711966873043396e-06, | |
| "loss": 0.0824, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.6419843151559366, | |
| "grad_norm": 0.8692035675048828, | |
| "learning_rate": 7.66362672485601e-06, | |
| "loss": 0.0769, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.6492795914645267, | |
| "grad_norm": 1.062687635421753, | |
| "learning_rate": 7.614936416110668e-06, | |
| "loss": 0.0834, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.6565748677731169, | |
| "grad_norm": 0.9357954859733582, | |
| "learning_rate": 7.565902347638806e-06, | |
| "loss": 0.0833, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.6638701440817071, | |
| "grad_norm": 0.873466432094574, | |
| "learning_rate": 7.5165309654625405e-06, | |
| "loss": 0.0803, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.6711654203902973, | |
| "grad_norm": 0.8454645276069641, | |
| "learning_rate": 7.466828759947271e-06, | |
| "loss": 0.0793, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.6784606966988874, | |
| "grad_norm": 1.1247007846832275, | |
| "learning_rate": 7.416802264948455e-06, | |
| "loss": 0.0757, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.6857559730074777, | |
| "grad_norm": 0.9007195234298706, | |
| "learning_rate": 7.366458056952668e-06, | |
| "loss": 0.0855, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.6930512493160679, | |
| "grad_norm": 0.775996208190918, | |
| "learning_rate": 7.315802754213062e-06, | |
| "loss": 0.079, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.700346525624658, | |
| "grad_norm": 1.1148663759231567, | |
| "learning_rate": 7.264843015879321e-06, | |
| "loss": 0.0881, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.7076418019332482, | |
| "grad_norm": 0.8462682366371155, | |
| "learning_rate": 7.213585541122261e-06, | |
| "loss": 0.0848, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.7149370782418384, | |
| "grad_norm": 0.8557429313659668, | |
| "learning_rate": 7.162037068253141e-06, | |
| "loss": 0.0839, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.7222323545504286, | |
| "grad_norm": 0.8856000900268555, | |
| "learning_rate": 7.110204373837857e-06, | |
| "loss": 0.0804, | |
| "step": 990 | |
| }, | |
| { | |
| "execution_accuracy": 47.0, | |
| "selected_difficulty": "challenging", | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.7295276308590188, | |
| "grad_norm": 0.7584331631660461, | |
| "learning_rate": 7.058094271806091e-06, | |
| "loss": 0.0818, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.736822907167609, | |
| "grad_norm": 0.9286295175552368, | |
| "learning_rate": 7.0057136125555456e-06, | |
| "loss": 0.0766, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.7441181834761992, | |
| "grad_norm": 1.1081056594848633, | |
| "learning_rate": 6.953069282051397e-06, | |
| "loss": 0.0835, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.7514134597847894, | |
| "grad_norm": 0.9020804762840271, | |
| "learning_rate": 6.900168200921065e-06, | |
| "loss": 0.0791, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.7587087360933795, | |
| "grad_norm": 1.0643606185913086, | |
| "learning_rate": 6.84701732354442e-06, | |
| "loss": 0.0866, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.7660040124019697, | |
| "grad_norm": 0.9293026924133301, | |
| "learning_rate": 6.79362363713957e-06, | |
| "loss": 0.0828, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.7732992887105599, | |
| "grad_norm": 0.9997085332870483, | |
| "learning_rate": 6.7399941608443096e-06, | |
| "loss": 0.0737, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.7805945650191501, | |
| "grad_norm": 0.8557327389717102, | |
| "learning_rate": 6.686135944793395e-06, | |
| "loss": 0.0836, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.7878898413277403, | |
| "grad_norm": 0.8752471208572388, | |
| "learning_rate": 6.632056069191723e-06, | |
| "loss": 0.0897, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.7951851176363305, | |
| "grad_norm": 0.9777392148971558, | |
| "learning_rate": 6.57776164338357e-06, | |
| "loss": 0.0805, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.8024803939449207, | |
| "grad_norm": 0.9013363122940063, | |
| "learning_rate": 6.523259804918001e-06, | |
| "loss": 0.0764, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.8097756702535108, | |
| "grad_norm": 0.930316686630249, | |
| "learning_rate": 6.4685577186105595e-06, | |
| "loss": 0.0858, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.817070946562101, | |
| "grad_norm": 1.2016055583953857, | |
| "learning_rate": 6.413662575601391e-06, | |
| "loss": 0.0809, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.8243662228706913, | |
| "grad_norm": 0.8230682611465454, | |
| "learning_rate": 6.358581592409881e-06, | |
| "loss": 0.0771, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.8316614991792814, | |
| "grad_norm": 0.9575796127319336, | |
| "learning_rate": 6.303322009985984e-06, | |
| "loss": 0.0893, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.8389567754878716, | |
| "grad_norm": 0.7613864541053772, | |
| "learning_rate": 6.247891092758319e-06, | |
| "loss": 0.0802, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.8462520517964618, | |
| "grad_norm": 0.8168739080429077, | |
| "learning_rate": 6.1922961276791925e-06, | |
| "loss": 0.0718, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.853547328105052, | |
| "grad_norm": 0.719241201877594, | |
| "learning_rate": 6.136544423266651e-06, | |
| "loss": 0.073, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.8608426044136421, | |
| "grad_norm": 1.0418319702148438, | |
| "learning_rate": 6.08064330864371e-06, | |
| "loss": 0.0816, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.8681378807222323, | |
| "grad_norm": 0.7831118106842041, | |
| "learning_rate": 6.024600132574855e-06, | |
| "loss": 0.0775, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.8754331570308226, | |
| "grad_norm": 0.9321058988571167, | |
| "learning_rate": 5.968422262499983e-06, | |
| "loss": 0.0777, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.8827284333394128, | |
| "grad_norm": 0.9959325194358826, | |
| "learning_rate": 5.912117083565874e-06, | |
| "loss": 0.0743, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.8900237096480029, | |
| "grad_norm": 0.7858604192733765, | |
| "learning_rate": 5.85569199765534e-06, | |
| "loss": 0.0742, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.8973189859565931, | |
| "grad_norm": 1.0771974325180054, | |
| "learning_rate": 5.799154422414174e-06, | |
| "loss": 0.0784, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.9046142622651833, | |
| "grad_norm": 1.0624542236328125, | |
| "learning_rate": 5.7425117902760195e-06, | |
| "loss": 0.0784, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.9119095385737734, | |
| "grad_norm": 0.7207911014556885, | |
| "learning_rate": 5.685771547485312e-06, | |
| "loss": 0.0694, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.9192048148823637, | |
| "grad_norm": 1.0225133895874023, | |
| "learning_rate": 5.628941153118388e-06, | |
| "loss": 0.0747, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.9265000911909539, | |
| "grad_norm": 1.0003268718719482, | |
| "learning_rate": 5.572028078102917e-06, | |
| "loss": 0.077, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.9337953674995441, | |
| "grad_norm": 0.9205290079116821, | |
| "learning_rate": 5.515039804235772e-06, | |
| "loss": 0.0764, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.9410906438081342, | |
| "grad_norm": 1.0423191785812378, | |
| "learning_rate": 5.457983823199475e-06, | |
| "loss": 0.0801, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.9483859201167244, | |
| "grad_norm": 0.8814120888710022, | |
| "learning_rate": 5.400867635577335e-06, | |
| "loss": 0.0801, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.9556811964253146, | |
| "grad_norm": 0.9548910856246948, | |
| "learning_rate": 5.343698749867421e-06, | |
| "loss": 0.0802, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.9629764727339047, | |
| "grad_norm": 0.7712908983230591, | |
| "learning_rate": 5.2864846814955e-06, | |
| "loss": 0.077, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.970271749042495, | |
| "grad_norm": 0.9300876259803772, | |
| "learning_rate": 5.229232951827054e-06, | |
| "loss": 0.081, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.9775670253510852, | |
| "grad_norm": 0.932421863079071, | |
| "learning_rate": 5.17195108717852e-06, | |
| "loss": 0.0841, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.9848623016596754, | |
| "grad_norm": 0.9440054297447205, | |
| "learning_rate": 5.114646617827884e-06, | |
| "loss": 0.0714, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.9921575779682655, | |
| "grad_norm": 0.8986610174179077, | |
| "learning_rate": 5.057327077024745e-06, | |
| "loss": 0.0781, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.9994528542768557, | |
| "grad_norm": 0.9697067141532898, | |
| "learning_rate": 5e-06, | |
| "loss": 0.0863, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.006748130585446, | |
| "grad_norm": 0.5880870223045349, | |
| "learning_rate": 4.942672922975255e-06, | |
| "loss": 0.0443, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.014043406894036, | |
| "grad_norm": 0.8269129395484924, | |
| "learning_rate": 4.8853533821721175e-06, | |
| "loss": 0.0464, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.0213386832026263, | |
| "grad_norm": 0.9354344606399536, | |
| "learning_rate": 4.82804891282148e-06, | |
| "loss": 0.0436, | |
| "step": 1400 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2740, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 200, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.849931290145456e+18, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |