| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.0, |
| "eval_steps": 500, |
| "global_step": 2120, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.009433962264150943, |
| "grad_norm": 2.1125893592834473, |
| "learning_rate": 1.9915094339622643e-05, |
| "loss": 0.6661, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.018867924528301886, |
| "grad_norm": 4.568980693817139, |
| "learning_rate": 1.9820754716981134e-05, |
| "loss": 0.6476, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.02830188679245283, |
| "grad_norm": 4.55064582824707, |
| "learning_rate": 1.9726415094339625e-05, |
| "loss": 0.609, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.03773584905660377, |
| "grad_norm": 2.458791732788086, |
| "learning_rate": 1.9632075471698116e-05, |
| "loss": 0.5465, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.04716981132075472, |
| "grad_norm": 5.235497951507568, |
| "learning_rate": 1.9537735849056604e-05, |
| "loss": 0.4277, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.05660377358490566, |
| "grad_norm": 2.81006121635437, |
| "learning_rate": 1.9443396226415095e-05, |
| "loss": 0.3367, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.0660377358490566, |
| "grad_norm": 5.580306529998779, |
| "learning_rate": 1.9349056603773586e-05, |
| "loss": 0.232, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.07547169811320754, |
| "grad_norm": 6.972874641418457, |
| "learning_rate": 1.9254716981132077e-05, |
| "loss": 0.1852, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.08490566037735849, |
| "grad_norm": 2.1650376319885254, |
| "learning_rate": 1.9160377358490568e-05, |
| "loss": 0.1581, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.09433962264150944, |
| "grad_norm": 0.42251577973365784, |
| "learning_rate": 1.906603773584906e-05, |
| "loss": 0.1821, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.10377358490566038, |
| "grad_norm": 0.361060231924057, |
| "learning_rate": 1.8971698113207547e-05, |
| "loss": 0.3396, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.11320754716981132, |
| "grad_norm": 1.1767176389694214, |
| "learning_rate": 1.887735849056604e-05, |
| "loss": 0.0724, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.12264150943396226, |
| "grad_norm": 0.310159295797348, |
| "learning_rate": 1.878301886792453e-05, |
| "loss": 0.1826, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.1320754716981132, |
| "grad_norm": 0.29479867219924927, |
| "learning_rate": 1.868867924528302e-05, |
| "loss": 0.1407, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.14150943396226415, |
| "grad_norm": 0.25081831216812134, |
| "learning_rate": 1.859433962264151e-05, |
| "loss": 0.0804, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.1509433962264151, |
| "grad_norm": 7.194790840148926, |
| "learning_rate": 1.8500000000000002e-05, |
| "loss": 0.1394, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.16037735849056603, |
| "grad_norm": 0.19942767918109894, |
| "learning_rate": 1.8405660377358494e-05, |
| "loss": 0.0876, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.16981132075471697, |
| "grad_norm": 12.822174072265625, |
| "learning_rate": 1.8311320754716985e-05, |
| "loss": 0.1606, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.1792452830188679, |
| "grad_norm": 1.3328683376312256, |
| "learning_rate": 1.8216981132075472e-05, |
| "loss": 0.1835, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.18867924528301888, |
| "grad_norm": 0.17138229310512543, |
| "learning_rate": 1.8122641509433963e-05, |
| "loss": 0.2355, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.19811320754716982, |
| "grad_norm": 0.21127229928970337, |
| "learning_rate": 1.8028301886792454e-05, |
| "loss": 0.2031, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.20754716981132076, |
| "grad_norm": 2.211583137512207, |
| "learning_rate": 1.7933962264150946e-05, |
| "loss": 0.0223, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.2169811320754717, |
| "grad_norm": 0.17399153113365173, |
| "learning_rate": 1.7839622641509437e-05, |
| "loss": 0.2204, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.22641509433962265, |
| "grad_norm": 0.26820316910743713, |
| "learning_rate": 1.7745283018867928e-05, |
| "loss": 0.1022, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.2358490566037736, |
| "grad_norm": 0.15803369879722595, |
| "learning_rate": 1.7650943396226415e-05, |
| "loss": 0.0638, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.24528301886792453, |
| "grad_norm": 0.15400823950767517, |
| "learning_rate": 1.7556603773584907e-05, |
| "loss": 0.1257, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.25471698113207547, |
| "grad_norm": 0.7688778638839722, |
| "learning_rate": 1.7462264150943398e-05, |
| "loss": 0.2443, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.2641509433962264, |
| "grad_norm": 0.243111252784729, |
| "learning_rate": 1.736792452830189e-05, |
| "loss": 0.2897, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.27358490566037735, |
| "grad_norm": 5.870262145996094, |
| "learning_rate": 1.727358490566038e-05, |
| "loss": 0.2198, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.2830188679245283, |
| "grad_norm": 0.2840751111507416, |
| "learning_rate": 1.7179245283018867e-05, |
| "loss": 0.3108, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.29245283018867924, |
| "grad_norm": 0.1476099044084549, |
| "learning_rate": 1.708490566037736e-05, |
| "loss": 0.1688, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.3018867924528302, |
| "grad_norm": 0.3752176761627197, |
| "learning_rate": 1.699056603773585e-05, |
| "loss": 0.2876, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.3113207547169811, |
| "grad_norm": 30.54532241821289, |
| "learning_rate": 1.689622641509434e-05, |
| "loss": 0.1215, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.32075471698113206, |
| "grad_norm": 0.15333189070224762, |
| "learning_rate": 1.6801886792452832e-05, |
| "loss": 0.2964, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.330188679245283, |
| "grad_norm": 0.23220475018024445, |
| "learning_rate": 1.6707547169811323e-05, |
| "loss": 0.1506, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.33962264150943394, |
| "grad_norm": 121.28533172607422, |
| "learning_rate": 1.661320754716981e-05, |
| "loss": 0.2107, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.3490566037735849, |
| "grad_norm": 1.4383134841918945, |
| "learning_rate": 1.6518867924528305e-05, |
| "loss": 0.2117, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.3584905660377358, |
| "grad_norm": 0.14638499915599823, |
| "learning_rate": 1.6424528301886793e-05, |
| "loss": 0.1013, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.36792452830188677, |
| "grad_norm": 0.124303437769413, |
| "learning_rate": 1.6330188679245284e-05, |
| "loss": 0.0478, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.37735849056603776, |
| "grad_norm": 0.11256200820207596, |
| "learning_rate": 1.6235849056603775e-05, |
| "loss": 0.0689, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.3867924528301887, |
| "grad_norm": 0.10450290888547897, |
| "learning_rate": 1.6141509433962266e-05, |
| "loss": 0.1293, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.39622641509433965, |
| "grad_norm": 0.09205685555934906, |
| "learning_rate": 1.6047169811320754e-05, |
| "loss": 0.0672, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.4056603773584906, |
| "grad_norm": 1.0851916074752808, |
| "learning_rate": 1.5952830188679248e-05, |
| "loss": 0.1918, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.41509433962264153, |
| "grad_norm": 0.1460585743188858, |
| "learning_rate": 1.5858490566037736e-05, |
| "loss": 0.0866, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.42452830188679247, |
| "grad_norm": 0.10592811554670334, |
| "learning_rate": 1.5764150943396227e-05, |
| "loss": 0.0327, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.4339622641509434, |
| "grad_norm": 0.09917836636304855, |
| "learning_rate": 1.5669811320754718e-05, |
| "loss": 0.268, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.44339622641509435, |
| "grad_norm": 0.2683676481246948, |
| "learning_rate": 1.557547169811321e-05, |
| "loss": 0.2433, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.4528301886792453, |
| "grad_norm": 0.3096301853656769, |
| "learning_rate": 1.54811320754717e-05, |
| "loss": 0.0106, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.46226415094339623, |
| "grad_norm": 0.14542903006076813, |
| "learning_rate": 1.538679245283019e-05, |
| "loss": 0.1084, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.4716981132075472, |
| "grad_norm": 0.09798438102006912, |
| "learning_rate": 1.529245283018868e-05, |
| "loss": 0.1217, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.4811320754716981, |
| "grad_norm": 0.5679267644882202, |
| "learning_rate": 1.5198113207547172e-05, |
| "loss": 0.1084, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.49056603773584906, |
| "grad_norm": 0.07084541022777557, |
| "learning_rate": 1.5103773584905661e-05, |
| "loss": 0.1496, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.09527801722288132, |
| "learning_rate": 1.5009433962264152e-05, |
| "loss": 0.0397, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.5094339622641509, |
| "grad_norm": 2.7925026416778564, |
| "learning_rate": 1.4915094339622642e-05, |
| "loss": 0.0082, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.5188679245283019, |
| "grad_norm": 61.901344299316406, |
| "learning_rate": 1.4820754716981134e-05, |
| "loss": 0.2448, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.5283018867924528, |
| "grad_norm": 0.10737080127000809, |
| "learning_rate": 1.4726415094339624e-05, |
| "loss": 0.2122, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.5377358490566038, |
| "grad_norm": 0.08159242570400238, |
| "learning_rate": 1.4632075471698115e-05, |
| "loss": 0.2056, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.5471698113207547, |
| "grad_norm": 26.218284606933594, |
| "learning_rate": 1.4537735849056604e-05, |
| "loss": 0.2137, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.5566037735849056, |
| "grad_norm": 0.09104019403457642, |
| "learning_rate": 1.4443396226415095e-05, |
| "loss": 0.1235, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.5660377358490566, |
| "grad_norm": 44.143577575683594, |
| "learning_rate": 1.4349056603773586e-05, |
| "loss": 0.1708, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.5754716981132075, |
| "grad_norm": 0.09825887531042099, |
| "learning_rate": 1.4254716981132078e-05, |
| "loss": 0.2829, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.5849056603773585, |
| "grad_norm": 0.08215257525444031, |
| "learning_rate": 1.4160377358490567e-05, |
| "loss": 0.0978, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.5943396226415094, |
| "grad_norm": 0.0858820378780365, |
| "learning_rate": 1.4066037735849058e-05, |
| "loss": 0.1308, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.6037735849056604, |
| "grad_norm": 0.07820013910531998, |
| "learning_rate": 1.3971698113207547e-05, |
| "loss": 0.0083, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.6132075471698113, |
| "grad_norm": 0.07088133692741394, |
| "learning_rate": 1.387735849056604e-05, |
| "loss": 0.0343, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.6226415094339622, |
| "grad_norm": 0.07562941312789917, |
| "learning_rate": 1.378301886792453e-05, |
| "loss": 0.1503, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.6320754716981132, |
| "grad_norm": 0.10470731556415558, |
| "learning_rate": 1.368867924528302e-05, |
| "loss": 0.0474, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.6415094339622641, |
| "grad_norm": 0.06047428399324417, |
| "learning_rate": 1.359433962264151e-05, |
| "loss": 0.3286, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.6509433962264151, |
| "grad_norm": 0.12241201102733612, |
| "learning_rate": 1.3500000000000001e-05, |
| "loss": 0.163, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.660377358490566, |
| "grad_norm": 0.07547236233949661, |
| "learning_rate": 1.3405660377358492e-05, |
| "loss": 0.1182, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.6698113207547169, |
| "grad_norm": 0.06518665701150894, |
| "learning_rate": 1.3311320754716983e-05, |
| "loss": 0.1177, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.6792452830188679, |
| "grad_norm": 0.05459102615714073, |
| "learning_rate": 1.3216981132075473e-05, |
| "loss": 0.0056, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.6886792452830188, |
| "grad_norm": 0.06507308781147003, |
| "learning_rate": 1.3122641509433964e-05, |
| "loss": 0.194, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.6981132075471698, |
| "grad_norm": 0.06847833096981049, |
| "learning_rate": 1.3028301886792453e-05, |
| "loss": 0.0334, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.7075471698113207, |
| "grad_norm": 0.05916672945022583, |
| "learning_rate": 1.2933962264150946e-05, |
| "loss": 0.1326, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.7169811320754716, |
| "grad_norm": 0.05806839466094971, |
| "learning_rate": 1.2839622641509435e-05, |
| "loss": 0.0734, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.7264150943396226, |
| "grad_norm": 0.058394916355609894, |
| "learning_rate": 1.2745283018867926e-05, |
| "loss": 0.3154, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.7358490566037735, |
| "grad_norm": 0.08128712326288223, |
| "learning_rate": 1.2650943396226416e-05, |
| "loss": 0.1004, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.7452830188679245, |
| "grad_norm": 0.1772662252187729, |
| "learning_rate": 1.2556603773584907e-05, |
| "loss": 0.1408, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.7547169811320755, |
| "grad_norm": 0.07657451182603836, |
| "learning_rate": 1.2462264150943398e-05, |
| "loss": 0.1138, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.7641509433962265, |
| "grad_norm": 0.0920625552535057, |
| "learning_rate": 1.2367924528301887e-05, |
| "loss": 0.0064, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.7735849056603774, |
| "grad_norm": 3.893165111541748, |
| "learning_rate": 1.2273584905660379e-05, |
| "loss": 0.1015, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.7830188679245284, |
| "grad_norm": 0.06635797768831253, |
| "learning_rate": 1.2179245283018868e-05, |
| "loss": 0.1405, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.7924528301886793, |
| "grad_norm": 1.1072912216186523, |
| "learning_rate": 1.2084905660377359e-05, |
| "loss": 0.1225, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.8018867924528302, |
| "grad_norm": 0.04783385619521141, |
| "learning_rate": 1.1990566037735848e-05, |
| "loss": 0.0895, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.8113207547169812, |
| "grad_norm": 0.05416665971279144, |
| "learning_rate": 1.1896226415094341e-05, |
| "loss": 0.1095, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.8207547169811321, |
| "grad_norm": 0.10190918296575546, |
| "learning_rate": 1.180188679245283e-05, |
| "loss": 0.1404, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.8301886792452831, |
| "grad_norm": 92.75904083251953, |
| "learning_rate": 1.1707547169811322e-05, |
| "loss": 0.0269, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.839622641509434, |
| "grad_norm": 0.06948844343423843, |
| "learning_rate": 1.1613207547169811e-05, |
| "loss": 0.0046, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.8490566037735849, |
| "grad_norm": 0.0549422986805439, |
| "learning_rate": 1.1518867924528304e-05, |
| "loss": 0.1588, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.8584905660377359, |
| "grad_norm": 0.09215264767408371, |
| "learning_rate": 1.1424528301886793e-05, |
| "loss": 0.0639, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.8679245283018868, |
| "grad_norm": 0.0980132520198822, |
| "learning_rate": 1.1330188679245284e-05, |
| "loss": 0.21, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.8773584905660378, |
| "grad_norm": 0.06740464270114899, |
| "learning_rate": 1.1235849056603774e-05, |
| "loss": 0.0059, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.8867924528301887, |
| "grad_norm": 0.42118075489997864, |
| "learning_rate": 1.1141509433962265e-05, |
| "loss": 0.1012, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.8962264150943396, |
| "grad_norm": 0.09071939438581467, |
| "learning_rate": 1.1047169811320754e-05, |
| "loss": 0.0636, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.9056603773584906, |
| "grad_norm": 0.06605428457260132, |
| "learning_rate": 1.0952830188679247e-05, |
| "loss": 0.1424, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.9150943396226415, |
| "grad_norm": 0.054869458079338074, |
| "learning_rate": 1.0858490566037736e-05, |
| "loss": 0.1506, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.9245283018867925, |
| "grad_norm": 8.528816223144531, |
| "learning_rate": 1.0764150943396227e-05, |
| "loss": 0.0766, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.9339622641509434, |
| "grad_norm": 0.04189067333936691, |
| "learning_rate": 1.0669811320754717e-05, |
| "loss": 0.0383, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.9433962264150944, |
| "grad_norm": 0.19604367017745972, |
| "learning_rate": 1.057547169811321e-05, |
| "loss": 0.0474, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.9528301886792453, |
| "grad_norm": 0.04203261435031891, |
| "learning_rate": 1.0481132075471699e-05, |
| "loss": 0.0571, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.9622641509433962, |
| "grad_norm": 20.811012268066406, |
| "learning_rate": 1.038679245283019e-05, |
| "loss": 0.2019, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.9716981132075472, |
| "grad_norm": 0.10877203941345215, |
| "learning_rate": 1.029245283018868e-05, |
| "loss": 0.1336, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.9811320754716981, |
| "grad_norm": 0.10228094458580017, |
| "learning_rate": 1.019811320754717e-05, |
| "loss": 0.0273, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.9905660377358491, |
| "grad_norm": 1.2278637886047363, |
| "learning_rate": 1.010377358490566e-05, |
| "loss": 0.0452, |
| "step": 1050 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.03946210443973541, |
| "learning_rate": 1.0009433962264153e-05, |
| "loss": 0.0109, |
| "step": 1060 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_loss": 0.08500511199235916, |
| "eval_runtime": 1.6971, |
| "eval_samples_per_second": 555.67, |
| "eval_steps_per_second": 69.532, |
| "step": 1060 |
| }, |
| { |
| "epoch": 1.009433962264151, |
| "grad_norm": 0.03779453784227371, |
| "learning_rate": 9.915094339622642e-06, |
| "loss": 0.1822, |
| "step": 1070 |
| }, |
| { |
| "epoch": 1.0188679245283019, |
| "grad_norm": 0.040553513914346695, |
| "learning_rate": 9.820754716981133e-06, |
| "loss": 0.054, |
| "step": 1080 |
| }, |
| { |
| "epoch": 1.028301886792453, |
| "grad_norm": 0.02908136695623398, |
| "learning_rate": 9.726415094339623e-06, |
| "loss": 0.0062, |
| "step": 1090 |
| }, |
| { |
| "epoch": 1.0377358490566038, |
| "grad_norm": 0.04287223890423775, |
| "learning_rate": 9.632075471698114e-06, |
| "loss": 0.0039, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.0471698113207548, |
| "grad_norm": 0.03865998610854149, |
| "learning_rate": 9.537735849056605e-06, |
| "loss": 0.0446, |
| "step": 1110 |
| }, |
| { |
| "epoch": 1.0566037735849056, |
| "grad_norm": 0.04521052539348602, |
| "learning_rate": 9.443396226415094e-06, |
| "loss": 0.0029, |
| "step": 1120 |
| }, |
| { |
| "epoch": 1.0660377358490567, |
| "grad_norm": 0.03594714403152466, |
| "learning_rate": 9.349056603773585e-06, |
| "loss": 0.148, |
| "step": 1130 |
| }, |
| { |
| "epoch": 1.0754716981132075, |
| "grad_norm": 0.029539430513978004, |
| "learning_rate": 9.254716981132076e-06, |
| "loss": 0.0271, |
| "step": 1140 |
| }, |
| { |
| "epoch": 1.0849056603773586, |
| "grad_norm": 4.038683891296387, |
| "learning_rate": 9.160377358490566e-06, |
| "loss": 0.165, |
| "step": 1150 |
| }, |
| { |
| "epoch": 1.0943396226415094, |
| "grad_norm": 0.027737464755773544, |
| "learning_rate": 9.066037735849057e-06, |
| "loss": 0.0028, |
| "step": 1160 |
| }, |
| { |
| "epoch": 1.1037735849056605, |
| "grad_norm": 0.03591258078813553, |
| "learning_rate": 8.971698113207548e-06, |
| "loss": 0.1371, |
| "step": 1170 |
| }, |
| { |
| "epoch": 1.1132075471698113, |
| "grad_norm": 12.058833122253418, |
| "learning_rate": 8.877358490566039e-06, |
| "loss": 0.236, |
| "step": 1180 |
| }, |
| { |
| "epoch": 1.1226415094339623, |
| "grad_norm": 0.04443644732236862, |
| "learning_rate": 8.783018867924528e-06, |
| "loss": 0.003, |
| "step": 1190 |
| }, |
| { |
| "epoch": 1.1320754716981132, |
| "grad_norm": 40.05979919433594, |
| "learning_rate": 8.68867924528302e-06, |
| "loss": 0.0157, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.1415094339622642, |
| "grad_norm": 0.04003468528389931, |
| "learning_rate": 8.59433962264151e-06, |
| "loss": 0.0029, |
| "step": 1210 |
| }, |
| { |
| "epoch": 1.150943396226415, |
| "grad_norm": 2.571934223175049, |
| "learning_rate": 8.5e-06, |
| "loss": 0.1204, |
| "step": 1220 |
| }, |
| { |
| "epoch": 1.1603773584905661, |
| "grad_norm": 0.06405159831047058, |
| "learning_rate": 8.405660377358491e-06, |
| "loss": 0.0146, |
| "step": 1230 |
| }, |
| { |
| "epoch": 1.169811320754717, |
| "grad_norm": 8.398702621459961, |
| "learning_rate": 8.311320754716982e-06, |
| "loss": 0.0048, |
| "step": 1240 |
| }, |
| { |
| "epoch": 1.179245283018868, |
| "grad_norm": 19.97818946838379, |
| "learning_rate": 8.216981132075471e-06, |
| "loss": 0.0688, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.1886792452830188, |
| "grad_norm": 0.03282918781042099, |
| "learning_rate": 8.122641509433963e-06, |
| "loss": 0.0035, |
| "step": 1260 |
| }, |
| { |
| "epoch": 1.1981132075471699, |
| "grad_norm": 0.03513512760400772, |
| "learning_rate": 8.028301886792454e-06, |
| "loss": 0.0355, |
| "step": 1270 |
| }, |
| { |
| "epoch": 1.2075471698113207, |
| "grad_norm": 0.03274456784129143, |
| "learning_rate": 7.933962264150945e-06, |
| "loss": 0.188, |
| "step": 1280 |
| }, |
| { |
| "epoch": 1.2169811320754718, |
| "grad_norm": 0.03626378998160362, |
| "learning_rate": 7.839622641509434e-06, |
| "loss": 0.014, |
| "step": 1290 |
| }, |
| { |
| "epoch": 1.2264150943396226, |
| "grad_norm": 0.8501767516136169, |
| "learning_rate": 7.745283018867925e-06, |
| "loss": 0.1076, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.2358490566037736, |
| "grad_norm": 0.0267633106559515, |
| "learning_rate": 7.650943396226416e-06, |
| "loss": 0.0025, |
| "step": 1310 |
| }, |
| { |
| "epoch": 1.2452830188679245, |
| "grad_norm": 0.04418088495731354, |
| "learning_rate": 7.5566037735849066e-06, |
| "loss": 0.1331, |
| "step": 1320 |
| }, |
| { |
| "epoch": 1.2547169811320755, |
| "grad_norm": 6.747396469116211, |
| "learning_rate": 7.462264150943397e-06, |
| "loss": 0.3259, |
| "step": 1330 |
| }, |
| { |
| "epoch": 1.2641509433962264, |
| "grad_norm": 0.032822128385305405, |
| "learning_rate": 7.367924528301887e-06, |
| "loss": 0.1229, |
| "step": 1340 |
| }, |
| { |
| "epoch": 1.2735849056603774, |
| "grad_norm": 0.06315261125564575, |
| "learning_rate": 7.273584905660378e-06, |
| "loss": 0.0039, |
| "step": 1350 |
| }, |
| { |
| "epoch": 1.2830188679245282, |
| "grad_norm": 0.03962039574980736, |
| "learning_rate": 7.179245283018868e-06, |
| "loss": 0.0072, |
| "step": 1360 |
| }, |
| { |
| "epoch": 1.2924528301886793, |
| "grad_norm": 0.03726816922426224, |
| "learning_rate": 7.0849056603773594e-06, |
| "loss": 0.0138, |
| "step": 1370 |
| }, |
| { |
| "epoch": 1.3018867924528301, |
| "grad_norm": 0.04036370664834976, |
| "learning_rate": 6.99056603773585e-06, |
| "loss": 0.1472, |
| "step": 1380 |
| }, |
| { |
| "epoch": 1.3113207547169812, |
| "grad_norm": 0.03037635050714016, |
| "learning_rate": 6.89622641509434e-06, |
| "loss": 0.0069, |
| "step": 1390 |
| }, |
| { |
| "epoch": 1.320754716981132, |
| "grad_norm": 0.02670077607035637, |
| "learning_rate": 6.801886792452831e-06, |
| "loss": 0.0021, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.330188679245283, |
| "grad_norm": 0.05846339091658592, |
| "learning_rate": 6.707547169811321e-06, |
| "loss": 0.0022, |
| "step": 1410 |
| }, |
| { |
| "epoch": 1.3396226415094339, |
| "grad_norm": 9.76332950592041, |
| "learning_rate": 6.613207547169812e-06, |
| "loss": 0.1445, |
| "step": 1420 |
| }, |
| { |
| "epoch": 1.349056603773585, |
| "grad_norm": 0.048967599868774414, |
| "learning_rate": 6.5188679245283026e-06, |
| "loss": 0.0441, |
| "step": 1430 |
| }, |
| { |
| "epoch": 1.3584905660377358, |
| "grad_norm": 0.025179484859108925, |
| "learning_rate": 6.424528301886793e-06, |
| "loss": 0.0188, |
| "step": 1440 |
| }, |
| { |
| "epoch": 1.3679245283018868, |
| "grad_norm": 0.02890361286699772, |
| "learning_rate": 6.330188679245284e-06, |
| "loss": 0.0021, |
| "step": 1450 |
| }, |
| { |
| "epoch": 1.3773584905660377, |
| "grad_norm": 0.02526584453880787, |
| "learning_rate": 6.235849056603774e-06, |
| "loss": 0.0022, |
| "step": 1460 |
| }, |
| { |
| "epoch": 1.3867924528301887, |
| "grad_norm": 2.650209903717041, |
| "learning_rate": 6.141509433962265e-06, |
| "loss": 0.0482, |
| "step": 1470 |
| }, |
| { |
| "epoch": 1.3962264150943398, |
| "grad_norm": 0.0368034765124321, |
| "learning_rate": 6.0471698113207555e-06, |
| "loss": 0.0034, |
| "step": 1480 |
| }, |
| { |
| "epoch": 1.4056603773584906, |
| "grad_norm": 0.021781697869300842, |
| "learning_rate": 5.952830188679246e-06, |
| "loss": 0.0022, |
| "step": 1490 |
| }, |
| { |
| "epoch": 1.4150943396226414, |
| "grad_norm": 0.019001835957169533, |
| "learning_rate": 5.858490566037737e-06, |
| "loss": 0.2058, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.4245283018867925, |
| "grad_norm": 0.047584012150764465, |
| "learning_rate": 5.764150943396227e-06, |
| "loss": 0.1632, |
| "step": 1510 |
| }, |
| { |
| "epoch": 1.4339622641509435, |
| "grad_norm": 0.02781420387327671, |
| "learning_rate": 5.669811320754718e-06, |
| "loss": 0.0018, |
| "step": 1520 |
| }, |
| { |
| "epoch": 1.4433962264150944, |
| "grad_norm": 0.03572264313697815, |
| "learning_rate": 5.575471698113208e-06, |
| "loss": 0.0184, |
| "step": 1530 |
| }, |
| { |
| "epoch": 1.4528301886792452, |
| "grad_norm": 0.021896662190556526, |
| "learning_rate": 5.4811320754716994e-06, |
| "loss": 0.155, |
| "step": 1540 |
| }, |
| { |
| "epoch": 1.4622641509433962, |
| "grad_norm": 0.024771416559815407, |
| "learning_rate": 5.38679245283019e-06, |
| "loss": 0.0018, |
| "step": 1550 |
| }, |
| { |
| "epoch": 1.4716981132075473, |
| "grad_norm": 3.849034547805786, |
| "learning_rate": 5.29245283018868e-06, |
| "loss": 0.1006, |
| "step": 1560 |
| }, |
| { |
| "epoch": 1.4811320754716981, |
| "grad_norm": 0.12136558443307877, |
| "learning_rate": 5.198113207547171e-06, |
| "loss": 0.0024, |
| "step": 1570 |
| }, |
| { |
| "epoch": 1.490566037735849, |
| "grad_norm": 0.029468955472111702, |
| "learning_rate": 5.103773584905661e-06, |
| "loss": 0.0021, |
| "step": 1580 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 2.4301178455352783, |
| "learning_rate": 5.009433962264152e-06, |
| "loss": 0.0024, |
| "step": 1590 |
| }, |
| { |
| "epoch": 1.509433962264151, |
| "grad_norm": 0.024639485403895378, |
| "learning_rate": 4.915094339622642e-06, |
| "loss": 0.0019, |
| "step": 1600 |
| }, |
| { |
| "epoch": 1.5188679245283019, |
| "grad_norm": 5.811135768890381, |
| "learning_rate": 4.820754716981133e-06, |
| "loss": 0.0282, |
| "step": 1610 |
| }, |
| { |
| "epoch": 1.5283018867924527, |
| "grad_norm": 0.02190409228205681, |
| "learning_rate": 4.726415094339623e-06, |
| "loss": 0.0016, |
| "step": 1620 |
| }, |
| { |
| "epoch": 1.5377358490566038, |
| "grad_norm": 0.025405917316675186, |
| "learning_rate": 4.632075471698113e-06, |
| "loss": 0.0505, |
| "step": 1630 |
| }, |
| { |
| "epoch": 1.5471698113207548, |
| "grad_norm": 16.779964447021484, |
| "learning_rate": 4.537735849056604e-06, |
| "loss": 0.1169, |
| "step": 1640 |
| }, |
| { |
| "epoch": 1.5566037735849056, |
| "grad_norm": 43.401329040527344, |
| "learning_rate": 4.443396226415095e-06, |
| "loss": 0.0245, |
| "step": 1650 |
| }, |
| { |
| "epoch": 1.5660377358490565, |
| "grad_norm": 0.9246163964271545, |
| "learning_rate": 4.349056603773586e-06, |
| "loss": 0.0354, |
| "step": 1660 |
| }, |
| { |
| "epoch": 1.5754716981132075, |
| "grad_norm": 0.02293049544095993, |
| "learning_rate": 4.254716981132076e-06, |
| "loss": 0.0036, |
| "step": 1670 |
| }, |
| { |
| "epoch": 1.5849056603773586, |
| "grad_norm": 0.019884012639522552, |
| "learning_rate": 4.160377358490566e-06, |
| "loss": 0.0017, |
| "step": 1680 |
| }, |
| { |
| "epoch": 1.5943396226415094, |
| "grad_norm": 0.0297381728887558, |
| "learning_rate": 4.066037735849057e-06, |
| "loss": 0.002, |
| "step": 1690 |
| }, |
| { |
| "epoch": 1.6037735849056602, |
| "grad_norm": 0.01664115861058235, |
| "learning_rate": 3.9716981132075475e-06, |
| "loss": 0.0029, |
| "step": 1700 |
| }, |
| { |
| "epoch": 1.6132075471698113, |
| "grad_norm": 0.021111104637384415, |
| "learning_rate": 3.8773584905660386e-06, |
| "loss": 0.0019, |
| "step": 1710 |
| }, |
| { |
| "epoch": 1.6226415094339623, |
| "grad_norm": 5.916421890258789, |
| "learning_rate": 3.783018867924529e-06, |
| "loss": 0.0597, |
| "step": 1720 |
| }, |
| { |
| "epoch": 1.6320754716981132, |
| "grad_norm": 0.11948911845684052, |
| "learning_rate": 3.688679245283019e-06, |
| "loss": 0.0288, |
| "step": 1730 |
| }, |
| { |
| "epoch": 1.641509433962264, |
| "grad_norm": 0.029013920575380325, |
| "learning_rate": 3.5943396226415093e-06, |
| "loss": 0.0022, |
| "step": 1740 |
| }, |
| { |
| "epoch": 1.650943396226415, |
| "grad_norm": 0.021010151132941246, |
| "learning_rate": 3.5e-06, |
| "loss": 0.0021, |
| "step": 1750 |
| }, |
| { |
| "epoch": 1.6603773584905661, |
| "grad_norm": 0.020002691075205803, |
| "learning_rate": 3.4056603773584906e-06, |
| "loss": 0.0016, |
| "step": 1760 |
| }, |
| { |
| "epoch": 1.669811320754717, |
| "grad_norm": 0.02407264895737171, |
| "learning_rate": 3.3113207547169813e-06, |
| "loss": 0.0928, |
| "step": 1770 |
| }, |
| { |
| "epoch": 1.6792452830188678, |
| "grad_norm": 0.024595679715275764, |
| "learning_rate": 3.216981132075472e-06, |
| "loss": 0.0018, |
| "step": 1780 |
| }, |
| { |
| "epoch": 1.6886792452830188, |
| "grad_norm": 0.017810633406043053, |
| "learning_rate": 3.122641509433962e-06, |
| "loss": 0.1722, |
| "step": 1790 |
| }, |
| { |
| "epoch": 1.6981132075471699, |
| "grad_norm": 0.01612844318151474, |
| "learning_rate": 3.028301886792453e-06, |
| "loss": 0.1158, |
| "step": 1800 |
| }, |
| { |
| "epoch": 1.7075471698113207, |
| "grad_norm": 0.023075569421052933, |
| "learning_rate": 2.9339622641509435e-06, |
| "loss": 0.0016, |
| "step": 1810 |
| }, |
| { |
| "epoch": 1.7169811320754715, |
| "grad_norm": 0.01716834492981434, |
| "learning_rate": 2.839622641509434e-06, |
| "loss": 0.0015, |
| "step": 1820 |
| }, |
| { |
| "epoch": 1.7264150943396226, |
| "grad_norm": 0.0205362718552351, |
| "learning_rate": 2.745283018867925e-06, |
| "loss": 0.0017, |
| "step": 1830 |
| }, |
| { |
| "epoch": 1.7358490566037736, |
| "grad_norm": 0.023218706250190735, |
| "learning_rate": 2.6509433962264155e-06, |
| "loss": 0.0014, |
| "step": 1840 |
| }, |
| { |
| "epoch": 1.7452830188679245, |
| "grad_norm": 0.018329111859202385, |
| "learning_rate": 2.5566037735849057e-06, |
| "loss": 0.1204, |
| "step": 1850 |
| }, |
| { |
| "epoch": 1.7547169811320755, |
| "grad_norm": 0.019258936867117882, |
| "learning_rate": 2.4622641509433964e-06, |
| "loss": 0.0018, |
| "step": 1860 |
| }, |
| { |
| "epoch": 1.7641509433962264, |
| "grad_norm": 0.023526819422841072, |
| "learning_rate": 2.367924528301887e-06, |
| "loss": 0.0074, |
| "step": 1870 |
| }, |
| { |
| "epoch": 1.7735849056603774, |
| "grad_norm": 0.01679695025086403, |
| "learning_rate": 2.2735849056603777e-06, |
| "loss": 0.0014, |
| "step": 1880 |
| }, |
| { |
| "epoch": 1.7830188679245285, |
| "grad_norm": 0.014311902225017548, |
| "learning_rate": 2.1792452830188684e-06, |
| "loss": 0.0015, |
| "step": 1890 |
| }, |
| { |
| "epoch": 1.7924528301886793, |
| "grad_norm": 1.5082017183303833, |
| "learning_rate": 2.0849056603773586e-06, |
| "loss": 0.0708, |
| "step": 1900 |
| }, |
| { |
| "epoch": 1.8018867924528301, |
| "grad_norm": 52.86256408691406, |
| "learning_rate": 1.9905660377358493e-06, |
| "loss": 0.0666, |
| "step": 1910 |
| }, |
| { |
| "epoch": 1.8113207547169812, |
| "grad_norm": 0.020178191363811493, |
| "learning_rate": 1.89622641509434e-06, |
| "loss": 0.0694, |
| "step": 1920 |
| }, |
| { |
| "epoch": 1.8207547169811322, |
| "grad_norm": 0.01885703019797802, |
| "learning_rate": 1.8018867924528302e-06, |
| "loss": 0.0416, |
| "step": 1930 |
| }, |
| { |
| "epoch": 1.830188679245283, |
| "grad_norm": 0.021212713792920113, |
| "learning_rate": 1.7075471698113208e-06, |
| "loss": 0.0015, |
| "step": 1940 |
| }, |
| { |
| "epoch": 1.8396226415094339, |
| "grad_norm": 0.020716039463877678, |
| "learning_rate": 1.6132075471698113e-06, |
| "loss": 0.0446, |
| "step": 1950 |
| }, |
| { |
| "epoch": 1.849056603773585, |
| "grad_norm": 0.01979793794453144, |
| "learning_rate": 1.518867924528302e-06, |
| "loss": 0.0017, |
| "step": 1960 |
| }, |
| { |
| "epoch": 1.858490566037736, |
| "grad_norm": 0.029191287234425545, |
| "learning_rate": 1.4245283018867926e-06, |
| "loss": 0.0016, |
| "step": 1970 |
| }, |
| { |
| "epoch": 1.8679245283018868, |
| "grad_norm": 0.02217436581850052, |
| "learning_rate": 1.330188679245283e-06, |
| "loss": 0.0441, |
| "step": 1980 |
| }, |
| { |
| "epoch": 1.8773584905660377, |
| "grad_norm": 0.0228409580886364, |
| "learning_rate": 1.2358490566037737e-06, |
| "loss": 0.0016, |
| "step": 1990 |
| }, |
| { |
| "epoch": 1.8867924528301887, |
| "grad_norm": 0.022314254194498062, |
| "learning_rate": 1.1415094339622642e-06, |
| "loss": 0.1457, |
| "step": 2000 |
| }, |
| { |
| "epoch": 1.8962264150943398, |
| "grad_norm": 0.017337264493107796, |
| "learning_rate": 1.0471698113207548e-06, |
| "loss": 0.1144, |
| "step": 2010 |
| }, |
| { |
| "epoch": 1.9056603773584906, |
| "grad_norm": 0.021478446200489998, |
| "learning_rate": 9.528301886792454e-07, |
| "loss": 0.0014, |
| "step": 2020 |
| }, |
| { |
| "epoch": 1.9150943396226414, |
| "grad_norm": 0.021643251180648804, |
| "learning_rate": 8.584905660377358e-07, |
| "loss": 0.0284, |
| "step": 2030 |
| }, |
| { |
| "epoch": 1.9245283018867925, |
| "grad_norm": 0.02016671560704708, |
| "learning_rate": 7.641509433962264e-07, |
| "loss": 0.0482, |
| "step": 2040 |
| }, |
| { |
| "epoch": 1.9339622641509435, |
| "grad_norm": 0.019274510443210602, |
| "learning_rate": 6.698113207547171e-07, |
| "loss": 0.0016, |
| "step": 2050 |
| }, |
| { |
| "epoch": 1.9433962264150944, |
| "grad_norm": 0.014301498420536518, |
| "learning_rate": 5.754716981132076e-07, |
| "loss": 0.0045, |
| "step": 2060 |
| }, |
| { |
| "epoch": 1.9528301886792452, |
| "grad_norm": 0.01988300494849682, |
| "learning_rate": 4.811320754716982e-07, |
| "loss": 0.0015, |
| "step": 2070 |
| }, |
| { |
| "epoch": 1.9622641509433962, |
| "grad_norm": 0.02060701698064804, |
| "learning_rate": 3.867924528301887e-07, |
| "loss": 0.0044, |
| "step": 2080 |
| }, |
| { |
| "epoch": 1.9716981132075473, |
| "grad_norm": 41.62403106689453, |
| "learning_rate": 2.924528301886793e-07, |
| "loss": 0.1818, |
| "step": 2090 |
| }, |
| { |
| "epoch": 1.9811320754716981, |
| "grad_norm": 0.0649556890130043, |
| "learning_rate": 1.9811320754716982e-07, |
| "loss": 0.0016, |
| "step": 2100 |
| }, |
| { |
| "epoch": 1.990566037735849, |
| "grad_norm": 0.02280914969742298, |
| "learning_rate": 1.0377358490566039e-07, |
| "loss": 0.0414, |
| "step": 2110 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.01854662224650383, |
| "learning_rate": 9.433962264150945e-09, |
| "loss": 0.0015, |
| "step": 2120 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_loss": 0.06546525657176971, |
| "eval_runtime": 1.7009, |
| "eval_samples_per_second": 554.403, |
| "eval_steps_per_second": 69.374, |
| "step": 2120 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 2120, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 4461837276794880.0, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|