| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 14.993576017130621, |
| "eval_steps": 500, |
| "global_step": 1740, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.08565310492505353, |
| "grad_norm": 5.411635875701904, |
| "learning_rate": 9.999339889379647e-06, |
| "loss": 1.5609, |
| "num_input_tokens_seen": 255328, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.17130620985010706, |
| "grad_norm": 0.7214002013206482, |
| "learning_rate": 9.997058249278764e-06, |
| "loss": 0.2055, |
| "num_input_tokens_seen": 510640, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.2569593147751606, |
| "grad_norm": 0.47501006722450256, |
| "learning_rate": 9.993147673772869e-06, |
| "loss": 0.184, |
| "num_input_tokens_seen": 763808, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.3426124197002141, |
| "grad_norm": 0.20641829073429108, |
| "learning_rate": 9.987609437626955e-06, |
| "loss": 0.1767, |
| "num_input_tokens_seen": 1017472, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.4282655246252677, |
| "grad_norm": 0.49966660141944885, |
| "learning_rate": 9.98044534618898e-06, |
| "loss": 0.1821, |
| "num_input_tokens_seen": 1273488, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.5139186295503212, |
| "grad_norm": 0.8994255065917969, |
| "learning_rate": 9.971657734801385e-06, |
| "loss": 0.1819, |
| "num_input_tokens_seen": 1527744, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.5995717344753747, |
| "grad_norm": 1.7688418626785278, |
| "learning_rate": 9.961249468039806e-06, |
| "loss": 0.1797, |
| "num_input_tokens_seen": 1785520, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.6852248394004282, |
| "grad_norm": 0.7211973071098328, |
| "learning_rate": 9.949223938779286e-06, |
| "loss": 0.1765, |
| "num_input_tokens_seen": 2037648, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.7708779443254818, |
| "grad_norm": 0.94338458776474, |
| "learning_rate": 9.935585067088276e-06, |
| "loss": 0.1766, |
| "num_input_tokens_seen": 2292464, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.8565310492505354, |
| "grad_norm": 0.6227909326553345, |
| "learning_rate": 9.920337298950767e-06, |
| "loss": 0.1714, |
| "num_input_tokens_seen": 2547872, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.9421841541755889, |
| "grad_norm": 0.5941164493560791, |
| "learning_rate": 9.903485604816993e-06, |
| "loss": 0.1728, |
| "num_input_tokens_seen": 2801536, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.0342612419700214, |
| "grad_norm": 0.4057783782482147, |
| "learning_rate": 9.885035477983184e-06, |
| "loss": 0.1868, |
| "num_input_tokens_seen": 3074416, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.119914346895075, |
| "grad_norm": 0.5549430251121521, |
| "learning_rate": 9.864992932800845e-06, |
| "loss": 0.1679, |
| "num_input_tokens_seen": 3327568, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.2055674518201285, |
| "grad_norm": 0.41472136974334717, |
| "learning_rate": 9.843364502716225e-06, |
| "loss": 0.1671, |
| "num_input_tokens_seen": 3582240, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.291220556745182, |
| "grad_norm": 0.5156757235527039, |
| "learning_rate": 9.820157238140535e-06, |
| "loss": 0.1682, |
| "num_input_tokens_seen": 3838160, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.3768736616702355, |
| "grad_norm": 0.5046593546867371, |
| "learning_rate": 9.795378704151675e-06, |
| "loss": 0.1651, |
| "num_input_tokens_seen": 4092304, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.462526766595289, |
| "grad_norm": 0.5588434338569641, |
| "learning_rate": 9.76903697802817e-06, |
| "loss": 0.1649, |
| "num_input_tokens_seen": 4346640, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.5481798715203428, |
| "grad_norm": 0.46262454986572266, |
| "learning_rate": 9.741140646616161e-06, |
| "loss": 0.1669, |
| "num_input_tokens_seen": 4602192, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.633832976445396, |
| "grad_norm": 0.45427972078323364, |
| "learning_rate": 9.711698803530253e-06, |
| "loss": 0.1674, |
| "num_input_tokens_seen": 4858240, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.7194860813704498, |
| "grad_norm": 0.4514879882335663, |
| "learning_rate": 9.68072104618921e-06, |
| "loss": 0.1641, |
| "num_input_tokens_seen": 5113408, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.805139186295503, |
| "grad_norm": 0.7933849692344666, |
| "learning_rate": 9.648217472687385e-06, |
| "loss": 0.1614, |
| "num_input_tokens_seen": 5368352, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.8907922912205568, |
| "grad_norm": 0.6207934021949768, |
| "learning_rate": 9.614198678502965e-06, |
| "loss": 0.163, |
| "num_input_tokens_seen": 5622128, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.9764453961456103, |
| "grad_norm": 0.8193040490150452, |
| "learning_rate": 9.57867575304406e-06, |
| "loss": 0.1589, |
| "num_input_tokens_seen": 5876816, |
| "step": 230 |
| }, |
| { |
| "epoch": 2.068522483940043, |
| "grad_norm": 1.0469202995300293, |
| "learning_rate": 9.541660276033795e-06, |
| "loss": 0.1755, |
| "num_input_tokens_seen": 6145392, |
| "step": 240 |
| }, |
| { |
| "epoch": 2.154175588865096, |
| "grad_norm": 0.9274189472198486, |
| "learning_rate": 9.503164313735566e-06, |
| "loss": 0.1595, |
| "num_input_tokens_seen": 6399504, |
| "step": 250 |
| }, |
| { |
| "epoch": 2.23982869379015, |
| "grad_norm": 0.6875982880592346, |
| "learning_rate": 9.46320041501969e-06, |
| "loss": 0.1563, |
| "num_input_tokens_seen": 6654160, |
| "step": 260 |
| }, |
| { |
| "epoch": 2.325481798715203, |
| "grad_norm": 0.5835751295089722, |
| "learning_rate": 9.421781607272741e-06, |
| "loss": 0.1554, |
| "num_input_tokens_seen": 6910752, |
| "step": 270 |
| }, |
| { |
| "epoch": 2.411134903640257, |
| "grad_norm": 0.6475698351860046, |
| "learning_rate": 9.378921392150893e-06, |
| "loss": 0.1579, |
| "num_input_tokens_seen": 7166960, |
| "step": 280 |
| }, |
| { |
| "epoch": 2.4967880085653107, |
| "grad_norm": 0.6029316782951355, |
| "learning_rate": 9.33463374117867e-06, |
| "loss": 0.1577, |
| "num_input_tokens_seen": 7420288, |
| "step": 290 |
| }, |
| { |
| "epoch": 2.582441113490364, |
| "grad_norm": 0.6444355845451355, |
| "learning_rate": 9.288933091194524e-06, |
| "loss": 0.1564, |
| "num_input_tokens_seen": 7675184, |
| "step": 300 |
| }, |
| { |
| "epoch": 2.6680942184154177, |
| "grad_norm": 0.5541071891784668, |
| "learning_rate": 9.241834339644726e-06, |
| "loss": 0.1528, |
| "num_input_tokens_seen": 7926976, |
| "step": 310 |
| }, |
| { |
| "epoch": 2.753747323340471, |
| "grad_norm": 0.6703725457191467, |
| "learning_rate": 9.193352839727122e-06, |
| "loss": 0.1549, |
| "num_input_tokens_seen": 8184992, |
| "step": 320 |
| }, |
| { |
| "epoch": 2.8394004282655247, |
| "grad_norm": 0.6515584588050842, |
| "learning_rate": 9.143504395386302e-06, |
| "loss": 0.157, |
| "num_input_tokens_seen": 8439712, |
| "step": 330 |
| }, |
| { |
| "epoch": 2.925053533190578, |
| "grad_norm": 0.5527693629264832, |
| "learning_rate": 9.09230525616186e-06, |
| "loss": 0.157, |
| "num_input_tokens_seen": 8694080, |
| "step": 340 |
| }, |
| { |
| "epoch": 3.017130620985011, |
| "grad_norm": 0.6593677401542664, |
| "learning_rate": 9.039772111891383e-06, |
| "loss": 0.1672, |
| "num_input_tokens_seen": 8965488, |
| "step": 350 |
| }, |
| { |
| "epoch": 3.102783725910064, |
| "grad_norm": 0.5042828917503357, |
| "learning_rate": 8.985922087269916e-06, |
| "loss": 0.1483, |
| "num_input_tokens_seen": 9220480, |
| "step": 360 |
| }, |
| { |
| "epoch": 3.188436830835118, |
| "grad_norm": 0.4123888611793518, |
| "learning_rate": 8.930772736267675e-06, |
| "loss": 0.1532, |
| "num_input_tokens_seen": 9477024, |
| "step": 370 |
| }, |
| { |
| "epoch": 3.274089935760171, |
| "grad_norm": 0.7851901054382324, |
| "learning_rate": 8.874342036407815e-06, |
| "loss": 0.1508, |
| "num_input_tokens_seen": 9731840, |
| "step": 380 |
| }, |
| { |
| "epoch": 3.359743040685225, |
| "grad_norm": 0.7545840740203857, |
| "learning_rate": 8.816648382906154e-06, |
| "loss": 0.1516, |
| "num_input_tokens_seen": 9986704, |
| "step": 390 |
| }, |
| { |
| "epoch": 3.445396145610278, |
| "grad_norm": 0.7439327239990234, |
| "learning_rate": 8.757710582674708e-06, |
| "loss": 0.1506, |
| "num_input_tokens_seen": 10238720, |
| "step": 400 |
| }, |
| { |
| "epoch": 3.531049250535332, |
| "grad_norm": 0.8343164920806885, |
| "learning_rate": 8.697547848191037e-06, |
| "loss": 0.1516, |
| "num_input_tokens_seen": 10491856, |
| "step": 410 |
| }, |
| { |
| "epoch": 3.6167023554603857, |
| "grad_norm": 0.817565381526947, |
| "learning_rate": 8.63617979123539e-06, |
| "loss": 0.1542, |
| "num_input_tokens_seen": 10744240, |
| "step": 420 |
| }, |
| { |
| "epoch": 3.702355460385439, |
| "grad_norm": 0.5334470272064209, |
| "learning_rate": 8.573626416497669e-06, |
| "loss": 0.1446, |
| "num_input_tokens_seen": 10996768, |
| "step": 430 |
| }, |
| { |
| "epoch": 3.7880085653104922, |
| "grad_norm": 0.9441611766815186, |
| "learning_rate": 8.509908115056334e-06, |
| "loss": 0.1515, |
| "num_input_tokens_seen": 11254560, |
| "step": 440 |
| }, |
| { |
| "epoch": 3.873661670235546, |
| "grad_norm": 0.6177489757537842, |
| "learning_rate": 8.445045657731329e-06, |
| "loss": 0.1513, |
| "num_input_tokens_seen": 11512992, |
| "step": 450 |
| }, |
| { |
| "epoch": 3.9593147751605997, |
| "grad_norm": 0.5743350982666016, |
| "learning_rate": 8.379060188313244e-06, |
| "loss": 0.1458, |
| "num_input_tokens_seen": 11765808, |
| "step": 460 |
| }, |
| { |
| "epoch": 4.0513918629550325, |
| "grad_norm": 0.8525713086128235, |
| "learning_rate": 8.311973216670888e-06, |
| "loss": 0.1598, |
| "num_input_tokens_seen": 12036784, |
| "step": 470 |
| }, |
| { |
| "epoch": 4.137044967880086, |
| "grad_norm": 0.6399952173233032, |
| "learning_rate": 8.243806611739516e-06, |
| "loss": 0.1448, |
| "num_input_tokens_seen": 12290592, |
| "step": 480 |
| }, |
| { |
| "epoch": 4.222698072805139, |
| "grad_norm": 0.657546877861023, |
| "learning_rate": 8.17458259439202e-06, |
| "loss": 0.144, |
| "num_input_tokens_seen": 12542464, |
| "step": 490 |
| }, |
| { |
| "epoch": 4.308351177730192, |
| "grad_norm": 0.6414650678634644, |
| "learning_rate": 8.104323730195407e-06, |
| "loss": 0.1406, |
| "num_input_tokens_seen": 12796848, |
| "step": 500 |
| }, |
| { |
| "epoch": 4.394004282655247, |
| "grad_norm": 0.7480872869491577, |
| "learning_rate": 8.033052922054882e-06, |
| "loss": 0.1436, |
| "num_input_tokens_seen": 13051760, |
| "step": 510 |
| }, |
| { |
| "epoch": 4.4796573875803, |
| "grad_norm": 0.7025752067565918, |
| "learning_rate": 7.960793402748001e-06, |
| "loss": 0.147, |
| "num_input_tokens_seen": 13305808, |
| "step": 520 |
| }, |
| { |
| "epoch": 4.565310492505353, |
| "grad_norm": 0.5708986520767212, |
| "learning_rate": 7.887568727351262e-06, |
| "loss": 0.1456, |
| "num_input_tokens_seen": 13563056, |
| "step": 530 |
| }, |
| { |
| "epoch": 4.650963597430406, |
| "grad_norm": 0.6903087496757507, |
| "learning_rate": 7.813402765561664e-06, |
| "loss": 0.143, |
| "num_input_tokens_seen": 13816992, |
| "step": 540 |
| }, |
| { |
| "epoch": 4.736616702355461, |
| "grad_norm": 0.6083903908729553, |
| "learning_rate": 7.738319693915673e-06, |
| "loss": 0.1439, |
| "num_input_tokens_seen": 14071936, |
| "step": 550 |
| }, |
| { |
| "epoch": 4.822269807280514, |
| "grad_norm": 0.6583831906318665, |
| "learning_rate": 7.662343987908195e-06, |
| "loss": 0.147, |
| "num_input_tokens_seen": 14327440, |
| "step": 560 |
| }, |
| { |
| "epoch": 4.907922912205567, |
| "grad_norm": 0.8827478885650635, |
| "learning_rate": 7.585500414014077e-06, |
| "loss": 0.1467, |
| "num_input_tokens_seen": 14582832, |
| "step": 570 |
| }, |
| { |
| "epoch": 4.993576017130621, |
| "grad_norm": 0.8274891972541809, |
| "learning_rate": 7.507814021614761e-06, |
| "loss": 0.1478, |
| "num_input_tokens_seen": 14839136, |
| "step": 580 |
| }, |
| { |
| "epoch": 5.085653104925053, |
| "grad_norm": 1.3195112943649292, |
| "learning_rate": 7.429310134832709e-06, |
| "loss": 0.1517, |
| "num_input_tokens_seen": 15109264, |
| "step": 590 |
| }, |
| { |
| "epoch": 5.1713062098501075, |
| "grad_norm": 0.7981224656105042, |
| "learning_rate": 7.35001434427628e-06, |
| "loss": 0.1396, |
| "num_input_tokens_seen": 15363824, |
| "step": 600 |
| }, |
| { |
| "epoch": 5.256959314775161, |
| "grad_norm": 0.6522560715675354, |
| "learning_rate": 7.269952498697734e-06, |
| "loss": 0.142, |
| "num_input_tokens_seen": 15618576, |
| "step": 610 |
| }, |
| { |
| "epoch": 5.342612419700214, |
| "grad_norm": 0.7629905343055725, |
| "learning_rate": 7.189150696567081e-06, |
| "loss": 0.1384, |
| "num_input_tokens_seen": 15871056, |
| "step": 620 |
| }, |
| { |
| "epoch": 5.428265524625267, |
| "grad_norm": 0.9554848670959473, |
| "learning_rate": 7.10763527756453e-06, |
| "loss": 0.1405, |
| "num_input_tokens_seen": 16124976, |
| "step": 630 |
| }, |
| { |
| "epoch": 5.5139186295503215, |
| "grad_norm": 0.8175866603851318, |
| "learning_rate": 7.025432813994315e-06, |
| "loss": 0.1357, |
| "num_input_tokens_seen": 16381680, |
| "step": 640 |
| }, |
| { |
| "epoch": 5.599571734475375, |
| "grad_norm": 0.7990790009498596, |
| "learning_rate": 6.942570102122679e-06, |
| "loss": 0.1387, |
| "num_input_tokens_seen": 16638048, |
| "step": 650 |
| }, |
| { |
| "epoch": 5.685224839400428, |
| "grad_norm": 0.9116854667663574, |
| "learning_rate": 6.859074153442864e-06, |
| "loss": 0.1414, |
| "num_input_tokens_seen": 16894688, |
| "step": 660 |
| }, |
| { |
| "epoch": 5.770877944325482, |
| "grad_norm": 0.7633938789367676, |
| "learning_rate": 6.774972185869928e-06, |
| "loss": 0.1389, |
| "num_input_tokens_seen": 17147808, |
| "step": 670 |
| }, |
| { |
| "epoch": 5.856531049250536, |
| "grad_norm": 0.8924551606178284, |
| "learning_rate": 6.690291614868287e-06, |
| "loss": 0.1361, |
| "num_input_tokens_seen": 17403280, |
| "step": 680 |
| }, |
| { |
| "epoch": 5.942184154175589, |
| "grad_norm": 0.8566009998321533, |
| "learning_rate": 6.60506004451485e-06, |
| "loss": 0.1356, |
| "num_input_tokens_seen": 17657888, |
| "step": 690 |
| }, |
| { |
| "epoch": 6.034261241970022, |
| "grad_norm": 0.9057173132896423, |
| "learning_rate": 6.5193052585006666e-06, |
| "loss": 0.1483, |
| "num_input_tokens_seen": 17927520, |
| "step": 700 |
| }, |
| { |
| "epoch": 6.119914346895075, |
| "grad_norm": 0.9895085692405701, |
| "learning_rate": 6.433055211074042e-06, |
| "loss": 0.1308, |
| "num_input_tokens_seen": 18184352, |
| "step": 710 |
| }, |
| { |
| "epoch": 6.205567451820128, |
| "grad_norm": 1.0845868587493896, |
| "learning_rate": 6.346338017928036e-06, |
| "loss": 0.1269, |
| "num_input_tokens_seen": 18437792, |
| "step": 720 |
| }, |
| { |
| "epoch": 6.291220556745182, |
| "grad_norm": 1.021283745765686, |
| "learning_rate": 6.2591819470353424e-06, |
| "loss": 0.1301, |
| "num_input_tokens_seen": 18690144, |
| "step": 730 |
| }, |
| { |
| "epoch": 6.376873661670236, |
| "grad_norm": 1.1350120306015015, |
| "learning_rate": 6.171615409433525e-06, |
| "loss": 0.1275, |
| "num_input_tokens_seen": 18944688, |
| "step": 740 |
| }, |
| { |
| "epoch": 6.462526766595289, |
| "grad_norm": 1.0572874546051025, |
| "learning_rate": 6.0836669499636255e-06, |
| "loss": 0.1264, |
| "num_input_tokens_seen": 19199984, |
| "step": 750 |
| }, |
| { |
| "epoch": 6.548179871520342, |
| "grad_norm": 1.1884225606918335, |
| "learning_rate": 5.995365237965144e-06, |
| "loss": 0.1294, |
| "num_input_tokens_seen": 19452032, |
| "step": 760 |
| }, |
| { |
| "epoch": 6.6338329764453965, |
| "grad_norm": 0.9745492339134216, |
| "learning_rate": 5.906739057930439e-06, |
| "loss": 0.1262, |
| "num_input_tokens_seen": 19707040, |
| "step": 770 |
| }, |
| { |
| "epoch": 6.71948608137045, |
| "grad_norm": 1.090391755104065, |
| "learning_rate": 5.817817300121592e-06, |
| "loss": 0.1266, |
| "num_input_tokens_seen": 19962960, |
| "step": 780 |
| }, |
| { |
| "epoch": 6.805139186295503, |
| "grad_norm": 1.1640676259994507, |
| "learning_rate": 5.728628951152799e-06, |
| "loss": 0.1324, |
| "num_input_tokens_seen": 20219008, |
| "step": 790 |
| }, |
| { |
| "epoch": 6.890792291220556, |
| "grad_norm": 0.9813507199287415, |
| "learning_rate": 5.639203084541338e-06, |
| "loss": 0.1338, |
| "num_input_tokens_seen": 20473664, |
| "step": 800 |
| }, |
| { |
| "epoch": 6.9764453961456105, |
| "grad_norm": 1.11289644241333, |
| "learning_rate": 5.549568851230219e-06, |
| "loss": 0.1273, |
| "num_input_tokens_seen": 20727296, |
| "step": 810 |
| }, |
| { |
| "epoch": 7.0685224839400425, |
| "grad_norm": 1.5624918937683105, |
| "learning_rate": 5.459755470085595e-06, |
| "loss": 0.1332, |
| "num_input_tokens_seen": 20996432, |
| "step": 820 |
| }, |
| { |
| "epoch": 7.154175588865097, |
| "grad_norm": 1.3339862823486328, |
| "learning_rate": 5.369792218372026e-06, |
| "loss": 0.1104, |
| "num_input_tokens_seen": 21252272, |
| "step": 830 |
| }, |
| { |
| "epoch": 7.23982869379015, |
| "grad_norm": 1.5236716270446777, |
| "learning_rate": 5.2797084222087105e-06, |
| "loss": 0.1114, |
| "num_input_tokens_seen": 21508208, |
| "step": 840 |
| }, |
| { |
| "epoch": 7.325481798715203, |
| "grad_norm": 1.4154669046401978, |
| "learning_rate": 5.189533447009795e-06, |
| "loss": 0.1134, |
| "num_input_tokens_seen": 21765536, |
| "step": 850 |
| }, |
| { |
| "epoch": 7.4111349036402565, |
| "grad_norm": 1.5260732173919678, |
| "learning_rate": 5.099296687911858e-06, |
| "loss": 0.1102, |
| "num_input_tokens_seen": 22020160, |
| "step": 860 |
| }, |
| { |
| "epoch": 7.496788008565311, |
| "grad_norm": 1.2989623546600342, |
| "learning_rate": 5.009027560191732e-06, |
| "loss": 0.1122, |
| "num_input_tokens_seen": 22274400, |
| "step": 870 |
| }, |
| { |
| "epoch": 7.582441113490364, |
| "grad_norm": 1.4925442934036255, |
| "learning_rate": 4.918755489677729e-06, |
| "loss": 0.1094, |
| "num_input_tokens_seen": 22526464, |
| "step": 880 |
| }, |
| { |
| "epoch": 7.668094218415417, |
| "grad_norm": 1.3059921264648438, |
| "learning_rate": 4.828509903157451e-06, |
| "loss": 0.1128, |
| "num_input_tokens_seen": 22779664, |
| "step": 890 |
| }, |
| { |
| "epoch": 7.7537473233404715, |
| "grad_norm": 1.6819276809692383, |
| "learning_rate": 4.738320218785281e-06, |
| "loss": 0.1146, |
| "num_input_tokens_seen": 23036160, |
| "step": 900 |
| }, |
| { |
| "epoch": 7.839400428265525, |
| "grad_norm": 1.3909580707550049, |
| "learning_rate": 4.648215836492682e-06, |
| "loss": 0.1145, |
| "num_input_tokens_seen": 23292016, |
| "step": 910 |
| }, |
| { |
| "epoch": 7.925053533190578, |
| "grad_norm": 1.7210851907730103, |
| "learning_rate": 4.5582261284044385e-06, |
| "loss": 0.1156, |
| "num_input_tokens_seen": 23544800, |
| "step": 920 |
| }, |
| { |
| "epoch": 8.01713062098501, |
| "grad_norm": 1.2723944187164307, |
| "learning_rate": 4.468380429263973e-06, |
| "loss": 0.1197, |
| "num_input_tokens_seen": 23816288, |
| "step": 930 |
| }, |
| { |
| "epoch": 8.102783725910065, |
| "grad_norm": 1.9091925621032715, |
| "learning_rate": 4.378708026870825e-06, |
| "loss": 0.0916, |
| "num_input_tokens_seen": 24071488, |
| "step": 940 |
| }, |
| { |
| "epoch": 8.188436830835117, |
| "grad_norm": 1.7839370965957642, |
| "learning_rate": 4.289238152533465e-06, |
| "loss": 0.0893, |
| "num_input_tokens_seen": 24324720, |
| "step": 950 |
| }, |
| { |
| "epoch": 8.274089935760172, |
| "grad_norm": 2.00311541557312, |
| "learning_rate": 4.199999971540489e-06, |
| "loss": 0.0889, |
| "num_input_tokens_seen": 24579648, |
| "step": 960 |
| }, |
| { |
| "epoch": 8.359743040685224, |
| "grad_norm": 2.047337293624878, |
| "learning_rate": 4.111022573653366e-06, |
| "loss": 0.0873, |
| "num_input_tokens_seen": 24833840, |
| "step": 970 |
| }, |
| { |
| "epoch": 8.445396145610278, |
| "grad_norm": 1.9115785360336304, |
| "learning_rate": 4.0223349636237766e-06, |
| "loss": 0.0904, |
| "num_input_tokens_seen": 25089776, |
| "step": 980 |
| }, |
| { |
| "epoch": 8.531049250535332, |
| "grad_norm": 1.8445810079574585, |
| "learning_rate": 3.933966051738684e-06, |
| "loss": 0.088, |
| "num_input_tokens_seen": 25345264, |
| "step": 990 |
| }, |
| { |
| "epoch": 8.616702355460385, |
| "grad_norm": 1.6529115438461304, |
| "learning_rate": 3.845944644396194e-06, |
| "loss": 0.0919, |
| "num_input_tokens_seen": 25598112, |
| "step": 1000 |
| }, |
| { |
| "epoch": 8.702355460385439, |
| "grad_norm": 2.129995346069336, |
| "learning_rate": 3.758299434715268e-06, |
| "loss": 0.0906, |
| "num_input_tokens_seen": 25851728, |
| "step": 1010 |
| }, |
| { |
| "epoch": 8.788008565310493, |
| "grad_norm": 2.1039373874664307, |
| "learning_rate": 3.6710589931823837e-06, |
| "loss": 0.0895, |
| "num_input_tokens_seen": 26104704, |
| "step": 1020 |
| }, |
| { |
| "epoch": 8.873661670235546, |
| "grad_norm": 2.058598518371582, |
| "learning_rate": 3.584251758338151e-06, |
| "loss": 0.0923, |
| "num_input_tokens_seen": 26361680, |
| "step": 1030 |
| }, |
| { |
| "epoch": 8.9593147751606, |
| "grad_norm": 1.8930065631866455, |
| "learning_rate": 3.4979060275069576e-06, |
| "loss": 0.0908, |
| "num_input_tokens_seen": 26617536, |
| "step": 1040 |
| }, |
| { |
| "epoch": 9.051391862955033, |
| "grad_norm": 1.8233646154403687, |
| "learning_rate": 3.4120499475726266e-06, |
| "loss": 0.0847, |
| "num_input_tokens_seen": 26888160, |
| "step": 1050 |
| }, |
| { |
| "epoch": 9.137044967880085, |
| "grad_norm": 2.1758053302764893, |
| "learning_rate": 3.3267115058031418e-06, |
| "loss": 0.0657, |
| "num_input_tokens_seen": 27142528, |
| "step": 1060 |
| }, |
| { |
| "epoch": 9.222698072805139, |
| "grad_norm": 2.0327367782592773, |
| "learning_rate": 3.2419185207273816e-06, |
| "loss": 0.0662, |
| "num_input_tokens_seen": 27394144, |
| "step": 1070 |
| }, |
| { |
| "epoch": 9.308351177730193, |
| "grad_norm": 2.2035434246063232, |
| "learning_rate": 3.157698633066863e-06, |
| "loss": 0.0665, |
| "num_input_tokens_seen": 27649488, |
| "step": 1080 |
| }, |
| { |
| "epoch": 9.394004282655246, |
| "grad_norm": 1.9066494703292847, |
| "learning_rate": 3.0740792967254606e-06, |
| "loss": 0.0642, |
| "num_input_tokens_seen": 27904992, |
| "step": 1090 |
| }, |
| { |
| "epoch": 9.4796573875803, |
| "grad_norm": 2.2175674438476562, |
| "learning_rate": 2.991087769840001e-06, |
| "loss": 0.0625, |
| "num_input_tokens_seen": 28160336, |
| "step": 1100 |
| }, |
| { |
| "epoch": 9.565310492505354, |
| "grad_norm": 2.435115337371826, |
| "learning_rate": 2.9087511058947014e-06, |
| "loss": 0.0643, |
| "num_input_tokens_seen": 28417360, |
| "step": 1110 |
| }, |
| { |
| "epoch": 9.650963597430406, |
| "grad_norm": 2.237015724182129, |
| "learning_rate": 2.827096144902289e-06, |
| "loss": 0.0645, |
| "num_input_tokens_seen": 28670512, |
| "step": 1120 |
| }, |
| { |
| "epoch": 9.73661670235546, |
| "grad_norm": 2.473604202270508, |
| "learning_rate": 2.7461495046547436e-06, |
| "loss": 0.068, |
| "num_input_tokens_seen": 28927232, |
| "step": 1130 |
| }, |
| { |
| "epoch": 9.822269807280513, |
| "grad_norm": 2.220705270767212, |
| "learning_rate": 2.665937572046432e-06, |
| "loss": 0.0647, |
| "num_input_tokens_seen": 29182768, |
| "step": 1140 |
| }, |
| { |
| "epoch": 9.907922912205567, |
| "grad_norm": 2.652024269104004, |
| "learning_rate": 2.586486494472572e-06, |
| "loss": 0.0644, |
| "num_input_tokens_seen": 29437936, |
| "step": 1150 |
| }, |
| { |
| "epoch": 9.993576017130621, |
| "grad_norm": 2.180983304977417, |
| "learning_rate": 2.5078221713057048e-06, |
| "loss": 0.0658, |
| "num_input_tokens_seen": 29690944, |
| "step": 1160 |
| }, |
| { |
| "epoch": 10.085653104925054, |
| "grad_norm": 1.6538355350494385, |
| "learning_rate": 2.4299702454530605e-06, |
| "loss": 0.053, |
| "num_input_tokens_seen": 29964448, |
| "step": 1170 |
| }, |
| { |
| "epoch": 10.171306209850107, |
| "grad_norm": 2.229673147201538, |
| "learning_rate": 2.3529560949975184e-06, |
| "loss": 0.0446, |
| "num_input_tokens_seen": 30215952, |
| "step": 1180 |
| }, |
| { |
| "epoch": 10.25695931477516, |
| "grad_norm": 1.8106822967529297, |
| "learning_rate": 2.2768048249248648e-06, |
| "loss": 0.0449, |
| "num_input_tokens_seen": 30471952, |
| "step": 1190 |
| }, |
| { |
| "epoch": 10.342612419700215, |
| "grad_norm": 2.150508403778076, |
| "learning_rate": 2.201541258940129e-06, |
| "loss": 0.0422, |
| "num_input_tokens_seen": 30727376, |
| "step": 1200 |
| }, |
| { |
| "epoch": 10.428265524625267, |
| "grad_norm": 2.0471906661987305, |
| "learning_rate": 2.12718993137555e-06, |
| "loss": 0.0461, |
| "num_input_tokens_seen": 30983760, |
| "step": 1210 |
| }, |
| { |
| "epoch": 10.513918629550322, |
| "grad_norm": 2.299278497695923, |
| "learning_rate": 2.0537750791929296e-06, |
| "loss": 0.0458, |
| "num_input_tokens_seen": 31238720, |
| "step": 1220 |
| }, |
| { |
| "epoch": 10.599571734475374, |
| "grad_norm": 2.1924257278442383, |
| "learning_rate": 1.981320634082873e-06, |
| "loss": 0.0434, |
| "num_input_tokens_seen": 31494560, |
| "step": 1230 |
| }, |
| { |
| "epoch": 10.685224839400428, |
| "grad_norm": 2.3524584770202637, |
| "learning_rate": 1.909850214663575e-06, |
| "loss": 0.0452, |
| "num_input_tokens_seen": 31750784, |
| "step": 1240 |
| }, |
| { |
| "epoch": 10.770877944325482, |
| "grad_norm": 2.2468934059143066, |
| "learning_rate": 1.8393871187816526e-06, |
| "loss": 0.0447, |
| "num_input_tokens_seen": 32005120, |
| "step": 1250 |
| }, |
| { |
| "epoch": 10.856531049250535, |
| "grad_norm": 2.448117971420288, |
| "learning_rate": 1.7699543159175215e-06, |
| "loss": 0.0449, |
| "num_input_tokens_seen": 32258480, |
| "step": 1260 |
| }, |
| { |
| "epoch": 10.942184154175589, |
| "grad_norm": 2.0848143100738525, |
| "learning_rate": 1.7015744396978557e-06, |
| "loss": 0.0442, |
| "num_input_tokens_seen": 32510944, |
| "step": 1270 |
| }, |
| { |
| "epoch": 11.034261241970022, |
| "grad_norm": 1.6036432981491089, |
| "learning_rate": 1.634269780517483e-06, |
| "loss": 0.0435, |
| "num_input_tokens_seen": 32780608, |
| "step": 1280 |
| }, |
| { |
| "epoch": 11.119914346895074, |
| "grad_norm": 3.015963315963745, |
| "learning_rate": 1.568062278273197e-06, |
| "loss": 0.0286, |
| "num_input_tokens_seen": 33034112, |
| "step": 1290 |
| }, |
| { |
| "epoch": 11.205567451820128, |
| "grad_norm": 1.6929532289505005, |
| "learning_rate": 1.5029735152118125e-06, |
| "loss": 0.0308, |
| "num_input_tokens_seen": 33290224, |
| "step": 1300 |
| }, |
| { |
| "epoch": 11.291220556745182, |
| "grad_norm": 1.9741885662078857, |
| "learning_rate": 1.4390247088948073e-06, |
| "loss": 0.0309, |
| "num_input_tokens_seen": 33544448, |
| "step": 1310 |
| }, |
| { |
| "epoch": 11.376873661670235, |
| "grad_norm": 1.5955508947372437, |
| "learning_rate": 1.3762367052818527e-06, |
| "loss": 0.0275, |
| "num_input_tokens_seen": 33799536, |
| "step": 1320 |
| }, |
| { |
| "epoch": 11.462526766595289, |
| "grad_norm": 2.293123245239258, |
| "learning_rate": 1.3146299719354544e-06, |
| "loss": 0.0304, |
| "num_input_tokens_seen": 34055952, |
| "step": 1330 |
| }, |
| { |
| "epoch": 11.548179871520343, |
| "grad_norm": 1.8011912107467651, |
| "learning_rate": 1.254224591348983e-06, |
| "loss": 0.0299, |
| "num_input_tokens_seen": 34310000, |
| "step": 1340 |
| }, |
| { |
| "epoch": 11.633832976445396, |
| "grad_norm": 1.8339879512786865, |
| "learning_rate": 1.1950402544001849e-06, |
| "loss": 0.0311, |
| "num_input_tokens_seen": 34565680, |
| "step": 1350 |
| }, |
| { |
| "epoch": 11.71948608137045, |
| "grad_norm": 1.6808807849884033, |
| "learning_rate": 1.1370962539323837e-06, |
| "loss": 0.0314, |
| "num_input_tokens_seen": 34820768, |
| "step": 1360 |
| }, |
| { |
| "epoch": 11.805139186295504, |
| "grad_norm": 1.7647879123687744, |
| "learning_rate": 1.0804114784654158e-06, |
| "loss": 0.0311, |
| "num_input_tokens_seen": 35074016, |
| "step": 1370 |
| }, |
| { |
| "epoch": 11.890792291220556, |
| "grad_norm": 1.753990650177002, |
| "learning_rate": 1.0250044060383734e-06, |
| "loss": 0.0299, |
| "num_input_tokens_seen": 35328272, |
| "step": 1380 |
| }, |
| { |
| "epoch": 11.97644539614561, |
| "grad_norm": 2.10841965675354, |
| "learning_rate": 9.708930981861603e-07, |
| "loss": 0.03, |
| "num_input_tokens_seen": 35582880, |
| "step": 1390 |
| }, |
| { |
| "epoch": 12.068522483940043, |
| "grad_norm": 1.4194451570510864, |
| "learning_rate": 9.180951940518002e-07, |
| "loss": 0.026, |
| "num_input_tokens_seen": 35853280, |
| "step": 1400 |
| }, |
| { |
| "epoch": 12.154175588865096, |
| "grad_norm": 1.612318515777588, |
| "learning_rate": 8.666279046364595e-07, |
| "loss": 0.0208, |
| "num_input_tokens_seen": 36106816, |
| "step": 1410 |
| }, |
| { |
| "epoch": 12.23982869379015, |
| "grad_norm": 1.6022765636444092, |
| "learning_rate": 8.165080071890208e-07, |
| "loss": 0.0205, |
| "num_input_tokens_seen": 36359232, |
| "step": 1420 |
| }, |
| { |
| "epoch": 12.325481798715204, |
| "grad_norm": 1.608430027961731, |
| "learning_rate": 7.677518397370548e-07, |
| "loss": 0.0228, |
| "num_input_tokens_seen": 36614176, |
| "step": 1430 |
| }, |
| { |
| "epoch": 12.411134903640257, |
| "grad_norm": 1.4423803091049194, |
| "learning_rate": 7.203752957609672e-07, |
| "loss": 0.0207, |
| "num_input_tokens_seen": 36868400, |
| "step": 1440 |
| }, |
| { |
| "epoch": 12.49678800856531, |
| "grad_norm": 1.6684809923171997, |
| "learning_rate": 6.743938190130616e-07, |
| "loss": 0.0215, |
| "num_input_tokens_seen": 37121536, |
| "step": 1450 |
| }, |
| { |
| "epoch": 12.582441113490365, |
| "grad_norm": 1.7179003953933716, |
| "learning_rate": 6.298223984832047e-07, |
| "loss": 0.0216, |
| "num_input_tokens_seen": 37377168, |
| "step": 1460 |
| }, |
| { |
| "epoch": 12.668094218415417, |
| "grad_norm": 1.6454778909683228, |
| "learning_rate": 5.866755635127247e-07, |
| "loss": 0.0207, |
| "num_input_tokens_seen": 37632992, |
| "step": 1470 |
| }, |
| { |
| "epoch": 12.753747323340471, |
| "grad_norm": 1.8044767379760742, |
| "learning_rate": 5.449673790581611e-07, |
| "loss": 0.0217, |
| "num_input_tokens_seen": 37888640, |
| "step": 1480 |
| }, |
| { |
| "epoch": 12.839400428265524, |
| "grad_norm": 1.874295711517334, |
| "learning_rate": 5.04711441106382e-07, |
| "loss": 0.0197, |
| "num_input_tokens_seen": 38143760, |
| "step": 1490 |
| }, |
| { |
| "epoch": 12.925053533190578, |
| "grad_norm": 1.3250926733016968, |
| "learning_rate": 4.659208722425806e-07, |
| "loss": 0.0207, |
| "num_input_tokens_seen": 38398560, |
| "step": 1500 |
| }, |
| { |
| "epoch": 13.01713062098501, |
| "grad_norm": 1.2411588430404663, |
| "learning_rate": 4.2860831737258857e-07, |
| "loss": 0.0216, |
| "num_input_tokens_seen": 38670912, |
| "step": 1510 |
| }, |
| { |
| "epoch": 13.102783725910065, |
| "grad_norm": 1.3138427734375, |
| "learning_rate": 3.9278593960090873e-07, |
| "loss": 0.0167, |
| "num_input_tokens_seen": 38925872, |
| "step": 1520 |
| }, |
| { |
| "epoch": 13.188436830835117, |
| "grad_norm": 1.362457036972046, |
| "learning_rate": 3.5846541626579026e-07, |
| "loss": 0.0159, |
| "num_input_tokens_seen": 39183632, |
| "step": 1530 |
| }, |
| { |
| "epoch": 13.274089935760172, |
| "grad_norm": 1.515376091003418, |
| "learning_rate": 3.256579351326744e-07, |
| "loss": 0.0156, |
| "num_input_tokens_seen": 39440864, |
| "step": 1540 |
| }, |
| { |
| "epoch": 13.359743040685224, |
| "grad_norm": 1.4070255756378174, |
| "learning_rate": 2.94374190747212e-07, |
| "loss": 0.0166, |
| "num_input_tokens_seen": 39695712, |
| "step": 1550 |
| }, |
| { |
| "epoch": 13.445396145610278, |
| "grad_norm": 1.4853448867797852, |
| "learning_rate": 2.64624380949069e-07, |
| "loss": 0.0173, |
| "num_input_tokens_seen": 39950304, |
| "step": 1560 |
| }, |
| { |
| "epoch": 13.531049250535332, |
| "grad_norm": 1.542286992073059, |
| "learning_rate": 2.3641820354764755e-07, |
| "loss": 0.0165, |
| "num_input_tokens_seen": 40203616, |
| "step": 1570 |
| }, |
| { |
| "epoch": 13.616702355460385, |
| "grad_norm": 1.565663456916809, |
| "learning_rate": 2.0976485316080375e-07, |
| "loss": 0.0167, |
| "num_input_tokens_seen": 40458464, |
| "step": 1580 |
| }, |
| { |
| "epoch": 13.702355460385439, |
| "grad_norm": 1.3701163530349731, |
| "learning_rate": 1.846730182175993e-07, |
| "loss": 0.017, |
| "num_input_tokens_seen": 40711216, |
| "step": 1590 |
| }, |
| { |
| "epoch": 13.788008565310493, |
| "grad_norm": 1.4886751174926758, |
| "learning_rate": 1.6115087812605123e-07, |
| "loss": 0.015, |
| "num_input_tokens_seen": 40965856, |
| "step": 1600 |
| }, |
| { |
| "epoch": 13.873661670235546, |
| "grad_norm": 1.2140471935272217, |
| "learning_rate": 1.392061006068246e-07, |
| "loss": 0.0169, |
| "num_input_tokens_seen": 41220736, |
| "step": 1610 |
| }, |
| { |
| "epoch": 13.9593147751606, |
| "grad_norm": 1.314063549041748, |
| "learning_rate": 1.1884583919371251e-07, |
| "loss": 0.0164, |
| "num_input_tokens_seen": 41473952, |
| "step": 1620 |
| }, |
| { |
| "epoch": 14.051391862955033, |
| "grad_norm": 1.2103674411773682, |
| "learning_rate": 1.0007673090173808e-07, |
| "loss": 0.0168, |
| "num_input_tokens_seen": 41742832, |
| "step": 1630 |
| }, |
| { |
| "epoch": 14.137044967880085, |
| "grad_norm": 1.250216007232666, |
| "learning_rate": 8.29048940636279e-08, |
| "loss": 0.0153, |
| "num_input_tokens_seen": 41998320, |
| "step": 1640 |
| }, |
| { |
| "epoch": 14.222698072805139, |
| "grad_norm": 1.114964485168457, |
| "learning_rate": 6.733592633536124e-08, |
| "loss": 0.0148, |
| "num_input_tokens_seen": 42253104, |
| "step": 1650 |
| }, |
| { |
| "epoch": 14.308351177730193, |
| "grad_norm": 1.3133609294891357, |
| "learning_rate": 5.3374902871456965e-08, |
| "loss": 0.0151, |
| "num_input_tokens_seen": 42509584, |
| "step": 1660 |
| }, |
| { |
| "epoch": 14.394004282655246, |
| "grad_norm": 1.3046901226043701, |
| "learning_rate": 4.102637467057746e-08, |
| "loss": 0.0144, |
| "num_input_tokens_seen": 42764768, |
| "step": 1670 |
| }, |
| { |
| "epoch": 14.4796573875803, |
| "grad_norm": 1.3270611763000488, |
| "learning_rate": 3.029436709200084e-08, |
| "loss": 0.0142, |
| "num_input_tokens_seen": 43019376, |
| "step": 1680 |
| }, |
| { |
| "epoch": 14.565310492505354, |
| "grad_norm": 1.1487038135528564, |
| "learning_rate": 2.1182378543438408e-08, |
| "loss": 0.0159, |
| "num_input_tokens_seen": 43273248, |
| "step": 1690 |
| }, |
| { |
| "epoch": 14.650963597430406, |
| "grad_norm": 1.1392930746078491, |
| "learning_rate": 1.3693379340626867e-08, |
| "loss": 0.0148, |
| "num_input_tokens_seen": 43529200, |
| "step": 1700 |
| }, |
| { |
| "epoch": 14.73661670235546, |
| "grad_norm": 1.24246084690094, |
| "learning_rate": 7.829810739069521e-09, |
| "loss": 0.0144, |
| "num_input_tokens_seen": 43781760, |
| "step": 1710 |
| }, |
| { |
| "epoch": 14.822269807280513, |
| "grad_norm": 1.2764571905136108, |
| "learning_rate": 3.593584138237294e-09, |
| "loss": 0.0142, |
| "num_input_tokens_seen": 44036144, |
| "step": 1720 |
| }, |
| { |
| "epoch": 14.907922912205567, |
| "grad_norm": 1.4254299402236938, |
| "learning_rate": 9.860804584937988e-10, |
| "loss": 0.0144, |
| "num_input_tokens_seen": 44292256, |
| "step": 1730 |
| }, |
| { |
| "epoch": 14.993576017130621, |
| "grad_norm": 1.1011109352111816, |
| "learning_rate": 8.149690943204391e-12, |
| "loss": 0.014, |
| "num_input_tokens_seen": 44548112, |
| "step": 1740 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 1740, |
| "num_input_tokens_seen": 44548112, |
| "num_train_epochs": 15, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.9454424851110953e+18, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|