{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 3922, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 1.046875, "learning_rate": 5.089058524173028e-07, "loss": 1.4534, "step": 1 }, { "epoch": 0.0, "grad_norm": 0.9609375, "learning_rate": 2.544529262086514e-06, "loss": 1.2906, "step": 5 }, { "epoch": 0.0, "grad_norm": 0.90625, "learning_rate": 5.089058524173028e-06, "loss": 1.253, "step": 10 }, { "epoch": 0.0, "grad_norm": 0.78515625, "learning_rate": 7.633587786259543e-06, "loss": 1.2664, "step": 15 }, { "epoch": 0.01, "grad_norm": 0.59375, "learning_rate": 1.0178117048346055e-05, "loss": 1.1842, "step": 20 }, { "epoch": 0.01, "grad_norm": 0.59765625, "learning_rate": 1.2722646310432571e-05, "loss": 1.1703, "step": 25 }, { "epoch": 0.01, "grad_norm": 0.5625, "learning_rate": 1.5267175572519086e-05, "loss": 1.1021, "step": 30 }, { "epoch": 0.01, "grad_norm": 0.458984375, "learning_rate": 1.78117048346056e-05, "loss": 1.05, "step": 35 }, { "epoch": 0.01, "grad_norm": 0.416015625, "learning_rate": 2.035623409669211e-05, "loss": 0.9612, "step": 40 }, { "epoch": 0.01, "grad_norm": 0.35546875, "learning_rate": 2.2900763358778628e-05, "loss": 0.9693, "step": 45 }, { "epoch": 0.01, "grad_norm": 0.3125, "learning_rate": 2.5445292620865142e-05, "loss": 0.9687, "step": 50 }, { "epoch": 0.01, "grad_norm": 0.341796875, "learning_rate": 2.7989821882951656e-05, "loss": 0.8825, "step": 55 }, { "epoch": 0.02, "grad_norm": 0.3203125, "learning_rate": 3.053435114503817e-05, "loss": 0.8668, "step": 60 }, { "epoch": 0.02, "grad_norm": 0.29296875, "learning_rate": 3.307888040712468e-05, "loss": 0.8335, "step": 65 }, { "epoch": 0.02, "grad_norm": 0.318359375, "learning_rate": 3.56234096692112e-05, "loss": 0.8146, "step": 70 }, { "epoch": 0.02, "grad_norm": 0.3046875, "learning_rate": 3.816793893129771e-05, "loss": 0.8094, "step": 75 }, { "epoch": 0.02, "grad_norm": 0.33984375, "learning_rate": 4.071246819338422e-05, "loss": 0.7349, "step": 80 }, { "epoch": 0.02, "grad_norm": 0.3828125, "learning_rate": 4.325699745547074e-05, "loss": 0.7013, "step": 85 }, { "epoch": 0.02, "grad_norm": 0.375, "learning_rate": 4.5801526717557256e-05, "loss": 0.7324, "step": 90 }, { "epoch": 0.02, "grad_norm": 0.30859375, "learning_rate": 4.8346055979643766e-05, "loss": 0.7071, "step": 95 }, { "epoch": 0.03, "grad_norm": 0.306640625, "learning_rate": 5.0890585241730283e-05, "loss": 0.7077, "step": 100 }, { "epoch": 0.03, "grad_norm": 0.287109375, "learning_rate": 5.3435114503816794e-05, "loss": 0.7136, "step": 105 }, { "epoch": 0.03, "grad_norm": 0.28125, "learning_rate": 5.597964376590331e-05, "loss": 0.6851, "step": 110 }, { "epoch": 0.03, "grad_norm": 0.30078125, "learning_rate": 5.852417302798983e-05, "loss": 0.6655, "step": 115 }, { "epoch": 0.03, "grad_norm": 0.30859375, "learning_rate": 6.106870229007635e-05, "loss": 0.7056, "step": 120 }, { "epoch": 0.03, "grad_norm": 0.271484375, "learning_rate": 6.361323155216285e-05, "loss": 0.6643, "step": 125 }, { "epoch": 0.03, "grad_norm": 0.2890625, "learning_rate": 6.615776081424937e-05, "loss": 0.6525, "step": 130 }, { "epoch": 0.03, "grad_norm": 0.26953125, "learning_rate": 6.870229007633588e-05, "loss": 0.6733, "step": 135 }, { "epoch": 0.04, "grad_norm": 0.287109375, "learning_rate": 7.12468193384224e-05, "loss": 0.6742, "step": 140 }, { "epoch": 0.04, "grad_norm": 0.310546875, "learning_rate": 7.379134860050892e-05, "loss": 0.6985, "step": 145 }, { "epoch": 0.04, "grad_norm": 0.3125, "learning_rate": 7.633587786259542e-05, "loss": 0.61, "step": 150 }, { "epoch": 0.04, "grad_norm": 0.326171875, "learning_rate": 7.888040712468194e-05, "loss": 0.6655, "step": 155 }, { "epoch": 0.04, "grad_norm": 0.271484375, "learning_rate": 8.142493638676844e-05, "loss": 0.6217, "step": 160 }, { "epoch": 0.04, "grad_norm": 0.2890625, "learning_rate": 8.396946564885496e-05, "loss": 0.5982, "step": 165 }, { "epoch": 0.04, "grad_norm": 0.306640625, "learning_rate": 8.651399491094148e-05, "loss": 0.6715, "step": 170 }, { "epoch": 0.04, "grad_norm": 0.314453125, "learning_rate": 8.9058524173028e-05, "loss": 0.6818, "step": 175 }, { "epoch": 0.05, "grad_norm": 0.296875, "learning_rate": 9.160305343511451e-05, "loss": 0.7241, "step": 180 }, { "epoch": 0.05, "grad_norm": 0.302734375, "learning_rate": 9.414758269720102e-05, "loss": 0.6986, "step": 185 }, { "epoch": 0.05, "grad_norm": 0.330078125, "learning_rate": 9.669211195928753e-05, "loss": 0.6512, "step": 190 }, { "epoch": 0.05, "grad_norm": 0.259765625, "learning_rate": 9.923664122137405e-05, "loss": 0.6016, "step": 195 }, { "epoch": 0.05, "grad_norm": 0.2734375, "learning_rate": 0.00010178117048346057, "loss": 0.6165, "step": 200 }, { "epoch": 0.05, "grad_norm": 0.296875, "learning_rate": 0.00010432569974554708, "loss": 0.6832, "step": 205 }, { "epoch": 0.05, "grad_norm": 0.294921875, "learning_rate": 0.00010687022900763359, "loss": 0.6087, "step": 210 }, { "epoch": 0.05, "grad_norm": 0.302734375, "learning_rate": 0.00010941475826972009, "loss": 0.6266, "step": 215 }, { "epoch": 0.06, "grad_norm": 0.310546875, "learning_rate": 0.00011195928753180662, "loss": 0.6328, "step": 220 }, { "epoch": 0.06, "grad_norm": 0.294921875, "learning_rate": 0.00011450381679389313, "loss": 0.6222, "step": 225 }, { "epoch": 0.06, "grad_norm": 0.32421875, "learning_rate": 0.00011704834605597966, "loss": 0.6983, "step": 230 }, { "epoch": 0.06, "grad_norm": 0.318359375, "learning_rate": 0.00011959287531806616, "loss": 0.5978, "step": 235 }, { "epoch": 0.06, "grad_norm": 0.259765625, "learning_rate": 0.0001221374045801527, "loss": 0.6174, "step": 240 }, { "epoch": 0.06, "grad_norm": 0.28515625, "learning_rate": 0.00012468193384223918, "loss": 0.6544, "step": 245 }, { "epoch": 0.06, "grad_norm": 0.314453125, "learning_rate": 0.0001272264631043257, "loss": 0.5861, "step": 250 }, { "epoch": 0.07, "grad_norm": 0.318359375, "learning_rate": 0.00012977099236641222, "loss": 0.6288, "step": 255 }, { "epoch": 0.07, "grad_norm": 0.326171875, "learning_rate": 0.00013231552162849873, "loss": 0.6122, "step": 260 }, { "epoch": 0.07, "grad_norm": 0.294921875, "learning_rate": 0.00013486005089058525, "loss": 0.6029, "step": 265 }, { "epoch": 0.07, "grad_norm": 0.259765625, "learning_rate": 0.00013740458015267177, "loss": 0.5742, "step": 270 }, { "epoch": 0.07, "grad_norm": 0.3359375, "learning_rate": 0.00013994910941475828, "loss": 0.6598, "step": 275 }, { "epoch": 0.07, "grad_norm": 0.31640625, "learning_rate": 0.0001424936386768448, "loss": 0.6627, "step": 280 }, { "epoch": 0.07, "grad_norm": 0.3359375, "learning_rate": 0.0001450381679389313, "loss": 0.6372, "step": 285 }, { "epoch": 0.07, "grad_norm": 0.2578125, "learning_rate": 0.00014758269720101784, "loss": 0.6057, "step": 290 }, { "epoch": 0.08, "grad_norm": 0.263671875, "learning_rate": 0.00015012722646310433, "loss": 0.5944, "step": 295 }, { "epoch": 0.08, "grad_norm": 0.2578125, "learning_rate": 0.00015267175572519084, "loss": 0.5982, "step": 300 }, { "epoch": 0.08, "grad_norm": 0.2890625, "learning_rate": 0.00015521628498727736, "loss": 0.5889, "step": 305 }, { "epoch": 0.08, "grad_norm": 0.255859375, "learning_rate": 0.00015776081424936388, "loss": 0.6026, "step": 310 }, { "epoch": 0.08, "grad_norm": 0.2890625, "learning_rate": 0.0001603053435114504, "loss": 0.5539, "step": 315 }, { "epoch": 0.08, "grad_norm": 0.255859375, "learning_rate": 0.00016284987277353689, "loss": 0.6334, "step": 320 }, { "epoch": 0.08, "grad_norm": 0.310546875, "learning_rate": 0.00016539440203562343, "loss": 0.612, "step": 325 }, { "epoch": 0.08, "grad_norm": 0.212890625, "learning_rate": 0.00016793893129770992, "loss": 0.6216, "step": 330 }, { "epoch": 0.09, "grad_norm": 0.25390625, "learning_rate": 0.00017048346055979644, "loss": 0.6328, "step": 335 }, { "epoch": 0.09, "grad_norm": 0.2412109375, "learning_rate": 0.00017302798982188295, "loss": 0.5892, "step": 340 }, { "epoch": 0.09, "grad_norm": 0.265625, "learning_rate": 0.00017557251908396947, "loss": 0.5984, "step": 345 }, { "epoch": 0.09, "grad_norm": 0.275390625, "learning_rate": 0.000178117048346056, "loss": 0.6217, "step": 350 }, { "epoch": 0.09, "grad_norm": 0.2734375, "learning_rate": 0.0001806615776081425, "loss": 0.5835, "step": 355 }, { "epoch": 0.09, "grad_norm": 0.2890625, "learning_rate": 0.00018320610687022902, "loss": 0.6361, "step": 360 }, { "epoch": 0.09, "grad_norm": 0.28515625, "learning_rate": 0.00018575063613231554, "loss": 0.5791, "step": 365 }, { "epoch": 0.09, "grad_norm": 0.2734375, "learning_rate": 0.00018829516539440203, "loss": 0.6529, "step": 370 }, { "epoch": 0.1, "grad_norm": 0.240234375, "learning_rate": 0.00019083969465648857, "loss": 0.6094, "step": 375 }, { "epoch": 0.1, "grad_norm": 0.23828125, "learning_rate": 0.00019338422391857506, "loss": 0.5728, "step": 380 }, { "epoch": 0.1, "grad_norm": 0.259765625, "learning_rate": 0.00019592875318066158, "loss": 0.6022, "step": 385 }, { "epoch": 0.1, "grad_norm": 0.279296875, "learning_rate": 0.0001984732824427481, "loss": 0.5601, "step": 390 }, { "epoch": 0.1, "grad_norm": 0.27734375, "learning_rate": 0.00019999984150107918, "loss": 0.6034, "step": 395 }, { "epoch": 0.1, "grad_norm": 0.271484375, "learning_rate": 0.00019999805839398995, "loss": 0.6251, "step": 400 }, { "epoch": 0.1, "grad_norm": 0.283203125, "learning_rate": 0.0001999942940916057, "loss": 0.6018, "step": 405 }, { "epoch": 0.1, "grad_norm": 0.197265625, "learning_rate": 0.0001999885486685061, "loss": 0.5812, "step": 410 }, { "epoch": 0.11, "grad_norm": 0.259765625, "learning_rate": 0.00019998082223852138, "loss": 0.6252, "step": 415 }, { "epoch": 0.11, "grad_norm": 0.255859375, "learning_rate": 0.00019997111495473018, "loss": 0.592, "step": 420 }, { "epoch": 0.11, "grad_norm": 0.28125, "learning_rate": 0.0001999594270094565, "loss": 0.5532, "step": 425 }, { "epoch": 0.11, "grad_norm": 0.314453125, "learning_rate": 0.00019994575863426585, "loss": 0.5884, "step": 430 }, { "epoch": 0.11, "grad_norm": 0.267578125, "learning_rate": 0.0001999301100999607, "loss": 0.634, "step": 435 }, { "epoch": 0.11, "grad_norm": 0.251953125, "learning_rate": 0.00019991248171657508, "loss": 0.6146, "step": 440 }, { "epoch": 0.11, "grad_norm": 0.26953125, "learning_rate": 0.00019989287383336853, "loss": 0.5627, "step": 445 }, { "epoch": 0.11, "grad_norm": 0.228515625, "learning_rate": 0.00019987128683881896, "loss": 0.529, "step": 450 }, { "epoch": 0.12, "grad_norm": 0.267578125, "learning_rate": 0.00019984772116061523, "loss": 0.6148, "step": 455 }, { "epoch": 0.12, "grad_norm": 0.255859375, "learning_rate": 0.00019982217726564856, "loss": 0.5908, "step": 460 }, { "epoch": 0.12, "grad_norm": 0.27734375, "learning_rate": 0.00019979465566000317, "loss": 0.5796, "step": 465 }, { "epoch": 0.12, "grad_norm": 0.2109375, "learning_rate": 0.00019976515688894638, "loss": 0.6234, "step": 470 }, { "epoch": 0.12, "grad_norm": 0.2578125, "learning_rate": 0.0001997336815369179, "loss": 0.6051, "step": 475 }, { "epoch": 0.12, "grad_norm": 0.279296875, "learning_rate": 0.00019970023022751793, "loss": 0.5882, "step": 480 }, { "epoch": 0.12, "grad_norm": 0.25390625, "learning_rate": 0.00019966480362349515, "loss": 0.5858, "step": 485 }, { "epoch": 0.12, "grad_norm": 0.2578125, "learning_rate": 0.00019962740242673342, "loss": 0.5626, "step": 490 }, { "epoch": 0.13, "grad_norm": 0.265625, "learning_rate": 0.00019958802737823786, "loss": 0.5921, "step": 495 }, { "epoch": 0.13, "grad_norm": 0.251953125, "learning_rate": 0.0001995466792581202, "loss": 0.5472, "step": 500 }, { "epoch": 0.13, "grad_norm": 0.255859375, "learning_rate": 0.00019950335888558338, "loss": 0.5575, "step": 505 }, { "epoch": 0.13, "grad_norm": 0.255859375, "learning_rate": 0.00019945806711890526, "loss": 0.5166, "step": 510 }, { "epoch": 0.13, "grad_norm": 0.302734375, "learning_rate": 0.00019941080485542156, "loss": 0.6087, "step": 515 }, { "epoch": 0.13, "grad_norm": 0.2734375, "learning_rate": 0.00019936157303150823, "loss": 0.562, "step": 520 }, { "epoch": 0.13, "grad_norm": 0.318359375, "learning_rate": 0.0001993103726225628, "loss": 0.6184, "step": 525 }, { "epoch": 0.14, "grad_norm": 0.25390625, "learning_rate": 0.00019925720464298511, "loss": 0.6244, "step": 530 }, { "epoch": 0.14, "grad_norm": 0.2392578125, "learning_rate": 0.00019920207014615703, "loss": 0.5759, "step": 535 }, { "epoch": 0.14, "grad_norm": 0.30078125, "learning_rate": 0.00019914497022442186, "loss": 0.5807, "step": 540 }, { "epoch": 0.14, "grad_norm": 0.267578125, "learning_rate": 0.0001990859060090625, "loss": 0.5753, "step": 545 }, { "epoch": 0.14, "grad_norm": 0.287109375, "learning_rate": 0.00019902487867027918, "loss": 0.562, "step": 550 }, { "epoch": 0.14, "grad_norm": 0.306640625, "learning_rate": 0.00019896188941716613, "loss": 0.55, "step": 555 }, { "epoch": 0.14, "grad_norm": 0.2490234375, "learning_rate": 0.00019889693949768766, "loss": 0.564, "step": 560 }, { "epoch": 0.14, "grad_norm": 0.2236328125, "learning_rate": 0.00019883003019865357, "loss": 0.5283, "step": 565 }, { "epoch": 0.15, "grad_norm": 0.1982421875, "learning_rate": 0.00019876116284569338, "loss": 0.5591, "step": 570 }, { "epoch": 0.15, "grad_norm": 0.240234375, "learning_rate": 0.00019869033880323044, "loss": 0.5353, "step": 575 }, { "epoch": 0.15, "grad_norm": 0.27734375, "learning_rate": 0.00019861755947445456, "loss": 0.5084, "step": 580 }, { "epoch": 0.15, "grad_norm": 0.3125, "learning_rate": 0.00019854282630129438, "loss": 0.5518, "step": 585 }, { "epoch": 0.15, "grad_norm": 0.23046875, "learning_rate": 0.0001984661407643887, "loss": 0.5609, "step": 590 }, { "epoch": 0.15, "grad_norm": 0.2890625, "learning_rate": 0.00019838750438305733, "loss": 0.5606, "step": 595 }, { "epoch": 0.15, "grad_norm": 0.306640625, "learning_rate": 0.00019830691871527078, "loss": 0.5606, "step": 600 }, { "epoch": 0.15, "grad_norm": 0.234375, "learning_rate": 0.00019822438535761953, "loss": 0.5632, "step": 605 }, { "epoch": 0.16, "grad_norm": 0.283203125, "learning_rate": 0.00019813990594528234, "loss": 0.5923, "step": 610 }, { "epoch": 0.16, "grad_norm": 0.3125, "learning_rate": 0.00019805348215199377, "loss": 0.5813, "step": 615 }, { "epoch": 0.16, "grad_norm": 0.275390625, "learning_rate": 0.0001979651156900113, "loss": 0.5534, "step": 620 }, { "epoch": 0.16, "grad_norm": 0.28515625, "learning_rate": 0.00019787480831008102, "loss": 0.4909, "step": 625 }, { "epoch": 0.16, "grad_norm": 0.234375, "learning_rate": 0.00019778256180140327, "loss": 0.5642, "step": 630 }, { "epoch": 0.16, "grad_norm": 0.271484375, "learning_rate": 0.00019768837799159704, "loss": 0.6093, "step": 635 }, { "epoch": 0.16, "grad_norm": 0.2470703125, "learning_rate": 0.00019759225874666373, "loss": 0.4626, "step": 640 }, { "epoch": 0.16, "grad_norm": 0.259765625, "learning_rate": 0.00019749420597095033, "loss": 0.5873, "step": 645 }, { "epoch": 0.17, "grad_norm": 0.2734375, "learning_rate": 0.00019739422160711147, "loss": 0.5864, "step": 650 }, { "epoch": 0.17, "grad_norm": 0.2890625, "learning_rate": 0.0001972923076360712, "loss": 0.5286, "step": 655 }, { "epoch": 0.17, "grad_norm": 0.359375, "learning_rate": 0.00019718846607698343, "loss": 0.5755, "step": 660 }, { "epoch": 0.17, "grad_norm": 0.2490234375, "learning_rate": 0.00019708269898719225, "loss": 0.53, "step": 665 }, { "epoch": 0.17, "grad_norm": 0.30859375, "learning_rate": 0.00019697500846219094, "loss": 0.5735, "step": 670 }, { "epoch": 0.17, "grad_norm": 0.23828125, "learning_rate": 0.00019686539663558053, "loss": 0.5132, "step": 675 }, { "epoch": 0.17, "grad_norm": 0.271484375, "learning_rate": 0.00019675386567902757, "loss": 0.5416, "step": 680 }, { "epoch": 0.17, "grad_norm": 0.30078125, "learning_rate": 0.00019664041780222103, "loss": 0.6035, "step": 685 }, { "epoch": 0.18, "grad_norm": 0.265625, "learning_rate": 0.00019652505525282848, "loss": 0.5411, "step": 690 }, { "epoch": 0.18, "grad_norm": 0.2578125, "learning_rate": 0.00019640778031645176, "loss": 0.5178, "step": 695 }, { "epoch": 0.18, "grad_norm": 0.287109375, "learning_rate": 0.0001962885953165815, "loss": 0.5137, "step": 700 }, { "epoch": 0.18, "grad_norm": 0.271484375, "learning_rate": 0.0001961675026145511, "loss": 0.5723, "step": 705 }, { "epoch": 0.18, "grad_norm": 0.271484375, "learning_rate": 0.00019604450460949013, "loss": 0.5401, "step": 710 }, { "epoch": 0.18, "grad_norm": 0.26171875, "learning_rate": 0.00019591960373827657, "loss": 0.5215, "step": 715 }, { "epoch": 0.18, "grad_norm": 0.29296875, "learning_rate": 0.00019579280247548865, "loss": 0.5151, "step": 720 }, { "epoch": 0.18, "grad_norm": 0.279296875, "learning_rate": 0.00019566410333335578, "loss": 0.5204, "step": 725 }, { "epoch": 0.19, "grad_norm": 0.2890625, "learning_rate": 0.00019553350886170883, "loss": 0.5436, "step": 730 }, { "epoch": 0.19, "grad_norm": 0.2197265625, "learning_rate": 0.00019540102164792956, "loss": 0.5291, "step": 735 }, { "epoch": 0.19, "grad_norm": 0.279296875, "learning_rate": 0.0001952666443168994, "loss": 0.5917, "step": 740 }, { "epoch": 0.19, "grad_norm": 0.298828125, "learning_rate": 0.0001951303795309474, "loss": 0.5848, "step": 745 }, { "epoch": 0.19, "grad_norm": 0.3046875, "learning_rate": 0.0001949922299897975, "loss": 0.5621, "step": 750 }, { "epoch": 0.19, "grad_norm": 0.20703125, "learning_rate": 0.00019485219843051502, "loss": 0.5872, "step": 755 }, { "epoch": 0.19, "grad_norm": 0.28515625, "learning_rate": 0.00019471028762745252, "loss": 0.5723, "step": 760 }, { "epoch": 0.2, "grad_norm": 0.259765625, "learning_rate": 0.00019456650039219474, "loss": 0.6229, "step": 765 }, { "epoch": 0.2, "grad_norm": 0.310546875, "learning_rate": 0.00019442083957350295, "loss": 0.5566, "step": 770 }, { "epoch": 0.2, "grad_norm": 0.25390625, "learning_rate": 0.00019427330805725843, "loss": 0.4698, "step": 775 }, { "epoch": 0.2, "grad_norm": 0.302734375, "learning_rate": 0.00019412390876640543, "loss": 0.52, "step": 780 }, { "epoch": 0.2, "grad_norm": 0.318359375, "learning_rate": 0.00019397264466089313, "loss": 0.4791, "step": 785 }, { "epoch": 0.2, "grad_norm": 0.359375, "learning_rate": 0.00019381951873761707, "loss": 0.5702, "step": 790 }, { "epoch": 0.2, "grad_norm": 0.259765625, "learning_rate": 0.0001936645340303597, "loss": 0.5061, "step": 795 }, { "epoch": 0.2, "grad_norm": 0.294921875, "learning_rate": 0.00019350769360973038, "loss": 0.5498, "step": 800 }, { "epoch": 0.21, "grad_norm": 0.2470703125, "learning_rate": 0.0001933490005831045, "loss": 0.4755, "step": 805 }, { "epoch": 0.21, "grad_norm": 0.30859375, "learning_rate": 0.00019318845809456186, "loss": 0.5354, "step": 810 }, { "epoch": 0.21, "grad_norm": 0.3125, "learning_rate": 0.00019302606932482448, "loss": 0.5161, "step": 815 }, { "epoch": 0.21, "grad_norm": 0.263671875, "learning_rate": 0.00019286183749119346, "loss": 0.4965, "step": 820 }, { "epoch": 0.21, "grad_norm": 0.24609375, "learning_rate": 0.00019269576584748536, "loss": 0.5485, "step": 825 }, { "epoch": 0.21, "grad_norm": 0.291015625, "learning_rate": 0.0001925278576839676, "loss": 0.4787, "step": 830 }, { "epoch": 0.21, "grad_norm": 0.291015625, "learning_rate": 0.00019235811632729353, "loss": 0.5091, "step": 835 }, { "epoch": 0.21, "grad_norm": 0.267578125, "learning_rate": 0.00019218654514043608, "loss": 0.5315, "step": 840 }, { "epoch": 0.22, "grad_norm": 0.275390625, "learning_rate": 0.0001920131475226216, "loss": 0.5157, "step": 845 }, { "epoch": 0.22, "grad_norm": 0.28515625, "learning_rate": 0.00019183792690926222, "loss": 0.5745, "step": 850 }, { "epoch": 0.22, "grad_norm": 0.33984375, "learning_rate": 0.00019166088677188793, "loss": 0.5265, "step": 855 }, { "epoch": 0.22, "grad_norm": 0.29296875, "learning_rate": 0.00019148203061807766, "loss": 0.5473, "step": 860 }, { "epoch": 0.22, "grad_norm": 0.28125, "learning_rate": 0.00019130136199138994, "loss": 0.4799, "step": 865 }, { "epoch": 0.22, "grad_norm": 0.2734375, "learning_rate": 0.00019111888447129256, "loss": 0.552, "step": 870 }, { "epoch": 0.22, "grad_norm": 0.2412109375, "learning_rate": 0.00019093460167309177, "loss": 0.4779, "step": 875 }, { "epoch": 0.22, "grad_norm": 0.34765625, "learning_rate": 0.00019074851724786064, "loss": 0.5315, "step": 880 }, { "epoch": 0.23, "grad_norm": 0.302734375, "learning_rate": 0.00019056063488236661, "loss": 0.4807, "step": 885 }, { "epoch": 0.23, "grad_norm": 0.322265625, "learning_rate": 0.00019037095829899849, "loss": 0.5849, "step": 890 }, { "epoch": 0.23, "grad_norm": 0.255859375, "learning_rate": 0.00019017949125569285, "loss": 0.5343, "step": 895 }, { "epoch": 0.23, "grad_norm": 0.2578125, "learning_rate": 0.00018998623754585935, "loss": 0.5233, "step": 900 }, { "epoch": 0.23, "grad_norm": 0.2490234375, "learning_rate": 0.00018979120099830578, "loss": 0.4892, "step": 905 }, { "epoch": 0.23, "grad_norm": 0.287109375, "learning_rate": 0.00018959438547716203, "loss": 0.5311, "step": 910 }, { "epoch": 0.23, "grad_norm": 0.2490234375, "learning_rate": 0.0001893957948818037, "loss": 0.5872, "step": 915 }, { "epoch": 0.23, "grad_norm": 0.30078125, "learning_rate": 0.00018919543314677463, "loss": 0.5303, "step": 920 }, { "epoch": 0.24, "grad_norm": 0.224609375, "learning_rate": 0.00018899330424170926, "loss": 0.5294, "step": 925 }, { "epoch": 0.24, "grad_norm": 0.26171875, "learning_rate": 0.00018878941217125367, "loss": 0.5178, "step": 930 }, { "epoch": 0.24, "grad_norm": 0.2451171875, "learning_rate": 0.00018858376097498637, "loss": 0.4638, "step": 935 }, { "epoch": 0.24, "grad_norm": 0.365234375, "learning_rate": 0.00018837635472733832, "loss": 0.5604, "step": 940 }, { "epoch": 0.24, "grad_norm": 0.3125, "learning_rate": 0.00018816719753751213, "loss": 0.5109, "step": 945 }, { "epoch": 0.24, "grad_norm": 0.2578125, "learning_rate": 0.00018795629354940064, "loss": 0.5107, "step": 950 }, { "epoch": 0.24, "grad_norm": 0.310546875, "learning_rate": 0.00018774364694150488, "loss": 0.5133, "step": 955 }, { "epoch": 0.24, "grad_norm": 0.2578125, "learning_rate": 0.00018752926192685125, "loss": 0.4816, "step": 960 }, { "epoch": 0.25, "grad_norm": 0.28125, "learning_rate": 0.000187313142752908, "loss": 0.5037, "step": 965 }, { "epoch": 0.25, "grad_norm": 0.294921875, "learning_rate": 0.00018709529370150124, "loss": 0.5207, "step": 970 }, { "epoch": 0.25, "grad_norm": 0.275390625, "learning_rate": 0.0001868757190887299, "loss": 0.4827, "step": 975 }, { "epoch": 0.25, "grad_norm": 0.361328125, "learning_rate": 0.00018665442326488033, "loss": 0.5168, "step": 980 }, { "epoch": 0.25, "grad_norm": 0.326171875, "learning_rate": 0.0001864314106143401, "loss": 0.5157, "step": 985 }, { "epoch": 0.25, "grad_norm": 0.240234375, "learning_rate": 0.00018620668555551116, "loss": 0.5182, "step": 990 }, { "epoch": 0.25, "grad_norm": 0.314453125, "learning_rate": 0.00018598025254072225, "loss": 0.5141, "step": 995 }, { "epoch": 0.25, "grad_norm": 0.298828125, "learning_rate": 0.00018575211605614066, "loss": 0.4837, "step": 1000 }, { "epoch": 0.26, "grad_norm": 0.330078125, "learning_rate": 0.00018552228062168355, "loss": 0.4921, "step": 1005 }, { "epoch": 0.26, "grad_norm": 0.263671875, "learning_rate": 0.00018529075079092803, "loss": 0.4074, "step": 1010 }, { "epoch": 0.26, "grad_norm": 0.2578125, "learning_rate": 0.0001850575311510214, "loss": 0.5471, "step": 1015 }, { "epoch": 0.26, "grad_norm": 0.375, "learning_rate": 0.00018482262632258975, "loss": 0.552, "step": 1020 }, { "epoch": 0.26, "grad_norm": 0.302734375, "learning_rate": 0.000184586040959647, "loss": 0.4955, "step": 1025 }, { "epoch": 0.26, "grad_norm": 0.31640625, "learning_rate": 0.00018434777974950218, "loss": 0.5257, "step": 1030 }, { "epoch": 0.26, "grad_norm": 0.318359375, "learning_rate": 0.0001841078474126668, "loss": 0.4816, "step": 1035 }, { "epoch": 0.27, "grad_norm": 0.294921875, "learning_rate": 0.00018386624870276138, "loss": 0.474, "step": 1040 }, { "epoch": 0.27, "grad_norm": 0.271484375, "learning_rate": 0.00018362298840642107, "loss": 0.4893, "step": 1045 }, { "epoch": 0.27, "grad_norm": 0.2421875, "learning_rate": 0.00018337807134320103, "loss": 0.4894, "step": 1050 }, { "epoch": 0.27, "grad_norm": 0.30859375, "learning_rate": 0.00018313150236548082, "loss": 0.5137, "step": 1055 }, { "epoch": 0.27, "grad_norm": 0.359375, "learning_rate": 0.00018288328635836826, "loss": 0.4951, "step": 1060 }, { "epoch": 0.27, "grad_norm": 0.28125, "learning_rate": 0.00018263342823960269, "loss": 0.514, "step": 1065 }, { "epoch": 0.27, "grad_norm": 0.36328125, "learning_rate": 0.0001823819329594575, "loss": 0.5667, "step": 1070 }, { "epoch": 0.27, "grad_norm": 0.298828125, "learning_rate": 0.00018212880550064214, "loss": 0.5274, "step": 1075 }, { "epoch": 0.28, "grad_norm": 0.25, "learning_rate": 0.0001818740508782032, "loss": 0.5278, "step": 1080 }, { "epoch": 0.28, "grad_norm": 0.34375, "learning_rate": 0.00018161767413942537, "loss": 0.4962, "step": 1085 }, { "epoch": 0.28, "grad_norm": 0.232421875, "learning_rate": 0.0001813596803637311, "loss": 0.4785, "step": 1090 }, { "epoch": 0.28, "grad_norm": 0.34765625, "learning_rate": 0.00018110007466258017, "loss": 0.5365, "step": 1095 }, { "epoch": 0.28, "grad_norm": 0.298828125, "learning_rate": 0.0001808388621793684, "loss": 0.4773, "step": 1100 }, { "epoch": 0.28, "grad_norm": 0.255859375, "learning_rate": 0.00018057604808932576, "loss": 0.5222, "step": 1105 }, { "epoch": 0.28, "grad_norm": 0.259765625, "learning_rate": 0.00018031163759941362, "loss": 0.5292, "step": 1110 }, { "epoch": 0.28, "grad_norm": 0.326171875, "learning_rate": 0.00018004563594822197, "loss": 0.5277, "step": 1115 }, { "epoch": 0.29, "grad_norm": 0.263671875, "learning_rate": 0.0001797780484058653, "loss": 0.4917, "step": 1120 }, { "epoch": 0.29, "grad_norm": 0.296875, "learning_rate": 0.00017950888027387833, "loss": 0.501, "step": 1125 }, { "epoch": 0.29, "grad_norm": 0.33984375, "learning_rate": 0.000179238136885111, "loss": 0.4991, "step": 1130 }, { "epoch": 0.29, "grad_norm": 0.28125, "learning_rate": 0.00017896582360362273, "loss": 0.4903, "step": 1135 }, { "epoch": 0.29, "grad_norm": 0.251953125, "learning_rate": 0.00017869194582457616, "loss": 0.5216, "step": 1140 }, { "epoch": 0.29, "grad_norm": 0.251953125, "learning_rate": 0.00017841650897413035, "loss": 0.5009, "step": 1145 }, { "epoch": 0.29, "grad_norm": 0.40625, "learning_rate": 0.0001781395185093332, "loss": 0.56, "step": 1150 }, { "epoch": 0.29, "grad_norm": 0.318359375, "learning_rate": 0.00017786097991801328, "loss": 0.5148, "step": 1155 }, { "epoch": 0.3, "grad_norm": 0.283203125, "learning_rate": 0.00017758089871867117, "loss": 0.5133, "step": 1160 }, { "epoch": 0.3, "grad_norm": 0.2890625, "learning_rate": 0.00017729928046037024, "loss": 0.5191, "step": 1165 }, { "epoch": 0.3, "grad_norm": 0.279296875, "learning_rate": 0.00017701613072262644, "loss": 0.4322, "step": 1170 }, { "epoch": 0.3, "grad_norm": 0.3125, "learning_rate": 0.000176731455115298, "loss": 0.5283, "step": 1175 }, { "epoch": 0.3, "grad_norm": 0.392578125, "learning_rate": 0.00017644525927847416, "loss": 0.5175, "step": 1180 }, { "epoch": 0.3, "grad_norm": 0.3125, "learning_rate": 0.00017615754888236347, "loss": 0.4731, "step": 1185 }, { "epoch": 0.3, "grad_norm": 0.314453125, "learning_rate": 0.00017586832962718143, "loss": 0.4347, "step": 1190 }, { "epoch": 0.3, "grad_norm": 0.337890625, "learning_rate": 0.0001755776072430376, "loss": 0.5024, "step": 1195 }, { "epoch": 0.31, "grad_norm": 0.30078125, "learning_rate": 0.00017528538748982198, "loss": 0.4423, "step": 1200 }, { "epoch": 0.31, "grad_norm": 0.287109375, "learning_rate": 0.00017499167615709095, "loss": 0.4574, "step": 1205 }, { "epoch": 0.31, "grad_norm": 0.357421875, "learning_rate": 0.0001746964790639527, "loss": 0.4691, "step": 1210 }, { "epoch": 0.31, "grad_norm": 0.259765625, "learning_rate": 0.00017439980205895158, "loss": 0.4838, "step": 1215 }, { "epoch": 0.31, "grad_norm": 0.30078125, "learning_rate": 0.00017410165101995267, "loss": 0.4605, "step": 1220 }, { "epoch": 0.31, "grad_norm": 0.3046875, "learning_rate": 0.00017380203185402503, "loss": 0.4571, "step": 1225 }, { "epoch": 0.31, "grad_norm": 0.322265625, "learning_rate": 0.00017350095049732466, "loss": 0.5028, "step": 1230 }, { "epoch": 0.31, "grad_norm": 0.263671875, "learning_rate": 0.0001731984129149771, "loss": 0.4862, "step": 1235 }, { "epoch": 0.32, "grad_norm": 0.26953125, "learning_rate": 0.00017289442510095906, "loss": 0.4735, "step": 1240 }, { "epoch": 0.32, "grad_norm": 0.322265625, "learning_rate": 0.00017258899307797976, "loss": 0.4943, "step": 1245 }, { "epoch": 0.32, "grad_norm": 0.29296875, "learning_rate": 0.00017228212289736153, "loss": 0.4617, "step": 1250 }, { "epoch": 0.32, "grad_norm": 0.26953125, "learning_rate": 0.00017197382063892006, "loss": 0.4763, "step": 1255 }, { "epoch": 0.32, "grad_norm": 0.240234375, "learning_rate": 0.00017166409241084374, "loss": 0.4584, "step": 1260 }, { "epoch": 0.32, "grad_norm": 0.298828125, "learning_rate": 0.00017135294434957282, "loss": 0.4857, "step": 1265 }, { "epoch": 0.32, "grad_norm": 0.33203125, "learning_rate": 0.00017104038261967768, "loss": 0.4834, "step": 1270 }, { "epoch": 0.33, "grad_norm": 0.33984375, "learning_rate": 0.00017072641341373692, "loss": 0.5171, "step": 1275 }, { "epoch": 0.33, "grad_norm": 0.392578125, "learning_rate": 0.00017041104295221438, "loss": 0.4695, "step": 1280 }, { "epoch": 0.33, "grad_norm": 0.341796875, "learning_rate": 0.0001700942774833362, "loss": 0.5051, "step": 1285 }, { "epoch": 0.33, "grad_norm": 0.283203125, "learning_rate": 0.00016977612328296674, "loss": 0.4571, "step": 1290 }, { "epoch": 0.33, "grad_norm": 0.2890625, "learning_rate": 0.00016945658665448447, "loss": 0.467, "step": 1295 }, { "epoch": 0.33, "grad_norm": 0.31640625, "learning_rate": 0.00016913567392865697, "loss": 0.5049, "step": 1300 }, { "epoch": 0.33, "grad_norm": 0.255859375, "learning_rate": 0.00016881339146351544, "loss": 0.4694, "step": 1305 }, { "epoch": 0.33, "grad_norm": 0.27734375, "learning_rate": 0.00016848974564422894, "loss": 0.4315, "step": 1310 }, { "epoch": 0.34, "grad_norm": 0.396484375, "learning_rate": 0.00016816474288297773, "loss": 0.4595, "step": 1315 }, { "epoch": 0.34, "grad_norm": 0.3984375, "learning_rate": 0.0001678383896188262, "loss": 0.4894, "step": 1320 }, { "epoch": 0.34, "grad_norm": 0.259765625, "learning_rate": 0.00016751069231759548, "loss": 0.471, "step": 1325 }, { "epoch": 0.34, "grad_norm": 0.33984375, "learning_rate": 0.00016718165747173507, "loss": 0.4694, "step": 1330 }, { "epoch": 0.34, "grad_norm": 0.373046875, "learning_rate": 0.00016685129160019444, "loss": 0.4806, "step": 1335 }, { "epoch": 0.34, "grad_norm": 0.40625, "learning_rate": 0.00016651960124829378, "loss": 0.4655, "step": 1340 }, { "epoch": 0.34, "grad_norm": 0.3359375, "learning_rate": 0.00016618659298759436, "loss": 0.4996, "step": 1345 }, { "epoch": 0.34, "grad_norm": 0.31640625, "learning_rate": 0.00016585227341576824, "loss": 0.4961, "step": 1350 }, { "epoch": 0.35, "grad_norm": 0.27734375, "learning_rate": 0.00016551664915646767, "loss": 0.4674, "step": 1355 }, { "epoch": 0.35, "grad_norm": 0.26171875, "learning_rate": 0.0001651797268591938, "loss": 0.5274, "step": 1360 }, { "epoch": 0.35, "grad_norm": 0.28125, "learning_rate": 0.00016484151319916488, "loss": 0.447, "step": 1365 }, { "epoch": 0.35, "grad_norm": 0.345703125, "learning_rate": 0.00016450201487718422, "loss": 0.4504, "step": 1370 }, { "epoch": 0.35, "grad_norm": 0.333984375, "learning_rate": 0.00016416123861950714, "loss": 0.4656, "step": 1375 }, { "epoch": 0.35, "grad_norm": 0.35546875, "learning_rate": 0.0001638191911777079, "loss": 0.5016, "step": 1380 }, { "epoch": 0.35, "grad_norm": 0.36328125, "learning_rate": 0.000163475879328546, "loss": 0.4829, "step": 1385 }, { "epoch": 0.35, "grad_norm": 0.263671875, "learning_rate": 0.00016313130987383157, "loss": 0.4662, "step": 1390 }, { "epoch": 0.36, "grad_norm": 0.251953125, "learning_rate": 0.00016278548964029108, "loss": 0.4765, "step": 1395 }, { "epoch": 0.36, "grad_norm": 0.34375, "learning_rate": 0.0001624384254794317, "loss": 0.4522, "step": 1400 }, { "epoch": 0.36, "grad_norm": 0.279296875, "learning_rate": 0.00016209012426740582, "loss": 0.4363, "step": 1405 }, { "epoch": 0.36, "grad_norm": 0.369140625, "learning_rate": 0.0001617405929048746, "loss": 0.4888, "step": 1410 }, { "epoch": 0.36, "grad_norm": 0.357421875, "learning_rate": 0.00016138983831687142, "loss": 0.4398, "step": 1415 }, { "epoch": 0.36, "grad_norm": 0.369140625, "learning_rate": 0.0001610378674526646, "loss": 0.5025, "step": 1420 }, { "epoch": 0.36, "grad_norm": 0.34375, "learning_rate": 0.00016068468728561977, "loss": 0.4561, "step": 1425 }, { "epoch": 0.36, "grad_norm": 0.326171875, "learning_rate": 0.0001603303048130616, "loss": 0.4637, "step": 1430 }, { "epoch": 0.37, "grad_norm": 0.35546875, "learning_rate": 0.0001599747270561353, "loss": 0.5011, "step": 1435 }, { "epoch": 0.37, "grad_norm": 0.2578125, "learning_rate": 0.00015961796105966745, "loss": 0.4757, "step": 1440 }, { "epoch": 0.37, "grad_norm": 0.35546875, "learning_rate": 0.00015926001389202638, "loss": 0.4707, "step": 1445 }, { "epoch": 0.37, "grad_norm": 0.35546875, "learning_rate": 0.0001589008926449823, "loss": 0.4305, "step": 1450 }, { "epoch": 0.37, "grad_norm": 0.33984375, "learning_rate": 0.00015854060443356658, "loss": 0.4506, "step": 1455 }, { "epoch": 0.37, "grad_norm": 0.3671875, "learning_rate": 0.0001581791563959309, "loss": 0.4594, "step": 1460 }, { "epoch": 0.37, "grad_norm": 0.30859375, "learning_rate": 0.00015781655569320577, "loss": 0.4565, "step": 1465 }, { "epoch": 0.37, "grad_norm": 0.33203125, "learning_rate": 0.00015745280950935883, "loss": 0.4848, "step": 1470 }, { "epoch": 0.38, "grad_norm": 0.392578125, "learning_rate": 0.00015708792505105223, "loss": 0.5282, "step": 1475 }, { "epoch": 0.38, "grad_norm": 0.37109375, "learning_rate": 0.00015672190954750005, "loss": 0.4261, "step": 1480 }, { "epoch": 0.38, "grad_norm": 0.2734375, "learning_rate": 0.00015635477025032504, "loss": 0.486, "step": 1485 }, { "epoch": 0.38, "grad_norm": 0.314453125, "learning_rate": 0.00015598651443341491, "loss": 0.462, "step": 1490 }, { "epoch": 0.38, "grad_norm": 0.291015625, "learning_rate": 0.00015561714939277822, "loss": 0.4475, "step": 1495 }, { "epoch": 0.38, "grad_norm": 0.44140625, "learning_rate": 0.00015524668244639993, "loss": 0.4765, "step": 1500 }, { "epoch": 0.38, "grad_norm": 0.390625, "learning_rate": 0.00015487512093409616, "loss": 0.4819, "step": 1505 }, { "epoch": 0.39, "grad_norm": 0.3828125, "learning_rate": 0.0001545024722173691, "loss": 0.4338, "step": 1510 }, { "epoch": 0.39, "grad_norm": 0.29296875, "learning_rate": 0.00015412874367926093, "loss": 0.4556, "step": 1515 }, { "epoch": 0.39, "grad_norm": 0.3828125, "learning_rate": 0.00015375394272420763, "loss": 0.4704, "step": 1520 }, { "epoch": 0.39, "grad_norm": 0.3046875, "learning_rate": 0.00015337807677789228, "loss": 0.4999, "step": 1525 }, { "epoch": 0.39, "grad_norm": 0.34375, "learning_rate": 0.0001530011532870979, "loss": 0.4351, "step": 1530 }, { "epoch": 0.39, "grad_norm": 0.2333984375, "learning_rate": 0.00015262317971955993, "loss": 0.444, "step": 1535 }, { "epoch": 0.39, "grad_norm": 0.3828125, "learning_rate": 0.00015224416356381833, "loss": 0.4488, "step": 1540 }, { "epoch": 0.39, "grad_norm": 0.34375, "learning_rate": 0.00015186411232906912, "loss": 0.4704, "step": 1545 }, { "epoch": 0.4, "grad_norm": 0.353515625, "learning_rate": 0.0001514830335450157, "loss": 0.4406, "step": 1550 }, { "epoch": 0.4, "grad_norm": 0.390625, "learning_rate": 0.0001511009347617196, "loss": 0.458, "step": 1555 }, { "epoch": 0.4, "grad_norm": 0.314453125, "learning_rate": 0.00015071782354945094, "loss": 0.4432, "step": 1560 }, { "epoch": 0.4, "grad_norm": 0.330078125, "learning_rate": 0.00015033370749853835, "loss": 0.4713, "step": 1565 }, { "epoch": 0.4, "grad_norm": 0.353515625, "learning_rate": 0.00014994859421921876, "loss": 0.4574, "step": 1570 }, { "epoch": 0.4, "grad_norm": 0.271484375, "learning_rate": 0.00014956249134148646, "loss": 0.4234, "step": 1575 }, { "epoch": 0.4, "grad_norm": 0.287109375, "learning_rate": 0.00014917540651494205, "loss": 0.471, "step": 1580 }, { "epoch": 0.4, "grad_norm": 0.3515625, "learning_rate": 0.0001487873474086408, "loss": 0.4152, "step": 1585 }, { "epoch": 0.41, "grad_norm": 0.30078125, "learning_rate": 0.00014839832171094074, "loss": 0.4751, "step": 1590 }, { "epoch": 0.41, "grad_norm": 0.306640625, "learning_rate": 0.00014800833712935033, "loss": 0.5005, "step": 1595 }, { "epoch": 0.41, "grad_norm": 0.35546875, "learning_rate": 0.00014761740139037578, "loss": 0.4411, "step": 1600 }, { "epoch": 0.41, "grad_norm": 0.37890625, "learning_rate": 0.00014722552223936792, "loss": 0.4463, "step": 1605 }, { "epoch": 0.41, "grad_norm": 0.380859375, "learning_rate": 0.00014683270744036882, "loss": 0.4677, "step": 1610 }, { "epoch": 0.41, "grad_norm": 0.376953125, "learning_rate": 0.00014643896477595792, "loss": 0.4811, "step": 1615 }, { "epoch": 0.41, "grad_norm": 0.443359375, "learning_rate": 0.0001460443020470978, "loss": 0.4741, "step": 1620 }, { "epoch": 0.41, "grad_norm": 0.392578125, "learning_rate": 0.00014564872707297966, "loss": 0.4751, "step": 1625 }, { "epoch": 0.42, "grad_norm": 0.376953125, "learning_rate": 0.00014525224769086853, "loss": 0.4897, "step": 1630 }, { "epoch": 0.42, "grad_norm": 0.341796875, "learning_rate": 0.0001448548717559477, "loss": 0.4843, "step": 1635 }, { "epoch": 0.42, "grad_norm": 0.494140625, "learning_rate": 0.00014445660714116344, "loss": 0.4625, "step": 1640 }, { "epoch": 0.42, "grad_norm": 0.3984375, "learning_rate": 0.0001440574617370687, "loss": 0.4661, "step": 1645 }, { "epoch": 0.42, "grad_norm": 0.265625, "learning_rate": 0.00014365744345166704, "loss": 0.4235, "step": 1650 }, { "epoch": 0.42, "grad_norm": 0.357421875, "learning_rate": 0.00014325656021025573, "loss": 0.4567, "step": 1655 }, { "epoch": 0.42, "grad_norm": 0.353515625, "learning_rate": 0.00014285481995526892, "loss": 0.4371, "step": 1660 }, { "epoch": 0.42, "grad_norm": 0.357421875, "learning_rate": 0.0001424522306461202, "loss": 0.456, "step": 1665 }, { "epoch": 0.43, "grad_norm": 0.328125, "learning_rate": 0.0001420488002590449, "loss": 0.4579, "step": 1670 }, { "epoch": 0.43, "grad_norm": 0.35546875, "learning_rate": 0.000141644536786942, "loss": 0.5085, "step": 1675 }, { "epoch": 0.43, "grad_norm": 0.333984375, "learning_rate": 0.00014123944823921596, "loss": 0.4936, "step": 1680 }, { "epoch": 0.43, "grad_norm": 0.455078125, "learning_rate": 0.00014083354264161782, "loss": 0.4667, "step": 1685 }, { "epoch": 0.43, "grad_norm": 0.341796875, "learning_rate": 0.00014042682803608637, "loss": 0.4848, "step": 1690 }, { "epoch": 0.43, "grad_norm": 0.322265625, "learning_rate": 0.0001400193124805886, "loss": 0.432, "step": 1695 }, { "epoch": 0.43, "grad_norm": 0.4375, "learning_rate": 0.00013961100404896035, "loss": 0.4491, "step": 1700 }, { "epoch": 0.43, "grad_norm": 0.3515625, "learning_rate": 0.00013920191083074612, "loss": 0.4099, "step": 1705 }, { "epoch": 0.44, "grad_norm": 0.408203125, "learning_rate": 0.00013879204093103885, "loss": 0.4092, "step": 1710 }, { "epoch": 0.44, "grad_norm": 0.392578125, "learning_rate": 0.00013838140247031944, "loss": 0.4587, "step": 1715 }, { "epoch": 0.44, "grad_norm": 0.3203125, "learning_rate": 0.00013797000358429563, "loss": 0.4563, "step": 1720 }, { "epoch": 0.44, "grad_norm": 0.416015625, "learning_rate": 0.00013755785242374112, "loss": 0.5002, "step": 1725 }, { "epoch": 0.44, "grad_norm": 0.31640625, "learning_rate": 0.0001371449571543338, "loss": 0.4465, "step": 1730 }, { "epoch": 0.44, "grad_norm": 0.314453125, "learning_rate": 0.00013673132595649418, "loss": 0.4596, "step": 1735 }, { "epoch": 0.44, "grad_norm": 0.33984375, "learning_rate": 0.0001363169670252232, "loss": 0.484, "step": 1740 }, { "epoch": 0.44, "grad_norm": 0.369140625, "learning_rate": 0.00013590188856993992, "loss": 0.4489, "step": 1745 }, { "epoch": 0.45, "grad_norm": 0.359375, "learning_rate": 0.00013548609881431883, "loss": 0.4362, "step": 1750 }, { "epoch": 0.45, "grad_norm": 0.369140625, "learning_rate": 0.0001350696059961269, "loss": 0.4953, "step": 1755 }, { "epoch": 0.45, "grad_norm": 0.36328125, "learning_rate": 0.00013465241836706056, "loss": 0.377, "step": 1760 }, { "epoch": 0.45, "grad_norm": 0.4296875, "learning_rate": 0.0001342345441925819, "loss": 0.3919, "step": 1765 }, { "epoch": 0.45, "grad_norm": 0.291015625, "learning_rate": 0.00013381599175175517, "loss": 0.4668, "step": 1770 }, { "epoch": 0.45, "grad_norm": 0.271484375, "learning_rate": 0.00013339676933708266, "loss": 0.4401, "step": 1775 }, { "epoch": 0.45, "grad_norm": 0.2412109375, "learning_rate": 0.00013297688525434035, "loss": 0.4791, "step": 1780 }, { "epoch": 0.46, "grad_norm": 0.3046875, "learning_rate": 0.00013255634782241349, "loss": 0.3973, "step": 1785 }, { "epoch": 0.46, "grad_norm": 0.404296875, "learning_rate": 0.00013213516537313163, "loss": 0.4692, "step": 1790 }, { "epoch": 0.46, "grad_norm": 0.306640625, "learning_rate": 0.0001317133462511037, "loss": 0.456, "step": 1795 }, { "epoch": 0.46, "grad_norm": 0.35546875, "learning_rate": 0.00013129089881355256, "loss": 0.4319, "step": 1800 }, { "epoch": 0.46, "grad_norm": 0.32421875, "learning_rate": 0.00013086783143014942, "loss": 0.3933, "step": 1805 }, { "epoch": 0.46, "grad_norm": 0.375, "learning_rate": 0.0001304441524828482, "loss": 0.427, "step": 1810 }, { "epoch": 0.46, "grad_norm": 0.3984375, "learning_rate": 0.00013001987036571916, "loss": 0.4683, "step": 1815 }, { "epoch": 0.46, "grad_norm": 0.365234375, "learning_rate": 0.00012959499348478293, "loss": 0.4522, "step": 1820 }, { "epoch": 0.47, "grad_norm": 0.2314453125, "learning_rate": 0.00012916953025784368, "loss": 0.4255, "step": 1825 }, { "epoch": 0.47, "grad_norm": 0.37109375, "learning_rate": 0.00012874348911432251, "loss": 0.4362, "step": 1830 }, { "epoch": 0.47, "grad_norm": 0.37890625, "learning_rate": 0.00012831687849509043, "loss": 0.4532, "step": 1835 }, { "epoch": 0.47, "grad_norm": 0.345703125, "learning_rate": 0.00012788970685230106, "loss": 0.4177, "step": 1840 }, { "epoch": 0.47, "grad_norm": 0.466796875, "learning_rate": 0.00012746198264922319, "loss": 0.4137, "step": 1845 }, { "epoch": 0.47, "grad_norm": 0.35546875, "learning_rate": 0.00012703371436007313, "loss": 0.4552, "step": 1850 }, { "epoch": 0.47, "grad_norm": 0.37890625, "learning_rate": 0.00012660491046984686, "loss": 0.4682, "step": 1855 }, { "epoch": 0.47, "grad_norm": 0.390625, "learning_rate": 0.00012617557947415179, "loss": 0.4947, "step": 1860 }, { "epoch": 0.48, "grad_norm": 0.37109375, "learning_rate": 0.0001257457298790386, "loss": 0.4205, "step": 1865 }, { "epoch": 0.48, "grad_norm": 0.322265625, "learning_rate": 0.00012531537020083258, "loss": 0.4037, "step": 1870 }, { "epoch": 0.48, "grad_norm": 0.3515625, "learning_rate": 0.000124884508965965, "loss": 0.4267, "step": 1875 }, { "epoch": 0.48, "grad_norm": 0.291015625, "learning_rate": 0.00012445315471080402, "loss": 0.3942, "step": 1880 }, { "epoch": 0.48, "grad_norm": 0.314453125, "learning_rate": 0.0001240213159814859, "loss": 0.4453, "step": 1885 }, { "epoch": 0.48, "grad_norm": 0.38671875, "learning_rate": 0.0001235890013337453, "loss": 0.4269, "step": 1890 }, { "epoch": 0.48, "grad_norm": 0.328125, "learning_rate": 0.00012315621933274597, "loss": 0.4226, "step": 1895 }, { "epoch": 0.48, "grad_norm": 0.416015625, "learning_rate": 0.00012272297855291103, "loss": 0.452, "step": 1900 }, { "epoch": 0.49, "grad_norm": 0.3828125, "learning_rate": 0.0001222892875777531, "loss": 0.4421, "step": 1905 }, { "epoch": 0.49, "grad_norm": 0.3984375, "learning_rate": 0.00012185515499970421, "loss": 0.4633, "step": 1910 }, { "epoch": 0.49, "grad_norm": 0.365234375, "learning_rate": 0.00012142058941994556, "loss": 0.4258, "step": 1915 }, { "epoch": 0.49, "grad_norm": 0.318359375, "learning_rate": 0.00012098559944823714, "loss": 0.4136, "step": 1920 }, { "epoch": 0.49, "grad_norm": 0.53125, "learning_rate": 0.00012055019370274714, "loss": 0.3866, "step": 1925 }, { "epoch": 0.49, "grad_norm": 0.482421875, "learning_rate": 0.00012011438080988118, "loss": 0.4743, "step": 1930 }, { "epoch": 0.49, "grad_norm": 0.373046875, "learning_rate": 0.00011967816940411145, "loss": 0.4307, "step": 1935 }, { "epoch": 0.49, "grad_norm": 0.326171875, "learning_rate": 0.00011924156812780558, "loss": 0.4401, "step": 1940 }, { "epoch": 0.5, "grad_norm": 0.3359375, "learning_rate": 0.0001188045856310555, "loss": 0.4417, "step": 1945 }, { "epoch": 0.5, "grad_norm": 0.3984375, "learning_rate": 0.00011836723057150594, "loss": 0.4071, "step": 1950 }, { "epoch": 0.5, "grad_norm": 0.384765625, "learning_rate": 0.00011792951161418301, "loss": 0.4008, "step": 1955 }, { "epoch": 0.5, "grad_norm": 0.46875, "learning_rate": 0.00011749143743132249, "loss": 0.442, "step": 1960 }, { "epoch": 0.5, "grad_norm": 0.31640625, "learning_rate": 0.000117053016702198, "loss": 0.4082, "step": 1965 }, { "epoch": 0.5, "grad_norm": 0.306640625, "learning_rate": 0.00011661425811294902, "loss": 0.4163, "step": 1970 }, { "epoch": 0.5, "grad_norm": 0.423828125, "learning_rate": 0.00011617517035640888, "loss": 0.4438, "step": 1975 }, { "epoch": 0.5, "grad_norm": 0.443359375, "learning_rate": 0.00011573576213193245, "loss": 0.46, "step": 1980 }, { "epoch": 0.51, "grad_norm": 0.33203125, "learning_rate": 0.00011529604214522385, "loss": 0.5053, "step": 1985 }, { "epoch": 0.51, "grad_norm": 0.326171875, "learning_rate": 0.00011485601910816388, "loss": 0.4295, "step": 1990 }, { "epoch": 0.51, "grad_norm": 0.345703125, "learning_rate": 0.00011441570173863756, "loss": 0.4254, "step": 1995 }, { "epoch": 0.51, "grad_norm": 0.341796875, "learning_rate": 0.00011397509876036125, "loss": 0.4208, "step": 2000 }, { "epoch": 0.51, "grad_norm": 0.310546875, "learning_rate": 0.00011353421890270992, "loss": 0.4441, "step": 2005 }, { "epoch": 0.51, "grad_norm": 0.419921875, "learning_rate": 0.00011309307090054415, "loss": 0.4299, "step": 2010 }, { "epoch": 0.51, "grad_norm": 0.380859375, "learning_rate": 0.00011265166349403713, "loss": 0.4684, "step": 2015 }, { "epoch": 0.52, "grad_norm": 0.287109375, "learning_rate": 0.00011221000542850142, "loss": 0.4475, "step": 2020 }, { "epoch": 0.52, "grad_norm": 0.259765625, "learning_rate": 0.00011176810545421572, "loss": 0.4259, "step": 2025 }, { "epoch": 0.52, "grad_norm": 0.287109375, "learning_rate": 0.00011132597232625153, "loss": 0.4661, "step": 2030 }, { "epoch": 0.52, "grad_norm": 0.37890625, "learning_rate": 0.00011088361480429965, "loss": 0.4305, "step": 2035 }, { "epoch": 0.52, "grad_norm": 0.2470703125, "learning_rate": 0.00011044104165249665, "loss": 0.4598, "step": 2040 }, { "epoch": 0.52, "grad_norm": 0.31640625, "learning_rate": 0.00010999826163925127, "loss": 0.4168, "step": 2045 }, { "epoch": 0.52, "grad_norm": 0.29296875, "learning_rate": 0.00010955528353707059, "loss": 0.4045, "step": 2050 }, { "epoch": 0.52, "grad_norm": 0.3984375, "learning_rate": 0.00010911211612238632, "loss": 0.4082, "step": 2055 }, { "epoch": 0.53, "grad_norm": 0.28515625, "learning_rate": 0.00010866876817538097, "loss": 0.4321, "step": 2060 }, { "epoch": 0.53, "grad_norm": 0.28125, "learning_rate": 0.0001082252484798137, "loss": 0.4071, "step": 2065 }, { "epoch": 0.53, "grad_norm": 0.306640625, "learning_rate": 0.00010778156582284649, "loss": 0.4588, "step": 2070 }, { "epoch": 0.53, "grad_norm": 0.234375, "learning_rate": 0.00010733772899486991, "loss": 0.4414, "step": 2075 }, { "epoch": 0.53, "grad_norm": 0.302734375, "learning_rate": 0.00010689374678932906, "loss": 0.393, "step": 2080 }, { "epoch": 0.53, "grad_norm": 0.451171875, "learning_rate": 0.00010644962800254932, "loss": 0.4724, "step": 2085 }, { "epoch": 0.53, "grad_norm": 0.462890625, "learning_rate": 0.00010600538143356201, "loss": 0.4428, "step": 2090 }, { "epoch": 0.53, "grad_norm": 0.3828125, "learning_rate": 0.00010556101588393015, "loss": 0.391, "step": 2095 }, { "epoch": 0.54, "grad_norm": 0.333984375, "learning_rate": 0.00010511654015757407, "loss": 0.3976, "step": 2100 }, { "epoch": 0.54, "grad_norm": 0.3515625, "learning_rate": 0.00010467196306059687, "loss": 0.4012, "step": 2105 }, { "epoch": 0.54, "grad_norm": 0.2890625, "learning_rate": 0.00010422729340111015, "loss": 0.468, "step": 2110 }, { "epoch": 0.54, "grad_norm": 0.37109375, "learning_rate": 0.00010378253998905932, "loss": 0.431, "step": 2115 }, { "epoch": 0.54, "grad_norm": 0.33984375, "learning_rate": 0.00010333771163604911, "loss": 0.4248, "step": 2120 }, { "epoch": 0.54, "grad_norm": 0.337890625, "learning_rate": 0.0001028928171551691, "loss": 0.389, "step": 2125 }, { "epoch": 0.54, "grad_norm": 0.388671875, "learning_rate": 0.0001024478653608189, "loss": 0.418, "step": 2130 }, { "epoch": 0.54, "grad_norm": 0.337890625, "learning_rate": 0.0001020028650685337, "loss": 0.4409, "step": 2135 }, { "epoch": 0.55, "grad_norm": 0.259765625, "learning_rate": 0.00010155782509480954, "loss": 0.4182, "step": 2140 }, { "epoch": 0.55, "grad_norm": 0.3515625, "learning_rate": 0.00010111275425692863, "loss": 0.3616, "step": 2145 }, { "epoch": 0.55, "grad_norm": 0.3671875, "learning_rate": 0.0001006676613727847, "loss": 0.384, "step": 2150 }, { "epoch": 0.55, "grad_norm": 0.255859375, "learning_rate": 0.00010022255526070823, "loss": 0.4133, "step": 2155 }, { "epoch": 0.55, "grad_norm": 0.36328125, "learning_rate": 9.97774447392918e-05, "loss": 0.4635, "step": 2160 }, { "epoch": 0.55, "grad_norm": 0.259765625, "learning_rate": 9.933233862721532e-05, "loss": 0.4389, "step": 2165 }, { "epoch": 0.55, "grad_norm": 0.29296875, "learning_rate": 9.888724574307139e-05, "loss": 0.4488, "step": 2170 }, { "epoch": 0.55, "grad_norm": 0.267578125, "learning_rate": 9.844217490519049e-05, "loss": 0.4116, "step": 2175 }, { "epoch": 0.56, "grad_norm": 0.37109375, "learning_rate": 9.799713493146633e-05, "loss": 0.3794, "step": 2180 }, { "epoch": 0.56, "grad_norm": 0.34375, "learning_rate": 9.755213463918114e-05, "loss": 0.3887, "step": 2185 }, { "epoch": 0.56, "grad_norm": 0.30859375, "learning_rate": 9.710718284483094e-05, "loss": 0.4222, "step": 2190 }, { "epoch": 0.56, "grad_norm": 0.4140625, "learning_rate": 9.66622883639509e-05, "loss": 0.4451, "step": 2195 }, { "epoch": 0.56, "grad_norm": 0.255859375, "learning_rate": 9.621746001094072e-05, "loss": 0.4213, "step": 2200 }, { "epoch": 0.56, "grad_norm": 0.32421875, "learning_rate": 9.577270659888988e-05, "loss": 0.4378, "step": 2205 }, { "epoch": 0.56, "grad_norm": 0.3203125, "learning_rate": 9.532803693940315e-05, "loss": 0.4315, "step": 2210 }, { "epoch": 0.56, "grad_norm": 0.4296875, "learning_rate": 9.488345984242597e-05, "loss": 0.4118, "step": 2215 }, { "epoch": 0.57, "grad_norm": 0.3125, "learning_rate": 9.443898411606988e-05, "loss": 0.4118, "step": 2220 }, { "epoch": 0.57, "grad_norm": 0.3515625, "learning_rate": 9.399461856643802e-05, "loss": 0.5094, "step": 2225 }, { "epoch": 0.57, "grad_norm": 0.291015625, "learning_rate": 9.355037199745072e-05, "loss": 0.391, "step": 2230 }, { "epoch": 0.57, "grad_norm": 0.33203125, "learning_rate": 9.310625321067096e-05, "loss": 0.3856, "step": 2235 }, { "epoch": 0.57, "grad_norm": 0.35546875, "learning_rate": 9.26622710051301e-05, "loss": 0.4223, "step": 2240 }, { "epoch": 0.57, "grad_norm": 0.5078125, "learning_rate": 9.221843417715352e-05, "loss": 0.3672, "step": 2245 }, { "epoch": 0.57, "grad_norm": 0.51171875, "learning_rate": 9.17747515201863e-05, "loss": 0.433, "step": 2250 }, { "epoch": 0.57, "grad_norm": 0.412109375, "learning_rate": 9.133123182461904e-05, "loss": 0.3983, "step": 2255 }, { "epoch": 0.58, "grad_norm": 0.32421875, "learning_rate": 9.088788387761366e-05, "loss": 0.4376, "step": 2260 }, { "epoch": 0.58, "grad_norm": 0.3828125, "learning_rate": 9.044471646292942e-05, "loss": 0.4773, "step": 2265 }, { "epoch": 0.58, "grad_norm": 0.341796875, "learning_rate": 9.000173836074874e-05, "loss": 0.461, "step": 2270 }, { "epoch": 0.58, "grad_norm": 0.3046875, "learning_rate": 8.955895834750334e-05, "loss": 0.4507, "step": 2275 }, { "epoch": 0.58, "grad_norm": 0.8671875, "learning_rate": 8.911638519570036e-05, "loss": 0.3797, "step": 2280 }, { "epoch": 0.58, "grad_norm": 0.365234375, "learning_rate": 8.867402767374848e-05, "loss": 0.3717, "step": 2285 }, { "epoch": 0.58, "grad_norm": 0.4140625, "learning_rate": 8.823189454578428e-05, "loss": 0.3888, "step": 2290 }, { "epoch": 0.59, "grad_norm": 0.3125, "learning_rate": 8.778999457149863e-05, "loss": 0.4496, "step": 2295 }, { "epoch": 0.59, "grad_norm": 0.4296875, "learning_rate": 8.73483365059629e-05, "loss": 0.3736, "step": 2300 }, { "epoch": 0.59, "grad_norm": 0.384765625, "learning_rate": 8.69069290994559e-05, "loss": 0.4593, "step": 2305 }, { "epoch": 0.59, "grad_norm": 0.361328125, "learning_rate": 8.646578109729014e-05, "loss": 0.4097, "step": 2310 }, { "epoch": 0.59, "grad_norm": 0.31640625, "learning_rate": 8.60249012396388e-05, "loss": 0.4032, "step": 2315 }, { "epoch": 0.59, "grad_norm": 0.298828125, "learning_rate": 8.558429826136248e-05, "loss": 0.3813, "step": 2320 }, { "epoch": 0.59, "grad_norm": 0.291015625, "learning_rate": 8.514398089183616e-05, "loss": 0.4135, "step": 2325 }, { "epoch": 0.59, "grad_norm": 0.37109375, "learning_rate": 8.470395785477619e-05, "loss": 0.3782, "step": 2330 }, { "epoch": 0.6, "grad_norm": 0.345703125, "learning_rate": 8.426423786806756e-05, "loss": 0.4147, "step": 2335 }, { "epoch": 0.6, "grad_norm": 0.40234375, "learning_rate": 8.382482964359115e-05, "loss": 0.4261, "step": 2340 }, { "epoch": 0.6, "grad_norm": 0.376953125, "learning_rate": 8.338574188705101e-05, "loss": 0.4219, "step": 2345 }, { "epoch": 0.6, "grad_norm": 0.34375, "learning_rate": 8.294698329780203e-05, "loss": 0.3911, "step": 2350 }, { "epoch": 0.6, "grad_norm": 0.318359375, "learning_rate": 8.250856256867753e-05, "loss": 0.376, "step": 2355 }, { "epoch": 0.6, "grad_norm": 0.306640625, "learning_rate": 8.207048838581701e-05, "loss": 0.3934, "step": 2360 }, { "epoch": 0.6, "grad_norm": 0.33984375, "learning_rate": 8.163276942849408e-05, "loss": 0.3958, "step": 2365 }, { "epoch": 0.6, "grad_norm": 0.37109375, "learning_rate": 8.119541436894452e-05, "loss": 0.396, "step": 2370 }, { "epoch": 0.61, "grad_norm": 0.23828125, "learning_rate": 8.075843187219443e-05, "loss": 0.4779, "step": 2375 }, { "epoch": 0.61, "grad_norm": 0.36328125, "learning_rate": 8.032183059588858e-05, "loss": 0.4172, "step": 2380 }, { "epoch": 0.61, "grad_norm": 0.3828125, "learning_rate": 7.988561919011884e-05, "loss": 0.3797, "step": 2385 }, { "epoch": 0.61, "grad_norm": 0.25390625, "learning_rate": 7.94498062972529e-05, "loss": 0.3962, "step": 2390 }, { "epoch": 0.61, "grad_norm": 0.4140625, "learning_rate": 7.901440055176287e-05, "loss": 0.3965, "step": 2395 }, { "epoch": 0.61, "grad_norm": 0.3359375, "learning_rate": 7.857941058005447e-05, "loss": 0.444, "step": 2400 }, { "epoch": 0.61, "grad_norm": 0.375, "learning_rate": 7.814484500029581e-05, "loss": 0.3988, "step": 2405 }, { "epoch": 0.61, "grad_norm": 0.35546875, "learning_rate": 7.771071242224693e-05, "loss": 0.4253, "step": 2410 }, { "epoch": 0.62, "grad_norm": 0.390625, "learning_rate": 7.727702144708899e-05, "loss": 0.445, "step": 2415 }, { "epoch": 0.62, "grad_norm": 0.265625, "learning_rate": 7.684378066725406e-05, "loss": 0.4292, "step": 2420 }, { "epoch": 0.62, "grad_norm": 0.341796875, "learning_rate": 7.641099866625472e-05, "loss": 0.4002, "step": 2425 }, { "epoch": 0.62, "grad_norm": 0.30859375, "learning_rate": 7.597868401851411e-05, "loss": 0.388, "step": 2430 }, { "epoch": 0.62, "grad_norm": 0.33203125, "learning_rate": 7.554684528919598e-05, "loss": 0.4417, "step": 2435 }, { "epoch": 0.62, "grad_norm": 0.341796875, "learning_rate": 7.511549103403505e-05, "loss": 0.3966, "step": 2440 }, { "epoch": 0.62, "grad_norm": 0.265625, "learning_rate": 7.468462979916744e-05, "loss": 0.3563, "step": 2445 }, { "epoch": 0.62, "grad_norm": 0.423828125, "learning_rate": 7.425427012096142e-05, "loss": 0.3639, "step": 2450 }, { "epoch": 0.63, "grad_norm": 0.291015625, "learning_rate": 7.382442052584821e-05, "loss": 0.4633, "step": 2455 }, { "epoch": 0.63, "grad_norm": 0.361328125, "learning_rate": 7.339508953015316e-05, "loss": 0.3974, "step": 2460 }, { "epoch": 0.63, "grad_norm": 0.353515625, "learning_rate": 7.296628563992689e-05, "loss": 0.4446, "step": 2465 }, { "epoch": 0.63, "grad_norm": 0.37890625, "learning_rate": 7.253801735077684e-05, "loss": 0.3914, "step": 2470 }, { "epoch": 0.63, "grad_norm": 0.337890625, "learning_rate": 7.211029314769897e-05, "loss": 0.4249, "step": 2475 }, { "epoch": 0.63, "grad_norm": 0.36328125, "learning_rate": 7.168312150490959e-05, "loss": 0.41, "step": 2480 }, { "epoch": 0.63, "grad_norm": 0.34765625, "learning_rate": 7.125651088567751e-05, "loss": 0.4528, "step": 2485 }, { "epoch": 0.63, "grad_norm": 0.28515625, "learning_rate": 7.083046974215634e-05, "loss": 0.38, "step": 2490 }, { "epoch": 0.64, "grad_norm": 0.333984375, "learning_rate": 7.040500651521708e-05, "loss": 0.3793, "step": 2495 }, { "epoch": 0.64, "grad_norm": 0.291015625, "learning_rate": 6.998012963428084e-05, "loss": 0.3859, "step": 2500 }, { "epoch": 0.64, "grad_norm": 0.330078125, "learning_rate": 6.95558475171518e-05, "loss": 0.3929, "step": 2505 }, { "epoch": 0.64, "grad_norm": 0.40234375, "learning_rate": 6.913216856985056e-05, "loss": 0.3616, "step": 2510 }, { "epoch": 0.64, "grad_norm": 0.38671875, "learning_rate": 6.870910118644744e-05, "loss": 0.3599, "step": 2515 }, { "epoch": 0.64, "grad_norm": 0.37890625, "learning_rate": 6.828665374889629e-05, "loss": 0.4315, "step": 2520 }, { "epoch": 0.64, "grad_norm": 0.32421875, "learning_rate": 6.786483462686835e-05, "loss": 0.5134, "step": 2525 }, { "epoch": 0.65, "grad_norm": 0.333984375, "learning_rate": 6.744365217758651e-05, "loss": 0.3792, "step": 2530 }, { "epoch": 0.65, "grad_norm": 0.41796875, "learning_rate": 6.702311474565965e-05, "loss": 0.419, "step": 2535 }, { "epoch": 0.65, "grad_norm": 0.314453125, "learning_rate": 6.660323066291735e-05, "loss": 0.4255, "step": 2540 }, { "epoch": 0.65, "grad_norm": 0.3046875, "learning_rate": 6.618400824824482e-05, "loss": 0.4608, "step": 2545 }, { "epoch": 0.65, "grad_norm": 0.30859375, "learning_rate": 6.576545580741811e-05, "loss": 0.4734, "step": 2550 }, { "epoch": 0.65, "grad_norm": 0.44921875, "learning_rate": 6.534758163293944e-05, "loss": 0.4022, "step": 2555 }, { "epoch": 0.65, "grad_norm": 0.375, "learning_rate": 6.493039400387312e-05, "loss": 0.393, "step": 2560 }, { "epoch": 0.65, "grad_norm": 0.39453125, "learning_rate": 6.451390118568124e-05, "loss": 0.4285, "step": 2565 }, { "epoch": 0.66, "grad_norm": 0.37890625, "learning_rate": 6.409811143006012e-05, "loss": 0.3768, "step": 2570 }, { "epoch": 0.66, "grad_norm": 0.33203125, "learning_rate": 6.368303297477685e-05, "loss": 0.4258, "step": 2575 }, { "epoch": 0.66, "grad_norm": 0.3671875, "learning_rate": 6.326867404350587e-05, "loss": 0.4105, "step": 2580 }, { "epoch": 0.66, "grad_norm": 0.47265625, "learning_rate": 6.285504284566623e-05, "loss": 0.3375, "step": 2585 }, { "epoch": 0.66, "grad_norm": 0.357421875, "learning_rate": 6.244214757625891e-05, "loss": 0.4004, "step": 2590 }, { "epoch": 0.66, "grad_norm": 0.388671875, "learning_rate": 6.202999641570438e-05, "loss": 0.4723, "step": 2595 }, { "epoch": 0.66, "grad_norm": 0.375, "learning_rate": 6.161859752968059e-05, "loss": 0.3726, "step": 2600 }, { "epoch": 0.66, "grad_norm": 0.31640625, "learning_rate": 6.120795906896116e-05, "loss": 0.3782, "step": 2605 }, { "epoch": 0.67, "grad_norm": 0.376953125, "learning_rate": 6.079808916925391e-05, "loss": 0.3922, "step": 2610 }, { "epoch": 0.67, "grad_norm": 0.33203125, "learning_rate": 6.038899595103968e-05, "loss": 0.3924, "step": 2615 }, { "epoch": 0.67, "grad_norm": 0.388671875, "learning_rate": 5.9980687519411436e-05, "loss": 0.4182, "step": 2620 }, { "epoch": 0.67, "grad_norm": 0.283203125, "learning_rate": 5.95731719639137e-05, "loss": 0.483, "step": 2625 }, { "epoch": 0.67, "grad_norm": 0.373046875, "learning_rate": 5.91664573583822e-05, "loss": 0.4156, "step": 2630 }, { "epoch": 0.67, "grad_norm": 0.35546875, "learning_rate": 5.876055176078407e-05, "loss": 0.4438, "step": 2635 }, { "epoch": 0.67, "grad_norm": 0.298828125, "learning_rate": 5.8355463213058025e-05, "loss": 0.4392, "step": 2640 }, { "epoch": 0.67, "grad_norm": 0.33984375, "learning_rate": 5.795119974095514e-05, "loss": 0.4561, "step": 2645 }, { "epoch": 0.68, "grad_norm": 0.302734375, "learning_rate": 5.754776935387982e-05, "loss": 0.3761, "step": 2650 }, { "epoch": 0.68, "grad_norm": 0.388671875, "learning_rate": 5.714518004473111e-05, "loss": 0.3753, "step": 2655 }, { "epoch": 0.68, "grad_norm": 0.345703125, "learning_rate": 5.674343978974431e-05, "loss": 0.3991, "step": 2660 }, { "epoch": 0.68, "grad_norm": 0.326171875, "learning_rate": 5.634255654833297e-05, "loss": 0.3749, "step": 2665 }, { "epoch": 0.68, "grad_norm": 0.48828125, "learning_rate": 5.594253826293129e-05, "loss": 0.3925, "step": 2670 }, { "epoch": 0.68, "grad_norm": 0.33984375, "learning_rate": 5.554339285883656e-05, "loss": 0.3903, "step": 2675 }, { "epoch": 0.68, "grad_norm": 0.48046875, "learning_rate": 5.514512824405228e-05, "loss": 0.3997, "step": 2680 }, { "epoch": 0.68, "grad_norm": 0.42578125, "learning_rate": 5.474775230913147e-05, "loss": 0.4459, "step": 2685 }, { "epoch": 0.69, "grad_norm": 0.453125, "learning_rate": 5.435127292702033e-05, "loss": 0.3924, "step": 2690 }, { "epoch": 0.69, "grad_norm": 0.455078125, "learning_rate": 5.395569795290221e-05, "loss": 0.4301, "step": 2695 }, { "epoch": 0.69, "grad_norm": 0.314453125, "learning_rate": 5.356103522404208e-05, "loss": 0.4262, "step": 2700 }, { "epoch": 0.69, "grad_norm": 0.4140625, "learning_rate": 5.316729255963118e-05, "loss": 0.4261, "step": 2705 }, { "epoch": 0.69, "grad_norm": 0.3984375, "learning_rate": 5.277447776063208e-05, "loss": 0.3991, "step": 2710 }, { "epoch": 0.69, "grad_norm": 0.384765625, "learning_rate": 5.238259860962423e-05, "loss": 0.3679, "step": 2715 }, { "epoch": 0.69, "grad_norm": 0.369140625, "learning_rate": 5.199166287064967e-05, "loss": 0.3751, "step": 2720 }, { "epoch": 0.69, "grad_norm": 0.302734375, "learning_rate": 5.160167828905926e-05, "loss": 0.4476, "step": 2725 }, { "epoch": 0.7, "grad_norm": 0.359375, "learning_rate": 5.121265259135921e-05, "loss": 0.3787, "step": 2730 }, { "epoch": 0.7, "grad_norm": 0.400390625, "learning_rate": 5.0824593485057946e-05, "loss": 0.362, "step": 2735 }, { "epoch": 0.7, "grad_norm": 0.38671875, "learning_rate": 5.043750865851354e-05, "loss": 0.3852, "step": 2740 }, { "epoch": 0.7, "grad_norm": 0.421875, "learning_rate": 5.005140578078127e-05, "loss": 0.3985, "step": 2745 }, { "epoch": 0.7, "grad_norm": 0.353515625, "learning_rate": 4.966629250146167e-05, "loss": 0.3807, "step": 2750 }, { "epoch": 0.7, "grad_norm": 0.279296875, "learning_rate": 4.928217645054909e-05, "loss": 0.3646, "step": 2755 }, { "epoch": 0.7, "grad_norm": 0.287109375, "learning_rate": 4.889906523828041e-05, "loss": 0.3996, "step": 2760 }, { "epoch": 0.7, "grad_norm": 0.3671875, "learning_rate": 4.851696645498429e-05, "loss": 0.4218, "step": 2765 }, { "epoch": 0.71, "grad_norm": 0.41015625, "learning_rate": 4.813588767093088e-05, "loss": 0.4394, "step": 2770 }, { "epoch": 0.71, "grad_norm": 0.2890625, "learning_rate": 4.7755836436181676e-05, "loss": 0.3549, "step": 2775 }, { "epoch": 0.71, "grad_norm": 0.384765625, "learning_rate": 4.737682028044007e-05, "loss": 0.4035, "step": 2780 }, { "epoch": 0.71, "grad_norm": 0.322265625, "learning_rate": 4.699884671290211e-05, "loss": 0.3727, "step": 2785 }, { "epoch": 0.71, "grad_norm": 0.330078125, "learning_rate": 4.662192322210771e-05, "loss": 0.3926, "step": 2790 }, { "epoch": 0.71, "grad_norm": 0.439453125, "learning_rate": 4.6246057275792356e-05, "loss": 0.3738, "step": 2795 }, { "epoch": 0.71, "grad_norm": 0.32421875, "learning_rate": 4.587125632073906e-05, "loss": 0.4018, "step": 2800 }, { "epoch": 0.72, "grad_norm": 0.416015625, "learning_rate": 4.54975277826309e-05, "loss": 0.4074, "step": 2805 }, { "epoch": 0.72, "grad_norm": 0.3046875, "learning_rate": 4.512487906590385e-05, "loss": 0.4087, "step": 2810 }, { "epoch": 0.72, "grad_norm": 0.349609375, "learning_rate": 4.4753317553600084e-05, "loss": 0.3621, "step": 2815 }, { "epoch": 0.72, "grad_norm": 0.46484375, "learning_rate": 4.438285060722176e-05, "loss": 0.4324, "step": 2820 }, { "epoch": 0.72, "grad_norm": 0.37890625, "learning_rate": 4.401348556658509e-05, "loss": 0.4352, "step": 2825 }, { "epoch": 0.72, "grad_norm": 0.3046875, "learning_rate": 4.3645229749674987e-05, "loss": 0.4221, "step": 2830 }, { "epoch": 0.72, "grad_norm": 0.5078125, "learning_rate": 4.3278090452499984e-05, "loss": 0.3938, "step": 2835 }, { "epoch": 0.72, "grad_norm": 0.40625, "learning_rate": 4.291207494894781e-05, "loss": 0.4412, "step": 2840 }, { "epoch": 0.73, "grad_norm": 0.34375, "learning_rate": 4.254719049064121e-05, "loss": 0.3785, "step": 2845 }, { "epoch": 0.73, "grad_norm": 0.298828125, "learning_rate": 4.218344430679426e-05, "loss": 0.4147, "step": 2850 }, { "epoch": 0.73, "grad_norm": 0.435546875, "learning_rate": 4.182084360406917e-05, "loss": 0.3998, "step": 2855 }, { "epoch": 0.73, "grad_norm": 0.33984375, "learning_rate": 4.145939556643348e-05, "loss": 0.3945, "step": 2860 }, { "epoch": 0.73, "grad_norm": 0.421875, "learning_rate": 4.109910735501774e-05, "loss": 0.422, "step": 2865 }, { "epoch": 0.73, "grad_norm": 0.3359375, "learning_rate": 4.073998610797366e-05, "loss": 0.4348, "step": 2870 }, { "epoch": 0.73, "grad_norm": 0.2734375, "learning_rate": 4.0382038940332615e-05, "loss": 0.408, "step": 2875 }, { "epoch": 0.73, "grad_norm": 0.3125, "learning_rate": 4.002527294386476e-05, "loss": 0.3963, "step": 2880 }, { "epoch": 0.74, "grad_norm": 0.357421875, "learning_rate": 3.966969518693845e-05, "loss": 0.3804, "step": 2885 }, { "epoch": 0.74, "grad_norm": 0.416015625, "learning_rate": 3.931531271438027e-05, "loss": 0.4101, "step": 2890 }, { "epoch": 0.74, "grad_norm": 0.380859375, "learning_rate": 3.896213254733543e-05, "loss": 0.3588, "step": 2895 }, { "epoch": 0.74, "grad_norm": 0.271484375, "learning_rate": 3.861016168312861e-05, "loss": 0.3458, "step": 2900 }, { "epoch": 0.74, "grad_norm": 0.333984375, "learning_rate": 3.8259407095125446e-05, "loss": 0.4511, "step": 2905 }, { "epoch": 0.74, "grad_norm": 0.40234375, "learning_rate": 3.790987573259422e-05, "loss": 0.44, "step": 2910 }, { "epoch": 0.74, "grad_norm": 0.37890625, "learning_rate": 3.756157452056832e-05, "loss": 0.4312, "step": 2915 }, { "epoch": 0.74, "grad_norm": 0.306640625, "learning_rate": 3.7214510359708934e-05, "loss": 0.4258, "step": 2920 }, { "epoch": 0.75, "grad_norm": 0.44140625, "learning_rate": 3.6868690126168425e-05, "loss": 0.4092, "step": 2925 }, { "epoch": 0.75, "grad_norm": 0.2890625, "learning_rate": 3.652412067145402e-05, "loss": 0.4247, "step": 2930 }, { "epoch": 0.75, "grad_norm": 0.416015625, "learning_rate": 3.6180808822292077e-05, "loss": 0.3827, "step": 2935 }, { "epoch": 0.75, "grad_norm": 0.35546875, "learning_rate": 3.583876138049287e-05, "loss": 0.4062, "step": 2940 }, { "epoch": 0.75, "grad_norm": 0.306640625, "learning_rate": 3.549798512281579e-05, "loss": 0.392, "step": 2945 }, { "epoch": 0.75, "grad_norm": 0.380859375, "learning_rate": 3.5158486800835114e-05, "loss": 0.3869, "step": 2950 }, { "epoch": 0.75, "grad_norm": 0.328125, "learning_rate": 3.482027314080621e-05, "loss": 0.4215, "step": 2955 }, { "epoch": 0.75, "grad_norm": 0.3515625, "learning_rate": 3.4483350843532335e-05, "loss": 0.3947, "step": 2960 }, { "epoch": 0.76, "grad_norm": 0.423828125, "learning_rate": 3.414772658423176e-05, "loss": 0.4174, "step": 2965 }, { "epoch": 0.76, "grad_norm": 0.31640625, "learning_rate": 3.381340701240564e-05, "loss": 0.4834, "step": 2970 }, { "epoch": 0.76, "grad_norm": 0.3828125, "learning_rate": 3.348039875170622e-05, "loss": 0.3834, "step": 2975 }, { "epoch": 0.76, "grad_norm": 0.416015625, "learning_rate": 3.3148708399805574e-05, "loss": 0.4661, "step": 2980 }, { "epoch": 0.76, "grad_norm": 0.333984375, "learning_rate": 3.2818342528264944e-05, "loss": 0.3878, "step": 2985 }, { "epoch": 0.76, "grad_norm": 0.375, "learning_rate": 3.248930768240454e-05, "loss": 0.4531, "step": 2990 }, { "epoch": 0.76, "grad_norm": 0.345703125, "learning_rate": 3.216161038117379e-05, "loss": 0.388, "step": 2995 }, { "epoch": 0.76, "grad_norm": 0.396484375, "learning_rate": 3.1835257117022276e-05, "loss": 0.4092, "step": 3000 }, { "epoch": 0.77, "grad_norm": 0.431640625, "learning_rate": 3.151025435577106e-05, "loss": 0.4103, "step": 3005 }, { "epoch": 0.77, "grad_norm": 0.3125, "learning_rate": 3.118660853648457e-05, "loss": 0.4309, "step": 3010 }, { "epoch": 0.77, "grad_norm": 0.490234375, "learning_rate": 3.0864326071343064e-05, "loss": 0.361, "step": 3015 }, { "epoch": 0.77, "grad_norm": 0.322265625, "learning_rate": 3.0543413345515526e-05, "loss": 0.3627, "step": 3020 }, { "epoch": 0.77, "grad_norm": 0.37109375, "learning_rate": 3.022387671703325e-05, "loss": 0.444, "step": 3025 }, { "epoch": 0.77, "grad_norm": 0.322265625, "learning_rate": 2.9905722516663803e-05, "loss": 0.44, "step": 3030 }, { "epoch": 0.77, "grad_norm": 0.314453125, "learning_rate": 2.9588957047785614e-05, "loss": 0.3949, "step": 3035 }, { "epoch": 0.78, "grad_norm": 0.265625, "learning_rate": 2.9273586586263103e-05, "loss": 0.4309, "step": 3040 }, { "epoch": 0.78, "grad_norm": 0.3671875, "learning_rate": 2.8959617380322336e-05, "loss": 0.4052, "step": 3045 }, { "epoch": 0.78, "grad_norm": 0.38671875, "learning_rate": 2.8647055650427214e-05, "loss": 0.4693, "step": 3050 }, { "epoch": 0.78, "grad_norm": 0.2578125, "learning_rate": 2.8335907589156264e-05, "loss": 0.3854, "step": 3055 }, { "epoch": 0.78, "grad_norm": 0.4140625, "learning_rate": 2.802617936107993e-05, "loss": 0.3751, "step": 3060 }, { "epoch": 0.78, "grad_norm": 0.302734375, "learning_rate": 2.7717877102638446e-05, "loss": 0.3518, "step": 3065 }, { "epoch": 0.78, "grad_norm": 0.353515625, "learning_rate": 2.7411006922020244e-05, "loss": 0.3935, "step": 3070 }, { "epoch": 0.78, "grad_norm": 0.40625, "learning_rate": 2.7105574899040942e-05, "loss": 0.3661, "step": 3075 }, { "epoch": 0.79, "grad_norm": 0.3359375, "learning_rate": 2.6801587085022906e-05, "loss": 0.3446, "step": 3080 }, { "epoch": 0.79, "grad_norm": 0.404296875, "learning_rate": 2.6499049502675344e-05, "loss": 0.4103, "step": 3085 }, { "epoch": 0.79, "grad_norm": 0.32421875, "learning_rate": 2.6197968145974993e-05, "loss": 0.4104, "step": 3090 }, { "epoch": 0.79, "grad_norm": 0.36328125, "learning_rate": 2.589834898004735e-05, "loss": 0.3658, "step": 3095 }, { "epoch": 0.79, "grad_norm": 0.392578125, "learning_rate": 2.5600197941048442e-05, "loss": 0.4091, "step": 3100 }, { "epoch": 0.79, "grad_norm": 0.26171875, "learning_rate": 2.5303520936047366e-05, "loss": 0.4013, "step": 3105 }, { "epoch": 0.79, "grad_norm": 0.3828125, "learning_rate": 2.5008323842909075e-05, "loss": 0.3581, "step": 3110 }, { "epoch": 0.79, "grad_norm": 0.296875, "learning_rate": 2.471461251017808e-05, "loss": 0.3613, "step": 3115 }, { "epoch": 0.8, "grad_norm": 0.4375, "learning_rate": 2.4422392756962452e-05, "loss": 0.3939, "step": 3120 }, { "epoch": 0.8, "grad_norm": 0.42578125, "learning_rate": 2.4131670372818603e-05, "loss": 0.402, "step": 3125 }, { "epoch": 0.8, "grad_norm": 0.34765625, "learning_rate": 2.3842451117636566e-05, "loss": 0.4436, "step": 3130 }, { "epoch": 0.8, "grad_norm": 0.31640625, "learning_rate": 2.355474072152588e-05, "loss": 0.3995, "step": 3135 }, { "epoch": 0.8, "grad_norm": 0.291015625, "learning_rate": 2.3268544884702037e-05, "loss": 0.3673, "step": 3140 }, { "epoch": 0.8, "grad_norm": 0.345703125, "learning_rate": 2.2983869277373604e-05, "loss": 0.3658, "step": 3145 }, { "epoch": 0.8, "grad_norm": 0.330078125, "learning_rate": 2.270071953962981e-05, "loss": 0.3789, "step": 3150 }, { "epoch": 0.8, "grad_norm": 0.34375, "learning_rate": 2.2419101281328857e-05, "loss": 0.442, "step": 3155 }, { "epoch": 0.81, "grad_norm": 0.45703125, "learning_rate": 2.2139020081986783e-05, "loss": 0.3913, "step": 3160 }, { "epoch": 0.81, "grad_norm": 0.4140625, "learning_rate": 2.186048149066684e-05, "loss": 0.3532, "step": 3165 }, { "epoch": 0.81, "grad_norm": 0.255859375, "learning_rate": 2.158349102586964e-05, "loss": 0.4629, "step": 3170 }, { "epoch": 0.81, "grad_norm": 0.326171875, "learning_rate": 2.130805417542384e-05, "loss": 0.3224, "step": 3175 }, { "epoch": 0.81, "grad_norm": 0.326171875, "learning_rate": 2.103417639637729e-05, "loss": 0.4609, "step": 3180 }, { "epoch": 0.81, "grad_norm": 0.3515625, "learning_rate": 2.0761863114889002e-05, "loss": 0.4257, "step": 3185 }, { "epoch": 0.81, "grad_norm": 0.4140625, "learning_rate": 2.0491119726121667e-05, "loss": 0.4391, "step": 3190 }, { "epoch": 0.81, "grad_norm": 0.3359375, "learning_rate": 2.0221951594134713e-05, "loss": 0.3992, "step": 3195 }, { "epoch": 0.82, "grad_norm": 0.353515625, "learning_rate": 1.995436405177804e-05, "loss": 0.3755, "step": 3200 }, { "epoch": 0.82, "grad_norm": 0.345703125, "learning_rate": 1.968836240058638e-05, "loss": 0.3586, "step": 3205 }, { "epoch": 0.82, "grad_norm": 0.4296875, "learning_rate": 1.9423951910674265e-05, "loss": 0.4388, "step": 3210 }, { "epoch": 0.82, "grad_norm": 0.37890625, "learning_rate": 1.916113782063159e-05, "loss": 0.422, "step": 3215 }, { "epoch": 0.82, "grad_norm": 0.35546875, "learning_rate": 1.8899925337419844e-05, "loss": 0.4077, "step": 3220 }, { "epoch": 0.82, "grad_norm": 0.3125, "learning_rate": 1.864031963626893e-05, "loss": 0.3461, "step": 3225 }, { "epoch": 0.82, "grad_norm": 0.384765625, "learning_rate": 1.8382325860574656e-05, "loss": 0.3394, "step": 3230 }, { "epoch": 0.82, "grad_norm": 0.375, "learning_rate": 1.8125949121796805e-05, "loss": 0.3918, "step": 3235 }, { "epoch": 0.83, "grad_norm": 0.40625, "learning_rate": 1.787119449935789e-05, "loss": 0.4278, "step": 3240 }, { "epoch": 0.83, "grad_norm": 0.41796875, "learning_rate": 1.7618067040542507e-05, "loss": 0.4128, "step": 3245 }, { "epoch": 0.83, "grad_norm": 0.326171875, "learning_rate": 1.736657176039732e-05, "loss": 0.3997, "step": 3250 }, { "epoch": 0.83, "grad_norm": 0.291015625, "learning_rate": 1.7116713641631744e-05, "loss": 0.3856, "step": 3255 }, { "epoch": 0.83, "grad_norm": 0.328125, "learning_rate": 1.6868497634519185e-05, "loss": 0.3838, "step": 3260 }, { "epoch": 0.83, "grad_norm": 0.357421875, "learning_rate": 1.6621928656798967e-05, "loss": 0.4358, "step": 3265 }, { "epoch": 0.83, "grad_norm": 0.47265625, "learning_rate": 1.637701159357895e-05, "loss": 0.4258, "step": 3270 }, { "epoch": 0.84, "grad_norm": 0.2734375, "learning_rate": 1.6133751297238654e-05, "loss": 0.4668, "step": 3275 }, { "epoch": 0.84, "grad_norm": 0.31640625, "learning_rate": 1.589215258733322e-05, "loss": 0.4056, "step": 3280 }, { "epoch": 0.84, "grad_norm": 0.310546875, "learning_rate": 1.565222025049785e-05, "loss": 0.3928, "step": 3285 }, { "epoch": 0.84, "grad_norm": 0.30078125, "learning_rate": 1.541395904035301e-05, "loss": 0.3993, "step": 3290 }, { "epoch": 0.84, "grad_norm": 0.412109375, "learning_rate": 1.5177373677410245e-05, "loss": 0.4256, "step": 3295 }, { "epoch": 0.84, "grad_norm": 0.447265625, "learning_rate": 1.4942468848978642e-05, "loss": 0.3931, "step": 3300 }, { "epoch": 0.84, "grad_norm": 0.42578125, "learning_rate": 1.4709249209071974e-05, "loss": 0.4088, "step": 3305 }, { "epoch": 0.84, "grad_norm": 0.423828125, "learning_rate": 1.4477719378316469e-05, "loss": 0.376, "step": 3310 }, { "epoch": 0.85, "grad_norm": 0.5, "learning_rate": 1.424788394385933e-05, "loss": 0.4155, "step": 3315 }, { "epoch": 0.85, "grad_norm": 0.33203125, "learning_rate": 1.4019747459277777e-05, "loss": 0.4336, "step": 3320 }, { "epoch": 0.85, "grad_norm": 0.427734375, "learning_rate": 1.379331444448886e-05, "loss": 0.4389, "step": 3325 }, { "epoch": 0.85, "grad_norm": 0.369140625, "learning_rate": 1.3568589385659914e-05, "loss": 0.3889, "step": 3330 }, { "epoch": 0.85, "grad_norm": 0.369140625, "learning_rate": 1.334557673511969e-05, "loss": 0.3689, "step": 3335 }, { "epoch": 0.85, "grad_norm": 0.27734375, "learning_rate": 1.3124280911270114e-05, "loss": 0.4072, "step": 3340 }, { "epoch": 0.85, "grad_norm": 0.29296875, "learning_rate": 1.290470629849876e-05, "loss": 0.3905, "step": 3345 }, { "epoch": 0.85, "grad_norm": 0.337890625, "learning_rate": 1.2686857247091987e-05, "loss": 0.3706, "step": 3350 }, { "epoch": 0.86, "grad_norm": 0.4609375, "learning_rate": 1.2470738073148769e-05, "loss": 0.3561, "step": 3355 }, { "epoch": 0.86, "grad_norm": 0.37109375, "learning_rate": 1.225635305849513e-05, "loss": 0.3786, "step": 3360 }, { "epoch": 0.86, "grad_norm": 0.416015625, "learning_rate": 1.2043706450599391e-05, "loss": 0.3541, "step": 3365 }, { "epoch": 0.86, "grad_norm": 0.349609375, "learning_rate": 1.1832802462487902e-05, "loss": 0.3837, "step": 3370 }, { "epoch": 0.86, "grad_norm": 0.359375, "learning_rate": 1.1623645272661698e-05, "loss": 0.3629, "step": 3375 }, { "epoch": 0.86, "grad_norm": 0.265625, "learning_rate": 1.1416239025013653e-05, "loss": 0.371, "step": 3380 }, { "epoch": 0.86, "grad_norm": 0.353515625, "learning_rate": 1.1210587828746355e-05, "loss": 0.3975, "step": 3385 }, { "epoch": 0.86, "grad_norm": 0.296875, "learning_rate": 1.1006695758290753e-05, "loss": 0.3682, "step": 3390 }, { "epoch": 0.87, "grad_norm": 0.3046875, "learning_rate": 1.0804566853225383e-05, "loss": 0.4048, "step": 3395 }, { "epoch": 0.87, "grad_norm": 0.369140625, "learning_rate": 1.0604205118196342e-05, "loss": 0.3566, "step": 3400 }, { "epoch": 0.87, "grad_norm": 0.42578125, "learning_rate": 1.0405614522837992e-05, "loss": 0.3956, "step": 3405 }, { "epoch": 0.87, "grad_norm": 0.34765625, "learning_rate": 1.0208799001694247e-05, "loss": 0.4019, "step": 3410 }, { "epoch": 0.87, "grad_norm": 0.421875, "learning_rate": 1.0013762454140663e-05, "loss": 0.4616, "step": 3415 }, { "epoch": 0.87, "grad_norm": 0.29296875, "learning_rate": 9.820508744307177e-06, "loss": 0.3845, "step": 3420 }, { "epoch": 0.87, "grad_norm": 0.443359375, "learning_rate": 9.629041701001507e-06, "loss": 0.4296, "step": 3425 }, { "epoch": 0.87, "grad_norm": 0.3203125, "learning_rate": 9.439365117633404e-06, "loss": 0.3861, "step": 3430 }, { "epoch": 0.88, "grad_norm": 0.47265625, "learning_rate": 9.251482752139352e-06, "loss": 0.3431, "step": 3435 }, { "epoch": 0.88, "grad_norm": 0.447265625, "learning_rate": 9.065398326908215e-06, "loss": 0.3941, "step": 3440 }, { "epoch": 0.88, "grad_norm": 0.39453125, "learning_rate": 8.881115528707462e-06, "loss": 0.3596, "step": 3445 }, { "epoch": 0.88, "grad_norm": 0.333984375, "learning_rate": 8.698638008610094e-06, "loss": 0.4148, "step": 3450 }, { "epoch": 0.88, "grad_norm": 0.43359375, "learning_rate": 8.517969381922364e-06, "loss": 0.3717, "step": 3455 }, { "epoch": 0.88, "grad_norm": 0.416015625, "learning_rate": 8.339113228112083e-06, "loss": 0.4585, "step": 3460 }, { "epoch": 0.88, "grad_norm": 0.306640625, "learning_rate": 8.162073090737776e-06, "loss": 0.4115, "step": 3465 }, { "epoch": 0.88, "grad_norm": 0.34375, "learning_rate": 7.986852477378404e-06, "loss": 0.3699, "step": 3470 }, { "epoch": 0.89, "grad_norm": 0.322265625, "learning_rate": 7.813454859563929e-06, "loss": 0.4215, "step": 3475 }, { "epoch": 0.89, "grad_norm": 0.330078125, "learning_rate": 7.641883672706496e-06, "loss": 0.3779, "step": 3480 }, { "epoch": 0.89, "grad_norm": 0.375, "learning_rate": 7.472142316032382e-06, "loss": 0.3675, "step": 3485 }, { "epoch": 0.89, "grad_norm": 0.349609375, "learning_rate": 7.304234152514655e-06, "loss": 0.4172, "step": 3490 }, { "epoch": 0.89, "grad_norm": 0.328125, "learning_rate": 7.138162508806556e-06, "loss": 0.423, "step": 3495 }, { "epoch": 0.89, "grad_norm": 0.310546875, "learning_rate": 6.973930675175533e-06, "loss": 0.368, "step": 3500 }, { "epoch": 0.89, "grad_norm": 0.291015625, "learning_rate": 6.811541905438145e-06, "loss": 0.4112, "step": 3505 }, { "epoch": 0.89, "grad_norm": 0.37109375, "learning_rate": 6.6509994168955204e-06, "loss": 0.4059, "step": 3510 }, { "epoch": 0.9, "grad_norm": 0.46484375, "learning_rate": 6.492306390269642e-06, "loss": 0.3527, "step": 3515 }, { "epoch": 0.9, "grad_norm": 0.279296875, "learning_rate": 6.335465969640331e-06, "loss": 0.3475, "step": 3520 }, { "epoch": 0.9, "grad_norm": 0.455078125, "learning_rate": 6.180481262382964e-06, "loss": 0.3912, "step": 3525 }, { "epoch": 0.9, "grad_norm": 0.41015625, "learning_rate": 6.027355339106877e-06, "loss": 0.428, "step": 3530 }, { "epoch": 0.9, "grad_norm": 0.369140625, "learning_rate": 5.876091233594572e-06, "loss": 0.4253, "step": 3535 }, { "epoch": 0.9, "grad_norm": 0.369140625, "learning_rate": 5.726691942741569e-06, "loss": 0.4701, "step": 3540 }, { "epoch": 0.9, "grad_norm": 0.400390625, "learning_rate": 5.579160426497065e-06, "loss": 0.3716, "step": 3545 }, { "epoch": 0.91, "grad_norm": 0.51171875, "learning_rate": 5.4334996078052705e-06, "loss": 0.3864, "step": 3550 }, { "epoch": 0.91, "grad_norm": 0.400390625, "learning_rate": 5.289712372547495e-06, "loss": 0.4586, "step": 3555 }, { "epoch": 0.91, "grad_norm": 0.310546875, "learning_rate": 5.147801569485e-06, "loss": 0.4011, "step": 3560 }, { "epoch": 0.91, "grad_norm": 0.353515625, "learning_rate": 5.007770010202528e-06, "loss": 0.3688, "step": 3565 }, { "epoch": 0.91, "grad_norm": 0.357421875, "learning_rate": 4.8696204690526155e-06, "loss": 0.3555, "step": 3570 }, { "epoch": 0.91, "grad_norm": 0.427734375, "learning_rate": 4.733355683100604e-06, "loss": 0.3783, "step": 3575 }, { "epoch": 0.91, "grad_norm": 0.42578125, "learning_rate": 4.598978352070437e-06, "loss": 0.386, "step": 3580 }, { "epoch": 0.91, "grad_norm": 0.423828125, "learning_rate": 4.466491138291185e-06, "loss": 0.3876, "step": 3585 }, { "epoch": 0.92, "grad_norm": 0.3984375, "learning_rate": 4.3358966666442455e-06, "loss": 0.4083, "step": 3590 }, { "epoch": 0.92, "grad_norm": 0.369140625, "learning_rate": 4.207197524511375e-06, "loss": 0.3974, "step": 3595 }, { "epoch": 0.92, "grad_norm": 0.4296875, "learning_rate": 4.0803962617234405e-06, "loss": 0.3811, "step": 3600 }, { "epoch": 0.92, "grad_norm": 0.34375, "learning_rate": 3.955495390509878e-06, "loss": 0.3775, "step": 3605 }, { "epoch": 0.92, "grad_norm": 0.45703125, "learning_rate": 3.8324973854489074e-06, "loss": 0.446, "step": 3610 }, { "epoch": 0.92, "grad_norm": 0.412109375, "learning_rate": 3.71140468341854e-06, "loss": 0.4419, "step": 3615 }, { "epoch": 0.92, "grad_norm": 0.41015625, "learning_rate": 3.592219683548259e-06, "loss": 0.3783, "step": 3620 }, { "epoch": 0.92, "grad_norm": 0.388671875, "learning_rate": 3.474944747171538e-06, "loss": 0.4398, "step": 3625 }, { "epoch": 0.93, "grad_norm": 0.357421875, "learning_rate": 3.35958219777901e-06, "loss": 0.4204, "step": 3630 }, { "epoch": 0.93, "grad_norm": 0.431640625, "learning_rate": 3.246134320972438e-06, "loss": 0.4019, "step": 3635 }, { "epoch": 0.93, "grad_norm": 0.34765625, "learning_rate": 3.134603364419475e-06, "loss": 0.3955, "step": 3640 }, { "epoch": 0.93, "grad_norm": 0.275390625, "learning_rate": 3.0249915378090854e-06, "loss": 0.414, "step": 3645 }, { "epoch": 0.93, "grad_norm": 0.341796875, "learning_rate": 2.9173010128077826e-06, "loss": 0.3865, "step": 3650 }, { "epoch": 0.93, "grad_norm": 0.294921875, "learning_rate": 2.811533923016607e-06, "loss": 0.4735, "step": 3655 }, { "epoch": 0.93, "grad_norm": 0.310546875, "learning_rate": 2.7076923639288486e-06, "loss": 0.3737, "step": 3660 }, { "epoch": 0.93, "grad_norm": 0.345703125, "learning_rate": 2.605778392888547e-06, "loss": 0.3935, "step": 3665 }, { "epoch": 0.94, "grad_norm": 0.353515625, "learning_rate": 2.505794029049702e-06, "loss": 0.3963, "step": 3670 }, { "epoch": 0.94, "grad_norm": 0.353515625, "learning_rate": 2.4077412533362707e-06, "loss": 0.4122, "step": 3675 }, { "epoch": 0.94, "grad_norm": 0.36328125, "learning_rate": 2.3116220084029694e-06, "loss": 0.3266, "step": 3680 }, { "epoch": 0.94, "grad_norm": 0.34765625, "learning_rate": 2.217438198596733e-06, "loss": 0.4134, "step": 3685 }, { "epoch": 0.94, "grad_norm": 0.388671875, "learning_rate": 2.125191689918993e-06, "loss": 0.4304, "step": 3690 }, { "epoch": 0.94, "grad_norm": 0.3828125, "learning_rate": 2.034884309988727e-06, "loss": 0.4402, "step": 3695 }, { "epoch": 0.94, "grad_norm": 0.3046875, "learning_rate": 1.9465178480062353e-06, "loss": 0.4104, "step": 3700 }, { "epoch": 0.94, "grad_norm": 0.298828125, "learning_rate": 1.860094054717687e-06, "loss": 0.4293, "step": 3705 }, { "epoch": 0.95, "grad_norm": 0.287109375, "learning_rate": 1.7756146423804742e-06, "loss": 0.317, "step": 3710 }, { "epoch": 0.95, "grad_norm": 0.388671875, "learning_rate": 1.6930812847292255e-06, "loss": 0.3738, "step": 3715 }, { "epoch": 0.95, "grad_norm": 0.3515625, "learning_rate": 1.6124956169426886e-06, "loss": 0.4004, "step": 3720 }, { "epoch": 0.95, "grad_norm": 0.39453125, "learning_rate": 1.5338592356113124e-06, "loss": 0.4346, "step": 3725 }, { "epoch": 0.95, "grad_norm": 0.24609375, "learning_rate": 1.4571736987056495e-06, "loss": 0.4255, "step": 3730 }, { "epoch": 0.95, "grad_norm": 0.341796875, "learning_rate": 1.3824405255454475e-06, "loss": 0.3787, "step": 3735 }, { "epoch": 0.95, "grad_norm": 0.390625, "learning_rate": 1.3096611967695517e-06, "loss": 0.3445, "step": 3740 }, { "epoch": 0.95, "grad_norm": 0.431640625, "learning_rate": 1.2388371543066045e-06, "loss": 0.4093, "step": 3745 }, { "epoch": 0.96, "grad_norm": 0.3671875, "learning_rate": 1.1699698013464488e-06, "loss": 0.3977, "step": 3750 }, { "epoch": 0.96, "grad_norm": 0.380859375, "learning_rate": 1.1030605023123364e-06, "loss": 0.373, "step": 3755 }, { "epoch": 0.96, "grad_norm": 0.34765625, "learning_rate": 1.0381105828338844e-06, "loss": 0.3833, "step": 3760 }, { "epoch": 0.96, "grad_norm": 0.275390625, "learning_rate": 9.751213297208184e-07, "loss": 0.434, "step": 3765 }, { "epoch": 0.96, "grad_norm": 0.35546875, "learning_rate": 9.140939909375034e-07, "loss": 0.3993, "step": 3770 }, { "epoch": 0.96, "grad_norm": 0.275390625, "learning_rate": 8.550297755781644e-07, "loss": 0.3842, "step": 3775 }, { "epoch": 0.96, "grad_norm": 0.345703125, "learning_rate": 7.979298538429935e-07, "loss": 0.462, "step": 3780 }, { "epoch": 0.97, "grad_norm": 0.408203125, "learning_rate": 7.427953570149027e-07, "loss": 0.3928, "step": 3785 }, { "epoch": 0.97, "grad_norm": 0.3359375, "learning_rate": 6.896273774371742e-07, "loss": 0.3982, "step": 3790 }, { "epoch": 0.97, "grad_norm": 0.314453125, "learning_rate": 6.384269684917676e-07, "loss": 0.4198, "step": 3795 }, { "epoch": 0.97, "grad_norm": 0.380859375, "learning_rate": 5.891951445784583e-07, "loss": 0.3316, "step": 3800 }, { "epoch": 0.97, "grad_norm": 0.279296875, "learning_rate": 5.419328810947755e-07, "loss": 0.4272, "step": 3805 }, { "epoch": 0.97, "grad_norm": 0.34375, "learning_rate": 4.966411144166405e-07, "loss": 0.3763, "step": 3810 }, { "epoch": 0.97, "grad_norm": 0.326171875, "learning_rate": 4.533207418798147e-07, "loss": 0.3747, "step": 3815 }, { "epoch": 0.97, "grad_norm": 0.28125, "learning_rate": 4.119726217621578e-07, "loss": 0.4468, "step": 3820 }, { "epoch": 0.98, "grad_norm": 0.41015625, "learning_rate": 3.7259757326657543e-07, "loss": 0.3989, "step": 3825 }, { "epoch": 0.98, "grad_norm": 0.306640625, "learning_rate": 3.351963765048427e-07, "loss": 0.3995, "step": 3830 }, { "epoch": 0.98, "grad_norm": 0.43359375, "learning_rate": 2.997697724820725e-07, "loss": 0.4069, "step": 3835 }, { "epoch": 0.98, "grad_norm": 0.384765625, "learning_rate": 2.663184630821158e-07, "loss": 0.397, "step": 3840 }, { "epoch": 0.98, "grad_norm": 0.287109375, "learning_rate": 2.3484311105360645e-07, "loss": 0.383, "step": 3845 }, { "epoch": 0.98, "grad_norm": 0.333984375, "learning_rate": 2.053443399968602e-07, "loss": 0.3835, "step": 3850 }, { "epoch": 0.98, "grad_norm": 0.337890625, "learning_rate": 1.778227343514627e-07, "loss": 0.4508, "step": 3855 }, { "epoch": 0.98, "grad_norm": 0.3203125, "learning_rate": 1.522788393847785e-07, "loss": 0.4136, "step": 3860 }, { "epoch": 0.99, "grad_norm": 0.375, "learning_rate": 1.287131611810599e-07, "loss": 0.3933, "step": 3865 }, { "epoch": 0.99, "grad_norm": 0.365234375, "learning_rate": 1.0712616663149932e-07, "loss": 0.3702, "step": 3870 }, { "epoch": 0.99, "grad_norm": 0.302734375, "learning_rate": 8.751828342491442e-08, "loss": 0.3742, "step": 3875 }, { "epoch": 0.99, "grad_norm": 0.349609375, "learning_rate": 6.988990003929941e-08, "loss": 0.4224, "step": 3880 }, { "epoch": 0.99, "grad_norm": 0.376953125, "learning_rate": 5.4241365734153396e-08, "loss": 0.4584, "step": 3885 }, { "epoch": 0.99, "grad_norm": 0.380859375, "learning_rate": 4.057299054350816e-08, "loss": 0.4101, "step": 3890 }, { "epoch": 0.99, "grad_norm": 0.33984375, "learning_rate": 2.8885045269833044e-08, "loss": 0.4152, "step": 3895 }, { "epoch": 0.99, "grad_norm": 0.294921875, "learning_rate": 1.9177761478639255e-08, "loss": 0.4205, "step": 3900 }, { "epoch": 1.0, "grad_norm": 0.41015625, "learning_rate": 1.1451331493916862e-08, "loss": 0.3096, "step": 3905 }, { "epoch": 1.0, "grad_norm": 0.3125, "learning_rate": 5.7059083943045154e-09, "loss": 0.3654, "step": 3910 }, { "epoch": 1.0, "grad_norm": 0.4921875, "learning_rate": 1.941606010069652e-09, "loss": 0.4881, "step": 3915 }, { "epoch": 1.0, "grad_norm": 0.32421875, "learning_rate": 1.5849892083252828e-10, "loss": 0.406, "step": 3920 }, { "epoch": 1.0, "eval_loss": 0.39284008741378784, "eval_runtime": 450.3981, "eval_samples_per_second": 3.71, "eval_steps_per_second": 0.464, "step": 3922 }, { "epoch": 1.0, "step": 3922, "total_flos": 2.757410325375484e+18, "train_loss": 0.47386226463779874, "train_runtime": 32073.4099, "train_samples_per_second": 0.978, "train_steps_per_second": 0.122 } ], "logging_steps": 5, "max_steps": 3922, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "total_flos": 2.757410325375484e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }