{ "best_metric": null, "best_model_checkpoint": null, "epoch": 19.51219512195122, "eval_steps": 100, "global_step": 500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.1951219512195122, "grad_norm": 3.5239181518554688, "learning_rate": 4.999451708687114e-05, "loss": 1.1522, "num_input_tokens_seen": 2231104, "step": 5 }, { "epoch": 0.3902439024390244, "grad_norm": 1.2978184223175049, "learning_rate": 4.997807075247146e-05, "loss": 0.8861, "num_input_tokens_seen": 4468608, "step": 10 }, { "epoch": 0.5853658536585366, "grad_norm": 1.0176712274551392, "learning_rate": 4.995066821070679e-05, "loss": 0.8351, "num_input_tokens_seen": 6779584, "step": 15 }, { "epoch": 0.7804878048780488, "grad_norm": 0.6517723798751831, "learning_rate": 4.991232148123761e-05, "loss": 0.7797, "num_input_tokens_seen": 9106880, "step": 20 }, { "epoch": 0.975609756097561, "grad_norm": 0.6675965785980225, "learning_rate": 4.9863047384206835e-05, "loss": 0.7575, "num_input_tokens_seen": 11435136, "step": 25 }, { "epoch": 1.170731707317073, "grad_norm": 1.225784182548523, "learning_rate": 4.980286753286195e-05, "loss": 0.7345, "num_input_tokens_seen": 13738880, "step": 30 }, { "epoch": 1.3658536585365852, "grad_norm": 0.4991011321544647, "learning_rate": 4.9731808324074717e-05, "loss": 0.7443, "num_input_tokens_seen": 16001920, "step": 35 }, { "epoch": 1.5609756097560976, "grad_norm": 0.7908358573913574, "learning_rate": 4.964990092676263e-05, "loss": 0.7211, "num_input_tokens_seen": 18282944, "step": 40 }, { "epoch": 1.7560975609756098, "grad_norm": 0.9223726391792297, "learning_rate": 4.9557181268217227e-05, "loss": 0.7306, "num_input_tokens_seen": 20575296, "step": 45 }, { "epoch": 1.951219512195122, "grad_norm": 0.8891558051109314, "learning_rate": 4.9453690018345144e-05, "loss": 0.7039, "num_input_tokens_seen": 22849856, "step": 50 }, { "epoch": 2.1463414634146343, "grad_norm": 1.3324096202850342, "learning_rate": 4.933947257182901e-05, "loss": 0.7044, "num_input_tokens_seen": 25176896, "step": 55 }, { "epoch": 2.341463414634146, "grad_norm": 1.0357320308685303, "learning_rate": 4.9214579028215776e-05, "loss": 0.6768, "num_input_tokens_seen": 27457472, "step": 60 }, { "epoch": 2.5365853658536586, "grad_norm": 0.6939280033111572, "learning_rate": 4.907906416994146e-05, "loss": 0.6906, "num_input_tokens_seen": 29772544, "step": 65 }, { "epoch": 2.7317073170731705, "grad_norm": 0.7578678727149963, "learning_rate": 4.893298743830168e-05, "loss": 0.6595, "num_input_tokens_seen": 32013760, "step": 70 }, { "epoch": 2.926829268292683, "grad_norm": 0.548279345035553, "learning_rate": 4.877641290737884e-05, "loss": 0.6514, "num_input_tokens_seen": 34342848, "step": 75 }, { "epoch": 3.1219512195121952, "grad_norm": 1.1982917785644531, "learning_rate": 4.860940925593703e-05, "loss": 0.6498, "num_input_tokens_seen": 36592000, "step": 80 }, { "epoch": 3.317073170731707, "grad_norm": 1.0446078777313232, "learning_rate": 4.843204973729729e-05, "loss": 0.6537, "num_input_tokens_seen": 38878528, "step": 85 }, { "epoch": 3.5121951219512195, "grad_norm": 0.7469108700752258, "learning_rate": 4.8244412147206284e-05, "loss": 0.5996, "num_input_tokens_seen": 41156672, "step": 90 }, { "epoch": 3.7073170731707314, "grad_norm": 0.7562854886054993, "learning_rate": 4.8046578789712515e-05, "loss": 0.6289, "num_input_tokens_seen": 43475456, "step": 95 }, { "epoch": 3.902439024390244, "grad_norm": 0.7673454880714417, "learning_rate": 4.783863644106502e-05, "loss": 0.6607, "num_input_tokens_seen": 45798272, "step": 100 }, { "epoch": 3.902439024390244, "eval_loss": 0.6353042125701904, "eval_runtime": 174.6543, "eval_samples_per_second": 12.07, "eval_steps_per_second": 0.504, "num_input_tokens_seen": 45798272, "step": 100 }, { "epoch": 4.097560975609756, "grad_norm": 0.6962050199508667, "learning_rate": 4.762067631165049e-05, "loss": 0.5924, "num_input_tokens_seen": 48035456, "step": 105 }, { "epoch": 4.2926829268292686, "grad_norm": 1.080012559890747, "learning_rate": 4.7392794005985326e-05, "loss": 0.6107, "num_input_tokens_seen": 50295680, "step": 110 }, { "epoch": 4.487804878048781, "grad_norm": 0.6142390966415405, "learning_rate": 4.715508948078037e-05, "loss": 0.6016, "num_input_tokens_seen": 52611072, "step": 115 }, { "epoch": 4.682926829268292, "grad_norm": 0.7695103883743286, "learning_rate": 4.690766700109659e-05, "loss": 0.5946, "num_input_tokens_seen": 54817920, "step": 120 }, { "epoch": 4.878048780487805, "grad_norm": 0.7141166925430298, "learning_rate": 4.665063509461097e-05, "loss": 0.5926, "num_input_tokens_seen": 57135552, "step": 125 }, { "epoch": 5.073170731707317, "grad_norm": 0.9515812993049622, "learning_rate": 4.638410650401267e-05, "loss": 0.5986, "num_input_tokens_seen": 59455680, "step": 130 }, { "epoch": 5.2682926829268295, "grad_norm": 1.5192577838897705, "learning_rate": 4.610819813755038e-05, "loss": 0.5619, "num_input_tokens_seen": 61782208, "step": 135 }, { "epoch": 5.463414634146342, "grad_norm": 1.0420284271240234, "learning_rate": 4.5823031017752485e-05, "loss": 0.5739, "num_input_tokens_seen": 64051520, "step": 140 }, { "epoch": 5.658536585365853, "grad_norm": 0.6723448038101196, "learning_rate": 4.5528730228342605e-05, "loss": 0.5673, "num_input_tokens_seen": 66320192, "step": 145 }, { "epoch": 5.853658536585366, "grad_norm": 0.8727495074272156, "learning_rate": 4.522542485937369e-05, "loss": 0.5607, "num_input_tokens_seen": 68535424, "step": 150 }, { "epoch": 6.048780487804878, "grad_norm": 1.1162189245224, "learning_rate": 4.491324795060491e-05, "loss": 0.5711, "num_input_tokens_seen": 70861632, "step": 155 }, { "epoch": 6.2439024390243905, "grad_norm": 1.0553621053695679, "learning_rate": 4.4592336433146e-05, "loss": 0.5437, "num_input_tokens_seen": 73124480, "step": 160 }, { "epoch": 6.439024390243903, "grad_norm": 1.105603575706482, "learning_rate": 4.426283106939474e-05, "loss": 0.5437, "num_input_tokens_seen": 75439232, "step": 165 }, { "epoch": 6.634146341463414, "grad_norm": 0.9295886158943176, "learning_rate": 4.3924876391293915e-05, "loss": 0.5321, "num_input_tokens_seen": 77689536, "step": 170 }, { "epoch": 6.829268292682927, "grad_norm": 0.9514282941818237, "learning_rate": 4.357862063693486e-05, "loss": 0.5075, "num_input_tokens_seen": 80000000, "step": 175 }, { "epoch": 7.024390243902439, "grad_norm": 0.641543447971344, "learning_rate": 4.3224215685535294e-05, "loss": 0.5246, "num_input_tokens_seen": 82294336, "step": 180 }, { "epoch": 7.219512195121951, "grad_norm": 0.6666741967201233, "learning_rate": 4.2861816990820084e-05, "loss": 0.5021, "num_input_tokens_seen": 84609856, "step": 185 }, { "epoch": 7.414634146341464, "grad_norm": 0.9410921931266785, "learning_rate": 4.249158351283414e-05, "loss": 0.4766, "num_input_tokens_seen": 86956608, "step": 190 }, { "epoch": 7.609756097560975, "grad_norm": 0.9510065913200378, "learning_rate": 4.211367764821722e-05, "loss": 0.4922, "num_input_tokens_seen": 89199040, "step": 195 }, { "epoch": 7.804878048780488, "grad_norm": 0.9778364896774292, "learning_rate": 4.172826515897146e-05, "loss": 0.5149, "num_input_tokens_seen": 91464512, "step": 200 }, { "epoch": 7.804878048780488, "eval_loss": 0.6269230246543884, "eval_runtime": 173.5455, "eval_samples_per_second": 12.147, "eval_steps_per_second": 0.507, "num_input_tokens_seen": 91464512, "step": 200 }, { "epoch": 8.0, "grad_norm": 1.1030495166778564, "learning_rate": 4.133551509975264e-05, "loss": 0.5019, "num_input_tokens_seen": 93698368, "step": 205 }, { "epoch": 8.195121951219512, "grad_norm": 0.8688378930091858, "learning_rate": 4.093559974371725e-05, "loss": 0.4275, "num_input_tokens_seen": 96008256, "step": 210 }, { "epoch": 8.390243902439025, "grad_norm": 0.9601365327835083, "learning_rate": 4.052869450695776e-05, "loss": 0.4545, "num_input_tokens_seen": 98232832, "step": 215 }, { "epoch": 8.585365853658537, "grad_norm": 0.9640222191810608, "learning_rate": 4.011497787155938e-05, "loss": 0.4577, "num_input_tokens_seen": 100520128, "step": 220 }, { "epoch": 8.78048780487805, "grad_norm": 0.7299013137817383, "learning_rate": 3.969463130731183e-05, "loss": 0.4589, "num_input_tokens_seen": 102814976, "step": 225 }, { "epoch": 8.975609756097562, "grad_norm": 0.6459980607032776, "learning_rate": 3.92678391921108e-05, "loss": 0.4482, "num_input_tokens_seen": 105142528, "step": 230 }, { "epoch": 9.170731707317072, "grad_norm": 0.8188865184783936, "learning_rate": 3.883478873108361e-05, "loss": 0.4301, "num_input_tokens_seen": 107394880, "step": 235 }, { "epoch": 9.365853658536585, "grad_norm": 0.7741889953613281, "learning_rate": 3.8395669874474915e-05, "loss": 0.3852, "num_input_tokens_seen": 109706560, "step": 240 }, { "epoch": 9.560975609756097, "grad_norm": 0.7010043263435364, "learning_rate": 3.795067523432826e-05, "loss": 0.3886, "num_input_tokens_seen": 111993664, "step": 245 }, { "epoch": 9.75609756097561, "grad_norm": 0.5523029565811157, "learning_rate": 3.7500000000000003e-05, "loss": 0.4028, "num_input_tokens_seen": 114255232, "step": 250 }, { "epoch": 9.951219512195122, "grad_norm": 0.5837187170982361, "learning_rate": 3.704384185254288e-05, "loss": 0.3847, "num_input_tokens_seen": 116578560, "step": 255 }, { "epoch": 10.146341463414634, "grad_norm": 0.7814375162124634, "learning_rate": 3.6582400877996546e-05, "loss": 0.3516, "num_input_tokens_seen": 118836800, "step": 260 }, { "epoch": 10.341463414634147, "grad_norm": 0.9329513907432556, "learning_rate": 3.611587947962319e-05, "loss": 0.3504, "num_input_tokens_seen": 121148608, "step": 265 }, { "epoch": 10.536585365853659, "grad_norm": 0.5801968574523926, "learning_rate": 3.564448228912682e-05, "loss": 0.3303, "num_input_tokens_seen": 123490176, "step": 270 }, { "epoch": 10.731707317073171, "grad_norm": 0.5544772744178772, "learning_rate": 3.516841607689501e-05, "loss": 0.336, "num_input_tokens_seen": 125741376, "step": 275 }, { "epoch": 10.926829268292684, "grad_norm": 0.5629393458366394, "learning_rate": 3.4687889661302576e-05, "loss": 0.337, "num_input_tokens_seen": 127996480, "step": 280 }, { "epoch": 11.121951219512194, "grad_norm": 0.6132429242134094, "learning_rate": 3.4203113817116957e-05, "loss": 0.3102, "num_input_tokens_seen": 130286208, "step": 285 }, { "epoch": 11.317073170731707, "grad_norm": 0.6265811920166016, "learning_rate": 3.3714301183045385e-05, "loss": 0.2777, "num_input_tokens_seen": 132600256, "step": 290 }, { "epoch": 11.512195121951219, "grad_norm": 0.7565940618515015, "learning_rate": 3.322166616846458e-05, "loss": 0.2875, "num_input_tokens_seen": 134908928, "step": 295 }, { "epoch": 11.707317073170731, "grad_norm": 0.8063022494316101, "learning_rate": 3.272542485937369e-05, "loss": 0.2768, "num_input_tokens_seen": 137181248, "step": 300 }, { "epoch": 11.707317073170731, "eval_loss": 0.718697726726532, "eval_runtime": 173.4793, "eval_samples_per_second": 12.151, "eval_steps_per_second": 0.507, "num_input_tokens_seen": 137181248, "step": 300 }, { "epoch": 11.902439024390244, "grad_norm": 0.7160292267799377, "learning_rate": 3.222579492361179e-05, "loss": 0.2823, "num_input_tokens_seen": 139385728, "step": 305 }, { "epoch": 12.097560975609756, "grad_norm": 0.6414353251457214, "learning_rate": 3.172299551538164e-05, "loss": 0.27, "num_input_tokens_seen": 141670400, "step": 310 }, { "epoch": 12.292682926829269, "grad_norm": 0.6131883859634399, "learning_rate": 3.121724717912138e-05, "loss": 0.2308, "num_input_tokens_seen": 143944512, "step": 315 }, { "epoch": 12.487804878048781, "grad_norm": 0.46672749519348145, "learning_rate": 3.0708771752766394e-05, "loss": 0.2236, "num_input_tokens_seen": 146200128, "step": 320 }, { "epoch": 12.682926829268293, "grad_norm": 0.48453348875045776, "learning_rate": 3.0197792270443982e-05, "loss": 0.2227, "num_input_tokens_seen": 148494912, "step": 325 }, { "epoch": 12.878048780487806, "grad_norm": 0.5639292597770691, "learning_rate": 2.9684532864643122e-05, "loss": 0.2236, "num_input_tokens_seen": 150765632, "step": 330 }, { "epoch": 13.073170731707316, "grad_norm": 0.45188578963279724, "learning_rate": 2.916921866790256e-05, "loss": 0.2193, "num_input_tokens_seen": 153025984, "step": 335 }, { "epoch": 13.268292682926829, "grad_norm": 0.5362280011177063, "learning_rate": 2.8652075714060295e-05, "loss": 0.1804, "num_input_tokens_seen": 155359744, "step": 340 }, { "epoch": 13.463414634146341, "grad_norm": 0.46796223521232605, "learning_rate": 2.8133330839107608e-05, "loss": 0.1763, "num_input_tokens_seen": 157653760, "step": 345 }, { "epoch": 13.658536585365853, "grad_norm": 0.6022464036941528, "learning_rate": 2.761321158169134e-05, "loss": 0.1785, "num_input_tokens_seen": 159970112, "step": 350 }, { "epoch": 13.853658536585366, "grad_norm": 0.6007694005966187, "learning_rate": 2.7091946083307896e-05, "loss": 0.1847, "num_input_tokens_seen": 162266048, "step": 355 }, { "epoch": 14.048780487804878, "grad_norm": 0.5330171585083008, "learning_rate": 2.656976298823284e-05, "loss": 0.1765, "num_input_tokens_seen": 164502784, "step": 360 }, { "epoch": 14.24390243902439, "grad_norm": 0.4508293569087982, "learning_rate": 2.604689134322999e-05, "loss": 0.1433, "num_input_tokens_seen": 166793792, "step": 365 }, { "epoch": 14.439024390243903, "grad_norm": 0.4898134469985962, "learning_rate": 2.5523560497083926e-05, "loss": 0.1436, "num_input_tokens_seen": 169105536, "step": 370 }, { "epoch": 14.634146341463415, "grad_norm": 0.5252612829208374, "learning_rate": 2.5e-05, "loss": 0.1456, "num_input_tokens_seen": 171380160, "step": 375 }, { "epoch": 14.829268292682928, "grad_norm": 0.45608317852020264, "learning_rate": 2.447643950291608e-05, "loss": 0.1361, "num_input_tokens_seen": 173636544, "step": 380 }, { "epoch": 15.024390243902438, "grad_norm": 0.5009132623672485, "learning_rate": 2.3953108656770016e-05, "loss": 0.1381, "num_input_tokens_seen": 175912832, "step": 385 }, { "epoch": 15.21951219512195, "grad_norm": 0.4505337178707123, "learning_rate": 2.3430237011767167e-05, "loss": 0.1121, "num_input_tokens_seen": 178157824, "step": 390 }, { "epoch": 15.414634146341463, "grad_norm": 0.43384668231010437, "learning_rate": 2.2908053916692117e-05, "loss": 0.1103, "num_input_tokens_seen": 180468736, "step": 395 }, { "epoch": 15.609756097560975, "grad_norm": 0.4433673620223999, "learning_rate": 2.238678841830867e-05, "loss": 0.1126, "num_input_tokens_seen": 182788736, "step": 400 }, { "epoch": 15.609756097560975, "eval_loss": 0.8370733261108398, "eval_runtime": 173.2462, "eval_samples_per_second": 12.168, "eval_steps_per_second": 0.508, "num_input_tokens_seen": 182788736, "step": 400 }, { "epoch": 15.804878048780488, "grad_norm": 0.39896300435066223, "learning_rate": 2.186666916089239e-05, "loss": 0.1114, "num_input_tokens_seen": 185050944, "step": 405 }, { "epoch": 16.0, "grad_norm": 0.3975840210914612, "learning_rate": 2.1347924285939714e-05, "loss": 0.1135, "num_input_tokens_seen": 187289920, "step": 410 }, { "epoch": 16.195121951219512, "grad_norm": 0.3381832242012024, "learning_rate": 2.0830781332097446e-05, "loss": 0.0935, "num_input_tokens_seen": 189562944, "step": 415 }, { "epoch": 16.390243902439025, "grad_norm": 0.3201667070388794, "learning_rate": 2.031546713535688e-05, "loss": 0.0851, "num_input_tokens_seen": 191813696, "step": 420 }, { "epoch": 16.585365853658537, "grad_norm": 0.38780859112739563, "learning_rate": 1.980220772955602e-05, "loss": 0.0897, "num_input_tokens_seen": 194094144, "step": 425 }, { "epoch": 16.78048780487805, "grad_norm": 0.31540414690971375, "learning_rate": 1.9291228247233605e-05, "loss": 0.0886, "num_input_tokens_seen": 196385984, "step": 430 }, { "epoch": 16.975609756097562, "grad_norm": 0.5678117275238037, "learning_rate": 1.8782752820878634e-05, "loss": 0.0899, "num_input_tokens_seen": 198700032, "step": 435 }, { "epoch": 17.170731707317074, "grad_norm": 0.4425772726535797, "learning_rate": 1.827700448461836e-05, "loss": 0.0753, "num_input_tokens_seen": 201036480, "step": 440 }, { "epoch": 17.365853658536587, "grad_norm": 0.32113131880760193, "learning_rate": 1.7774205076388206e-05, "loss": 0.0724, "num_input_tokens_seen": 203331520, "step": 445 }, { "epoch": 17.5609756097561, "grad_norm": 0.2662774920463562, "learning_rate": 1.7274575140626318e-05, "loss": 0.0711, "num_input_tokens_seen": 205627072, "step": 450 }, { "epoch": 17.75609756097561, "grad_norm": 0.27264896035194397, "learning_rate": 1.677833383153542e-05, "loss": 0.0719, "num_input_tokens_seen": 207845568, "step": 455 }, { "epoch": 17.951219512195124, "grad_norm": 0.29512110352516174, "learning_rate": 1.6285698816954624e-05, "loss": 0.0704, "num_input_tokens_seen": 210180736, "step": 460 }, { "epoch": 18.146341463414632, "grad_norm": 0.252193421125412, "learning_rate": 1.5796886182883053e-05, "loss": 0.0666, "num_input_tokens_seen": 212454016, "step": 465 }, { "epoch": 18.341463414634145, "grad_norm": 0.23161688446998596, "learning_rate": 1.5312110338697426e-05, "loss": 0.0564, "num_input_tokens_seen": 214714240, "step": 470 }, { "epoch": 18.536585365853657, "grad_norm": 0.25446033477783203, "learning_rate": 1.4831583923104999e-05, "loss": 0.0617, "num_input_tokens_seen": 217009664, "step": 475 }, { "epoch": 18.73170731707317, "grad_norm": 0.2264089584350586, "learning_rate": 1.4355517710873184e-05, "loss": 0.0587, "num_input_tokens_seen": 219270336, "step": 480 }, { "epoch": 18.926829268292682, "grad_norm": 0.21758581697940826, "learning_rate": 1.388412052037682e-05, "loss": 0.0573, "num_input_tokens_seen": 221604736, "step": 485 }, { "epoch": 19.121951219512194, "grad_norm": 0.21286001801490784, "learning_rate": 1.3417599122003464e-05, "loss": 0.0537, "num_input_tokens_seen": 223843968, "step": 490 }, { "epoch": 19.317073170731707, "grad_norm": 0.19537770748138428, "learning_rate": 1.2956158147457115e-05, "loss": 0.0505, "num_input_tokens_seen": 226198144, "step": 495 }, { "epoch": 19.51219512195122, "grad_norm": 0.18156716227531433, "learning_rate": 1.2500000000000006e-05, "loss": 0.0494, "num_input_tokens_seen": 228521280, "step": 500 }, { "epoch": 19.51219512195122, "eval_loss": 0.9114091396331787, "eval_runtime": 173.5404, "eval_samples_per_second": 12.147, "eval_steps_per_second": 0.507, "num_input_tokens_seen": 228521280, "step": 500 } ], "logging_steps": 5, "max_steps": 750, "num_input_tokens_seen": 228521280, "num_train_epochs": 30, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 9.694721607846593e+18, "train_batch_size": 8, "trial_name": null, "trial_params": null }