| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.994236311239193, | |
| "eval_steps": 20, | |
| "global_step": 346, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.011527377521613832, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.7662, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.023054755043227664, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.7773, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.0345821325648415, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.7593, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.04610951008645533, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.795, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.05763688760806916, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.7765, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.069164265129683, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.7819, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.08069164265129683, | |
| "grad_norm": 1.0932377576828003, | |
| "learning_rate": 2.5e-06, | |
| "loss": 0.7426, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.09221902017291066, | |
| "grad_norm": 1.0143316984176636, | |
| "learning_rate": 7.5e-06, | |
| "loss": 0.7667, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.1037463976945245, | |
| "grad_norm": 0.5817523002624512, | |
| "learning_rate": 1.25e-05, | |
| "loss": 0.6819, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.11527377521613832, | |
| "grad_norm": 0.7819337248802185, | |
| "learning_rate": 1.5e-05, | |
| "loss": 0.6171, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.11527377521613832, | |
| "eval_loss": 0.5882565379142761, | |
| "eval_runtime": 219.1309, | |
| "eval_samples_per_second": 21.325, | |
| "eval_steps_per_second": 0.169, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.12680115273775217, | |
| "grad_norm": 0.4540252089500427, | |
| "learning_rate": 2e-05, | |
| "loss": 0.5573, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.138328530259366, | |
| "grad_norm": 0.31643402576446533, | |
| "learning_rate": 2.5e-05, | |
| "loss": 0.5615, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.14985590778097982, | |
| "grad_norm": 0.2575995922088623, | |
| "learning_rate": 3e-05, | |
| "loss": 0.5225, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.16138328530259366, | |
| "grad_norm": 0.1953422576189041, | |
| "learning_rate": 3.5e-05, | |
| "loss": 0.5141, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.1729106628242075, | |
| "grad_norm": 0.17974254488945007, | |
| "learning_rate": 4e-05, | |
| "loss": 0.508, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.1844380403458213, | |
| "grad_norm": 0.16816851496696472, | |
| "learning_rate": 4.5e-05, | |
| "loss": 0.4763, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.19596541786743515, | |
| "grad_norm": 0.16267944872379303, | |
| "learning_rate": 5e-05, | |
| "loss": 0.4757, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.207492795389049, | |
| "grad_norm": 0.1356152594089508, | |
| "learning_rate": 4.9995821084255035e-05, | |
| "loss": 0.4714, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.21902017291066284, | |
| "grad_norm": 0.12293241173028946, | |
| "learning_rate": 4.998328588931674e-05, | |
| "loss": 0.4777, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.23054755043227665, | |
| "grad_norm": 0.12494528293609619, | |
| "learning_rate": 4.996239907149833e-05, | |
| "loss": 0.4649, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.23054755043227665, | |
| "eval_loss": 0.45945191383361816, | |
| "eval_runtime": 163.1604, | |
| "eval_samples_per_second": 28.641, | |
| "eval_steps_per_second": 0.227, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.2420749279538905, | |
| "grad_norm": 0.11425463855266571, | |
| "learning_rate": 4.993316838939995e-05, | |
| "loss": 0.4508, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.25360230547550433, | |
| "grad_norm": 0.11498431116342545, | |
| "learning_rate": 4.9895604701026735e-05, | |
| "loss": 0.4433, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.26512968299711814, | |
| "grad_norm": 0.12802694737911224, | |
| "learning_rate": 4.9849721959755464e-05, | |
| "loss": 0.451, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.276657060518732, | |
| "grad_norm": 0.11723160743713379, | |
| "learning_rate": 4.979553720915145e-05, | |
| "loss": 0.4344, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.2881844380403458, | |
| "grad_norm": 0.11519418656826019, | |
| "learning_rate": 4.9733070576637606e-05, | |
| "loss": 0.4464, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.29971181556195964, | |
| "grad_norm": 0.1040327250957489, | |
| "learning_rate": 4.9662345266017836e-05, | |
| "loss": 0.4367, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.3112391930835735, | |
| "grad_norm": 0.10918751358985901, | |
| "learning_rate": 4.9583387548857836e-05, | |
| "loss": 0.4411, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.3227665706051873, | |
| "grad_norm": 0.11520207673311234, | |
| "learning_rate": 4.949622675472627e-05, | |
| "loss": 0.4119, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.33429394812680113, | |
| "grad_norm": 0.11275804787874222, | |
| "learning_rate": 4.940089526030003e-05, | |
| "loss": 0.4387, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.345821325648415, | |
| "grad_norm": 0.11030508577823639, | |
| "learning_rate": 4.92974284773376e-05, | |
| "loss": 0.4265, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.345821325648415, | |
| "eval_loss": 0.4273378252983093, | |
| "eval_runtime": 163.6226, | |
| "eval_samples_per_second": 28.56, | |
| "eval_steps_per_second": 0.226, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.3573487031700288, | |
| "grad_norm": 0.1050853505730629, | |
| "learning_rate": 4.9185864839525065e-05, | |
| "loss": 0.4217, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.3688760806916426, | |
| "grad_norm": 0.1106187105178833, | |
| "learning_rate": 4.906624578819954e-05, | |
| "loss": 0.4244, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.3804034582132565, | |
| "grad_norm": 0.1203250139951706, | |
| "learning_rate": 4.893861575695544e-05, | |
| "loss": 0.4161, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.3919308357348703, | |
| "grad_norm": 0.13240748643875122, | |
| "learning_rate": 4.880302215513915e-05, | |
| "loss": 0.4205, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.4034582132564842, | |
| "grad_norm": 0.11961999535560608, | |
| "learning_rate": 4.865951535023847e-05, | |
| "loss": 0.4152, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.414985590778098, | |
| "grad_norm": 0.10617737472057343, | |
| "learning_rate": 4.850814864917309e-05, | |
| "loss": 0.4151, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.4265129682997118, | |
| "grad_norm": 0.1053657978773117, | |
| "learning_rate": 4.834897827849325e-05, | |
| "loss": 0.4286, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.43804034582132567, | |
| "grad_norm": 0.1263289898633957, | |
| "learning_rate": 4.818206336349391e-05, | |
| "loss": 0.4206, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.4495677233429395, | |
| "grad_norm": 0.11088111996650696, | |
| "learning_rate": 4.8007465906252065e-05, | |
| "loss": 0.4127, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.4610951008645533, | |
| "grad_norm": 0.10416624695062637, | |
| "learning_rate": 4.782525076259556e-05, | |
| "loss": 0.3924, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.4610951008645533, | |
| "eval_loss": 0.41243475675582886, | |
| "eval_runtime": 163.0434, | |
| "eval_samples_per_second": 28.661, | |
| "eval_steps_per_second": 0.227, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.47262247838616717, | |
| "grad_norm": 0.109275221824646, | |
| "learning_rate": 4.763548561801175e-05, | |
| "loss": 0.418, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.484149855907781, | |
| "grad_norm": 0.11440867185592651, | |
| "learning_rate": 4.743824096250513e-05, | |
| "loss": 0.4054, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.4956772334293948, | |
| "grad_norm": 0.11359245330095291, | |
| "learning_rate": 4.723359006441316e-05, | |
| "loss": 0.3969, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.5072046109510087, | |
| "grad_norm": 0.11708330363035202, | |
| "learning_rate": 4.7021608943190105e-05, | |
| "loss": 0.4052, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.5187319884726225, | |
| "grad_norm": 0.1086309477686882, | |
| "learning_rate": 4.680237634116884e-05, | |
| "loss": 0.4055, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.5302593659942363, | |
| "grad_norm": 0.1133558601140976, | |
| "learning_rate": 4.6575973694311364e-05, | |
| "loss": 0.4016, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.5417867435158501, | |
| "grad_norm": 0.11118675768375397, | |
| "learning_rate": 4.6342485101958634e-05, | |
| "loss": 0.4137, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.553314121037464, | |
| "grad_norm": 0.1130242571234703, | |
| "learning_rate": 4.610199729559106e-05, | |
| "loss": 0.4125, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.5648414985590778, | |
| "grad_norm": 0.10623504966497421, | |
| "learning_rate": 4.585459960661137e-05, | |
| "loss": 0.3989, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.5763688760806917, | |
| "grad_norm": 0.10171157866716385, | |
| "learning_rate": 4.5600383933161586e-05, | |
| "loss": 0.4099, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.5763688760806917, | |
| "eval_loss": 0.40298977494239807, | |
| "eval_runtime": 163.0144, | |
| "eval_samples_per_second": 28.666, | |
| "eval_steps_per_second": 0.227, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.5878962536023055, | |
| "grad_norm": 0.1252756267786026, | |
| "learning_rate": 4.5339444705986636e-05, | |
| "loss": 0.4047, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.5994236311239193, | |
| "grad_norm": 0.12425834685564041, | |
| "learning_rate": 4.507187885335716e-05, | |
| "loss": 0.4044, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.6109510086455331, | |
| "grad_norm": 0.12188662588596344, | |
| "learning_rate": 4.479778576506467e-05, | |
| "loss": 0.3985, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.622478386167147, | |
| "grad_norm": 0.1297188252210617, | |
| "learning_rate": 4.451726725550226e-05, | |
| "loss": 0.4105, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.6340057636887608, | |
| "grad_norm": 0.1405544877052307, | |
| "learning_rate": 4.423042752584471e-05, | |
| "loss": 0.3943, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.6455331412103746, | |
| "grad_norm": 0.1427697390317917, | |
| "learning_rate": 4.393737312534203e-05, | |
| "loss": 0.3929, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.6570605187319885, | |
| "grad_norm": 0.12234554439783096, | |
| "learning_rate": 4.363821291174073e-05, | |
| "loss": 0.385, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.6685878962536023, | |
| "grad_norm": 0.11913135647773743, | |
| "learning_rate": 4.3333058010847665e-05, | |
| "loss": 0.3944, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.6801152737752162, | |
| "grad_norm": 0.10236409306526184, | |
| "learning_rate": 4.302202177525126e-05, | |
| "loss": 0.4088, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.69164265129683, | |
| "grad_norm": 0.1148187592625618, | |
| "learning_rate": 4.270521974221572e-05, | |
| "loss": 0.3925, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.69164265129683, | |
| "eval_loss": 0.39593446254730225, | |
| "eval_runtime": 163.197, | |
| "eval_samples_per_second": 28.634, | |
| "eval_steps_per_second": 0.227, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.7031700288184438, | |
| "grad_norm": 0.1020105630159378, | |
| "learning_rate": 4.238276959076362e-05, | |
| "loss": 0.3959, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.7146974063400576, | |
| "grad_norm": 0.09900052100419998, | |
| "learning_rate": 4.205479109796293e-05, | |
| "loss": 0.3938, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.7262247838616714, | |
| "grad_norm": 0.11457400768995285, | |
| "learning_rate": 4.172140609443478e-05, | |
| "loss": 0.3941, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.7377521613832853, | |
| "grad_norm": 0.12308034300804138, | |
| "learning_rate": 4.138273841909831e-05, | |
| "loss": 0.3773, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.7492795389048992, | |
| "grad_norm": 0.11765991151332855, | |
| "learning_rate": 4.1038913873169606e-05, | |
| "loss": 0.3908, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.760806916426513, | |
| "grad_norm": 0.11604655534029007, | |
| "learning_rate": 4.069006017343161e-05, | |
| "loss": 0.4024, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.7723342939481268, | |
| "grad_norm": 0.10507753491401672, | |
| "learning_rate": 4.033630690479266e-05, | |
| "loss": 0.3874, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.7838616714697406, | |
| "grad_norm": 0.11194345355033875, | |
| "learning_rate": 3.99777854721509e-05, | |
| "loss": 0.3902, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.7953890489913544, | |
| "grad_norm": 0.10498375445604324, | |
| "learning_rate": 3.961462905158275e-05, | |
| "loss": 0.3922, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.8069164265129684, | |
| "grad_norm": 0.10028143227100372, | |
| "learning_rate": 3.924697254087344e-05, | |
| "loss": 0.3884, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.8069164265129684, | |
| "eval_loss": 0.3904818892478943, | |
| "eval_runtime": 165.7335, | |
| "eval_samples_per_second": 28.196, | |
| "eval_steps_per_second": 0.223, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.8184438040345822, | |
| "grad_norm": 0.10579356551170349, | |
| "learning_rate": 3.887495250940798e-05, | |
| "loss": 0.3945, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.829971181556196, | |
| "grad_norm": 0.10865680873394012, | |
| "learning_rate": 3.8498707147441186e-05, | |
| "loss": 0.396, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.8414985590778098, | |
| "grad_norm": 0.10962694138288498, | |
| "learning_rate": 3.8118376214765724e-05, | |
| "loss": 0.3925, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.8530259365994236, | |
| "grad_norm": 0.14297883212566376, | |
| "learning_rate": 3.773410098879694e-05, | |
| "loss": 0.3862, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.8645533141210374, | |
| "grad_norm": 0.12033689022064209, | |
| "learning_rate": 3.734602421209414e-05, | |
| "loss": 0.3923, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.8760806916426513, | |
| "grad_norm": 0.11992188543081284, | |
| "learning_rate": 3.6954290039337544e-05, | |
| "loss": 0.3864, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.8876080691642652, | |
| "grad_norm": 0.1202910915017128, | |
| "learning_rate": 3.6559043983780695e-05, | |
| "loss": 0.3965, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.899135446685879, | |
| "grad_norm": 0.1139882281422615, | |
| "learning_rate": 3.6160432863198227e-05, | |
| "loss": 0.3867, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.9106628242074928, | |
| "grad_norm": 0.10837512463331223, | |
| "learning_rate": 3.575860474534907e-05, | |
| "loss": 0.3734, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.9221902017291066, | |
| "grad_norm": 0.11608846485614777, | |
| "learning_rate": 3.535370889297532e-05, | |
| "loss": 0.3856, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.9221902017291066, | |
| "eval_loss": 0.3864249289035797, | |
| "eval_runtime": 162.9187, | |
| "eval_samples_per_second": 28.683, | |
| "eval_steps_per_second": 0.227, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.9337175792507204, | |
| "grad_norm": 0.11061898618936539, | |
| "learning_rate": 3.494589570835719e-05, | |
| "loss": 0.3874, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.9452449567723343, | |
| "grad_norm": 0.10494677722454071, | |
| "learning_rate": 3.4535316677444745e-05, | |
| "loss": 0.3862, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.9567723342939481, | |
| "grad_norm": 0.11414439976215363, | |
| "learning_rate": 3.412212431358704e-05, | |
| "loss": 0.382, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.968299711815562, | |
| "grad_norm": 0.12404631823301315, | |
| "learning_rate": 3.3706472100879635e-05, | |
| "loss": 0.3893, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.9798270893371758, | |
| "grad_norm": 0.11317316442728043, | |
| "learning_rate": 3.3288514437151505e-05, | |
| "loss": 0.379, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.9913544668587896, | |
| "grad_norm": 0.11200971901416779, | |
| "learning_rate": 3.286840657661259e-05, | |
| "loss": 0.3829, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 1.0028818443804035, | |
| "grad_norm": 0.11575999855995178, | |
| "learning_rate": 3.2446304572183155e-05, | |
| "loss": 0.3757, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 1.0144092219020173, | |
| "grad_norm": 0.1118895411491394, | |
| "learning_rate": 3.2022365217526515e-05, | |
| "loss": 0.3487, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 1.0259365994236311, | |
| "grad_norm": 0.11941341310739517, | |
| "learning_rate": 3.159674598880658e-05, | |
| "loss": 0.3611, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 1.037463976945245, | |
| "grad_norm": 0.12434028834104538, | |
| "learning_rate": 3.116960498619191e-05, | |
| "loss": 0.3527, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.037463976945245, | |
| "eval_loss": 0.38433459401130676, | |
| "eval_runtime": 167.462, | |
| "eval_samples_per_second": 27.905, | |
| "eval_steps_per_second": 0.221, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.0489913544668588, | |
| "grad_norm": 0.1257391721010208, | |
| "learning_rate": 3.0741100875127956e-05, | |
| "loss": 0.346, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 1.0605187319884726, | |
| "grad_norm": 0.1072888970375061, | |
| "learning_rate": 3.0311392827399266e-05, | |
| "loss": 0.351, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 1.0720461095100864, | |
| "grad_norm": 0.11550486832857132, | |
| "learning_rate": 2.9880640462003766e-05, | |
| "loss": 0.3584, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 1.0835734870317002, | |
| "grad_norm": 0.11017844080924988, | |
| "learning_rate": 2.944900378586073e-05, | |
| "loss": 0.3528, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 1.0951008645533142, | |
| "grad_norm": 0.11351864784955978, | |
| "learning_rate": 2.901664313437478e-05, | |
| "loss": 0.3548, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.106628242074928, | |
| "grad_norm": 0.11179786920547485, | |
| "learning_rate": 2.8583719111877844e-05, | |
| "loss": 0.3433, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 1.1181556195965419, | |
| "grad_norm": 0.10140883177518845, | |
| "learning_rate": 2.8150392531971137e-05, | |
| "loss": 0.3524, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 1.1296829971181557, | |
| "grad_norm": 0.10281901806592941, | |
| "learning_rate": 2.7716824357789567e-05, | |
| "loss": 0.3527, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 1.1412103746397695, | |
| "grad_norm": 0.09435153007507324, | |
| "learning_rate": 2.728317564221044e-05, | |
| "loss": 0.3521, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 1.1527377521613833, | |
| "grad_norm": 0.10692735016345978, | |
| "learning_rate": 2.684960746802887e-05, | |
| "loss": 0.3635, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.1527377521613833, | |
| "eval_loss": 0.3818141520023346, | |
| "eval_runtime": 163.2711, | |
| "eval_samples_per_second": 28.621, | |
| "eval_steps_per_second": 0.227, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.1642651296829971, | |
| "grad_norm": 0.10303134471178055, | |
| "learning_rate": 2.6416280888122165e-05, | |
| "loss": 0.341, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 1.175792507204611, | |
| "grad_norm": 0.1047578901052475, | |
| "learning_rate": 2.5983356865625224e-05, | |
| "loss": 0.3483, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 1.1873198847262247, | |
| "grad_norm": 0.09823092073202133, | |
| "learning_rate": 2.5550996214139283e-05, | |
| "loss": 0.3469, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 1.1988472622478386, | |
| "grad_norm": 0.10456107556819916, | |
| "learning_rate": 2.511935953799624e-05, | |
| "loss": 0.3498, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 1.2103746397694524, | |
| "grad_norm": 0.10154879838228226, | |
| "learning_rate": 2.4688607172600742e-05, | |
| "loss": 0.3494, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.2219020172910664, | |
| "grad_norm": 0.10301995277404785, | |
| "learning_rate": 2.4258899124872063e-05, | |
| "loss": 0.344, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 1.23342939481268, | |
| "grad_norm": 0.10198619216680527, | |
| "learning_rate": 2.3830395013808088e-05, | |
| "loss": 0.3453, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 1.244956772334294, | |
| "grad_norm": 0.10618147999048233, | |
| "learning_rate": 2.3403254011193433e-05, | |
| "loss": 0.3492, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 1.2564841498559078, | |
| "grad_norm": 0.10644455999135971, | |
| "learning_rate": 2.2977634782473493e-05, | |
| "loss": 0.3365, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 1.2680115273775217, | |
| "grad_norm": 0.10750509053468704, | |
| "learning_rate": 2.2553695427816847e-05, | |
| "loss": 0.3527, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.2680115273775217, | |
| "eval_loss": 0.37958407402038574, | |
| "eval_runtime": 162.0488, | |
| "eval_samples_per_second": 28.837, | |
| "eval_steps_per_second": 0.228, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.2795389048991355, | |
| "grad_norm": 0.10969573259353638, | |
| "learning_rate": 2.2131593423387414e-05, | |
| "loss": 0.3388, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 1.2910662824207493, | |
| "grad_norm": 0.11101813614368439, | |
| "learning_rate": 2.1711485562848504e-05, | |
| "loss": 0.3417, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 1.302593659942363, | |
| "grad_norm": 0.10727708041667938, | |
| "learning_rate": 2.1293527899120374e-05, | |
| "loss": 0.3478, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 1.314121037463977, | |
| "grad_norm": 0.11611964553594589, | |
| "learning_rate": 2.0877875686412966e-05, | |
| "loss": 0.34, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 1.3256484149855907, | |
| "grad_norm": 0.11851979792118073, | |
| "learning_rate": 2.0464683322555267e-05, | |
| "loss": 0.3413, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.3371757925072045, | |
| "grad_norm": 0.11880774050951004, | |
| "learning_rate": 2.0054104291642818e-05, | |
| "loss": 0.3479, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 1.3487031700288186, | |
| "grad_norm": 0.10287457704544067, | |
| "learning_rate": 1.964629110702469e-05, | |
| "loss": 0.3434, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 1.3602305475504322, | |
| "grad_norm": 0.10474801808595657, | |
| "learning_rate": 1.924139525465093e-05, | |
| "loss": 0.338, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 1.3717579250720462, | |
| "grad_norm": 0.10204192996025085, | |
| "learning_rate": 1.8839567136801772e-05, | |
| "loss": 0.3369, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 1.38328530259366, | |
| "grad_norm": 0.1050105094909668, | |
| "learning_rate": 1.8440956016219314e-05, | |
| "loss": 0.3407, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.38328530259366, | |
| "eval_loss": 0.3772360384464264, | |
| "eval_runtime": 161.7846, | |
| "eval_samples_per_second": 28.884, | |
| "eval_steps_per_second": 0.229, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.3948126801152738, | |
| "grad_norm": 0.10698170959949493, | |
| "learning_rate": 1.8045709960662465e-05, | |
| "loss": 0.355, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 1.4063400576368876, | |
| "grad_norm": 0.09979532659053802, | |
| "learning_rate": 1.765397578790587e-05, | |
| "loss": 0.3419, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 1.4178674351585014, | |
| "grad_norm": 0.09883076697587967, | |
| "learning_rate": 1.7265899011203065e-05, | |
| "loss": 0.3415, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 1.4293948126801153, | |
| "grad_norm": 0.10558874905109406, | |
| "learning_rate": 1.6881623785234285e-05, | |
| "loss": 0.353, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 1.440922190201729, | |
| "grad_norm": 0.1028595045208931, | |
| "learning_rate": 1.650129285255882e-05, | |
| "loss": 0.3537, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.4524495677233429, | |
| "grad_norm": 0.10255461186170578, | |
| "learning_rate": 1.6125047490592034e-05, | |
| "loss": 0.3393, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 1.4639769452449567, | |
| "grad_norm": 0.10520190000534058, | |
| "learning_rate": 1.5753027459126566e-05, | |
| "loss": 0.3469, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 1.4755043227665707, | |
| "grad_norm": 0.10457232594490051, | |
| "learning_rate": 1.5385370948417256e-05, | |
| "loss": 0.3548, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 1.4870317002881843, | |
| "grad_norm": 0.10024940967559814, | |
| "learning_rate": 1.5022214527849105e-05, | |
| "loss": 0.3403, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 1.4985590778097984, | |
| "grad_norm": 0.09729419648647308, | |
| "learning_rate": 1.4663693095207343e-05, | |
| "loss": 0.3372, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.4985590778097984, | |
| "eval_loss": 0.3750576674938202, | |
| "eval_runtime": 161.7883, | |
| "eval_samples_per_second": 28.883, | |
| "eval_steps_per_second": 0.229, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.510086455331412, | |
| "grad_norm": 0.09922486543655396, | |
| "learning_rate": 1.43099398265684e-05, | |
| "loss": 0.329, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 1.521613832853026, | |
| "grad_norm": 0.10505367070436478, | |
| "learning_rate": 1.3961086126830403e-05, | |
| "loss": 0.3404, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 1.5331412103746398, | |
| "grad_norm": 0.10513550788164139, | |
| "learning_rate": 1.3617261580901691e-05, | |
| "loss": 0.3436, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 1.5446685878962536, | |
| "grad_norm": 0.10580945760011673, | |
| "learning_rate": 1.3278593905565224e-05, | |
| "loss": 0.3448, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 1.5561959654178674, | |
| "grad_norm": 0.0965629443526268, | |
| "learning_rate": 1.2945208902037071e-05, | |
| "loss": 0.3323, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.5677233429394812, | |
| "grad_norm": 0.1028643250465393, | |
| "learning_rate": 1.261723040923638e-05, | |
| "loss": 0.341, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 1.579250720461095, | |
| "grad_norm": 0.10219055414199829, | |
| "learning_rate": 1.2294780257784277e-05, | |
| "loss": 0.3337, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 1.5907780979827089, | |
| "grad_norm": 0.10359715670347214, | |
| "learning_rate": 1.1977978224748735e-05, | |
| "loss": 0.3388, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 1.602305475504323, | |
| "grad_norm": 0.09491585195064545, | |
| "learning_rate": 1.166694198915234e-05, | |
| "loss": 0.3404, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 1.6138328530259365, | |
| "grad_norm": 0.10194993764162064, | |
| "learning_rate": 1.1361787088259273e-05, | |
| "loss": 0.3453, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.6138328530259365, | |
| "eval_loss": 0.3727053105831146, | |
| "eval_runtime": 161.5061, | |
| "eval_samples_per_second": 28.934, | |
| "eval_steps_per_second": 0.229, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.6253602305475505, | |
| "grad_norm": 0.09746123105287552, | |
| "learning_rate": 1.1062626874657977e-05, | |
| "loss": 0.3372, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 1.6368876080691641, | |
| "grad_norm": 0.1048843190073967, | |
| "learning_rate": 1.0769572474155296e-05, | |
| "loss": 0.3415, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 1.6484149855907781, | |
| "grad_norm": 0.10008594393730164, | |
| "learning_rate": 1.0482732744497742e-05, | |
| "loss": 0.3332, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 1.659942363112392, | |
| "grad_norm": 0.09605535864830017, | |
| "learning_rate": 1.020221423493533e-05, | |
| "loss": 0.3355, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 1.6714697406340058, | |
| "grad_norm": 0.0957765281200409, | |
| "learning_rate": 9.928121146642841e-06, | |
| "loss": 0.3511, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.6829971181556196, | |
| "grad_norm": 0.10064635425806046, | |
| "learning_rate": 9.660555294013373e-06, | |
| "loss": 0.3462, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 1.6945244956772334, | |
| "grad_norm": 0.09759914875030518, | |
| "learning_rate": 9.399616066838415e-06, | |
| "loss": 0.3476, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 1.7060518731988472, | |
| "grad_norm": 0.1018439456820488, | |
| "learning_rate": 9.145400393388629e-06, | |
| "loss": 0.3542, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 1.717579250720461, | |
| "grad_norm": 0.10098982602357864, | |
| "learning_rate": 8.898002704408944e-06, | |
| "loss": 0.3474, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 1.729106628242075, | |
| "grad_norm": 0.09507515281438828, | |
| "learning_rate": 8.657514898041381e-06, | |
| "loss": 0.3408, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.729106628242075, | |
| "eval_loss": 0.37131258845329285, | |
| "eval_runtime": 161.848, | |
| "eval_samples_per_second": 28.873, | |
| "eval_steps_per_second": 0.229, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.7406340057636887, | |
| "grad_norm": 0.09314768761396408, | |
| "learning_rate": 8.424026305688643e-06, | |
| "loss": 0.341, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 1.7521613832853027, | |
| "grad_norm": 0.09930098056793213, | |
| "learning_rate": 8.197623658831165e-06, | |
| "loss": 0.3397, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 1.7636887608069163, | |
| "grad_norm": 0.10165336728096008, | |
| "learning_rate": 7.978391056809904e-06, | |
| "loss": 0.3419, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 1.7752161383285303, | |
| "grad_norm": 0.09817034006118774, | |
| "learning_rate": 7.766409935586837e-06, | |
| "loss": 0.3456, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 1.7867435158501441, | |
| "grad_norm": 0.10100841522216797, | |
| "learning_rate": 7.5617590374948745e-06, | |
| "loss": 0.3464, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.798270893371758, | |
| "grad_norm": 0.10015368461608887, | |
| "learning_rate": 7.364514381988254e-06, | |
| "loss": 0.3402, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 1.8097982708933718, | |
| "grad_norm": 0.10384070128202438, | |
| "learning_rate": 7.174749237404444e-06, | |
| "loss": 0.3312, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 1.8213256484149856, | |
| "grad_norm": 0.10439052432775497, | |
| "learning_rate": 6.992534093747942e-06, | |
| "loss": 0.3333, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 1.8328530259365994, | |
| "grad_norm": 0.09687253832817078, | |
| "learning_rate": 6.8179366365061e-06, | |
| "loss": 0.3358, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 1.8443804034582132, | |
| "grad_norm": 0.10308189690113068, | |
| "learning_rate": 6.651021721506756e-06, | |
| "loss": 0.3342, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.8443804034582132, | |
| "eval_loss": 0.3700437545776367, | |
| "eval_runtime": 161.117, | |
| "eval_samples_per_second": 29.004, | |
| "eval_steps_per_second": 0.23, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.8559077809798272, | |
| "grad_norm": 0.09917671233415604, | |
| "learning_rate": 6.491851350826915e-06, | |
| "loss": 0.35, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 1.8674351585014408, | |
| "grad_norm": 0.09794910252094269, | |
| "learning_rate": 6.340484649761529e-06, | |
| "loss": 0.3382, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 1.8789625360230549, | |
| "grad_norm": 0.10062103718519211, | |
| "learning_rate": 6.196977844860848e-06, | |
| "loss": 0.3313, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 1.8904899135446684, | |
| "grad_norm": 0.09653456509113312, | |
| "learning_rate": 6.061384243044564e-06, | |
| "loss": 0.3364, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 1.9020172910662825, | |
| "grad_norm": 0.09764689952135086, | |
| "learning_rate": 5.933754211800459e-06, | |
| "loss": 0.332, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.9135446685878963, | |
| "grad_norm": 0.09577332437038422, | |
| "learning_rate": 5.814135160474937e-06, | |
| "loss": 0.339, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 1.92507204610951, | |
| "grad_norm": 0.0969650149345398, | |
| "learning_rate": 5.7025715226624036e-06, | |
| "loss": 0.3301, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 1.936599423631124, | |
| "grad_norm": 0.09393922984600067, | |
| "learning_rate": 5.5991047396999735e-06, | |
| "loss": 0.3374, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 1.9481268011527377, | |
| "grad_norm": 0.11683030426502228, | |
| "learning_rate": 5.503773245273734e-06, | |
| "loss": 0.3424, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 1.9596541786743515, | |
| "grad_norm": 0.09428226202726364, | |
| "learning_rate": 5.41661245114217e-06, | |
| "loss": 0.332, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.9596541786743515, | |
| "eval_loss": 0.36898180842399597, | |
| "eval_runtime": 161.0051, | |
| "eval_samples_per_second": 29.024, | |
| "eval_steps_per_second": 0.23, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.9711815561959654, | |
| "grad_norm": 0.09828232228755951, | |
| "learning_rate": 5.337654733982173e-06, | |
| "loss": 0.3386, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 1.9827089337175794, | |
| "grad_norm": 0.09727545082569122, | |
| "learning_rate": 5.2669294233624e-06, | |
| "loss": 0.331, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 1.994236311239193, | |
| "grad_norm": 0.09713173657655716, | |
| "learning_rate": 5.204462790848549e-06, | |
| "loss": 0.3478, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 1.994236311239193, | |
| "step": 346, | |
| "total_flos": 388981471576064.0, | |
| "train_loss": 0.3994986457976303, | |
| "train_runtime": 42730.5756, | |
| "train_samples_per_second": 4.155, | |
| "train_steps_per_second": 0.008 | |
| } | |
| ], | |
| "logging_steps": 2, | |
| "max_steps": 346, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 1000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 388981471576064.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |