| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 4.0, | |
| "eval_steps": 500, | |
| "global_step": 6932, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0005770340450086555, | |
| "grad_norm": 1.3359850523956383, | |
| "learning_rate": 2.8818443804034583e-08, | |
| "loss": 0.656, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0028851702250432777, | |
| "grad_norm": 1.415611871814656, | |
| "learning_rate": 1.4409221902017293e-07, | |
| "loss": 0.6481, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.005770340450086555, | |
| "grad_norm": 1.212781925465119, | |
| "learning_rate": 2.8818443804034586e-07, | |
| "loss": 0.6386, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.008655510675129832, | |
| "grad_norm": 1.2361637923329842, | |
| "learning_rate": 4.322766570605188e-07, | |
| "loss": 0.65, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.01154068090017311, | |
| "grad_norm": 1.3799250131478336, | |
| "learning_rate": 5.763688760806917e-07, | |
| "loss": 0.6423, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.014425851125216388, | |
| "grad_norm": 1.303690207496262, | |
| "learning_rate": 7.204610951008646e-07, | |
| "loss": 0.6492, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.017311021350259664, | |
| "grad_norm": 1.2010767582718596, | |
| "learning_rate": 8.645533141210376e-07, | |
| "loss": 0.6254, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.020196191575302943, | |
| "grad_norm": 1.123148811050096, | |
| "learning_rate": 1.0086455331412106e-06, | |
| "loss": 0.632, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.02308136180034622, | |
| "grad_norm": 1.017563618882216, | |
| "learning_rate": 1.1527377521613834e-06, | |
| "loss": 0.6176, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.025966532025389497, | |
| "grad_norm": 0.9670044654743236, | |
| "learning_rate": 1.2968299711815562e-06, | |
| "loss": 0.6237, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.028851702250432775, | |
| "grad_norm": 0.8817997118057925, | |
| "learning_rate": 1.4409221902017292e-06, | |
| "loss": 0.5945, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.031736872475476054, | |
| "grad_norm": 0.7910601022781902, | |
| "learning_rate": 1.5850144092219022e-06, | |
| "loss": 0.5666, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.03462204270051933, | |
| "grad_norm": 0.698866883185103, | |
| "learning_rate": 1.7291066282420752e-06, | |
| "loss": 0.5781, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.03750721292556261, | |
| "grad_norm": 0.7228128874208073, | |
| "learning_rate": 1.873198847262248e-06, | |
| "loss": 0.552, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.040392383150605886, | |
| "grad_norm": 0.6829934171012417, | |
| "learning_rate": 2.0172910662824213e-06, | |
| "loss": 0.5538, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.04327755337564916, | |
| "grad_norm": 0.6501090725880212, | |
| "learning_rate": 2.1613832853025936e-06, | |
| "loss": 0.5428, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.04616272360069244, | |
| "grad_norm": 0.5921206030532731, | |
| "learning_rate": 2.305475504322767e-06, | |
| "loss": 0.5523, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.04904789382573572, | |
| "grad_norm": 0.5843936221943186, | |
| "learning_rate": 2.4495677233429396e-06, | |
| "loss": 0.5295, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.05193306405077899, | |
| "grad_norm": 0.4971214384044829, | |
| "learning_rate": 2.5936599423631124e-06, | |
| "loss": 0.5293, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.054818234275822275, | |
| "grad_norm": 0.502778037587032, | |
| "learning_rate": 2.7377521613832852e-06, | |
| "loss": 0.5192, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.05770340450086555, | |
| "grad_norm": 0.48028251458263926, | |
| "learning_rate": 2.8818443804034585e-06, | |
| "loss": 0.526, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.060588574725908825, | |
| "grad_norm": 0.4958418647779635, | |
| "learning_rate": 3.0259365994236312e-06, | |
| "loss": 0.507, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.06347374495095211, | |
| "grad_norm": 0.4605312245170545, | |
| "learning_rate": 3.1700288184438045e-06, | |
| "loss": 0.5209, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.06635891517599539, | |
| "grad_norm": 0.45269794748416303, | |
| "learning_rate": 3.3141210374639773e-06, | |
| "loss": 0.5018, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.06924408540103866, | |
| "grad_norm": 0.4518223453145298, | |
| "learning_rate": 3.4582132564841505e-06, | |
| "loss": 0.5014, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.07212925562608194, | |
| "grad_norm": 0.4782643771875912, | |
| "learning_rate": 3.602305475504323e-06, | |
| "loss": 0.5152, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.07501442585112522, | |
| "grad_norm": 0.4420045459990657, | |
| "learning_rate": 3.746397694524496e-06, | |
| "loss": 0.4823, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.07789959607616849, | |
| "grad_norm": 0.42381033264166024, | |
| "learning_rate": 3.890489913544669e-06, | |
| "loss": 0.4956, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.08078476630121177, | |
| "grad_norm": 0.4406324685320003, | |
| "learning_rate": 4.0345821325648425e-06, | |
| "loss": 0.4746, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.08366993652625505, | |
| "grad_norm": 0.43316241824454993, | |
| "learning_rate": 4.1786743515850145e-06, | |
| "loss": 0.4746, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.08655510675129832, | |
| "grad_norm": 0.4422617049175986, | |
| "learning_rate": 4.322766570605187e-06, | |
| "loss": 0.4791, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.0894402769763416, | |
| "grad_norm": 0.41863788062040475, | |
| "learning_rate": 4.466858789625361e-06, | |
| "loss": 0.5032, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.09232544720138489, | |
| "grad_norm": 0.4448228403063216, | |
| "learning_rate": 4.610951008645534e-06, | |
| "loss": 0.4984, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.09521061742642815, | |
| "grad_norm": 0.4280125696477632, | |
| "learning_rate": 4.7550432276657065e-06, | |
| "loss": 0.4966, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.09809578765147144, | |
| "grad_norm": 0.3924527289307632, | |
| "learning_rate": 4.899135446685879e-06, | |
| "loss": 0.4707, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.10098095787651472, | |
| "grad_norm": 0.405378064505572, | |
| "learning_rate": 5.043227665706052e-06, | |
| "loss": 0.4684, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.10386612810155799, | |
| "grad_norm": 0.3967929757933598, | |
| "learning_rate": 5.187319884726225e-06, | |
| "loss": 0.4685, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.10675129832660127, | |
| "grad_norm": 0.39908256939712566, | |
| "learning_rate": 5.3314121037463985e-06, | |
| "loss": 0.4777, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.10963646855164455, | |
| "grad_norm": 0.4120577956537649, | |
| "learning_rate": 5.4755043227665705e-06, | |
| "loss": 0.4834, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.11252163877668782, | |
| "grad_norm": 0.3970501930373392, | |
| "learning_rate": 5.619596541786744e-06, | |
| "loss": 0.4698, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.1154068090017311, | |
| "grad_norm": 0.41521712362317653, | |
| "learning_rate": 5.763688760806917e-06, | |
| "loss": 0.4854, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.11829197922677438, | |
| "grad_norm": 0.38302949433328365, | |
| "learning_rate": 5.9077809798270905e-06, | |
| "loss": 0.4603, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.12117714945181765, | |
| "grad_norm": 0.40356716686480254, | |
| "learning_rate": 6.0518731988472625e-06, | |
| "loss": 0.4617, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.12406231967686093, | |
| "grad_norm": 0.4108590792560944, | |
| "learning_rate": 6.195965417867435e-06, | |
| "loss": 0.4821, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.12694748990190421, | |
| "grad_norm": 0.38579700259781685, | |
| "learning_rate": 6.340057636887609e-06, | |
| "loss": 0.4548, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.1298326601269475, | |
| "grad_norm": 0.4008887277935931, | |
| "learning_rate": 6.484149855907782e-06, | |
| "loss": 0.4826, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.13271783035199078, | |
| "grad_norm": 0.4324585130384216, | |
| "learning_rate": 6.6282420749279545e-06, | |
| "loss": 0.4524, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.13560300057703403, | |
| "grad_norm": 0.36547940995267386, | |
| "learning_rate": 6.772334293948127e-06, | |
| "loss": 0.4472, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.13848817080207732, | |
| "grad_norm": 0.4021304207485496, | |
| "learning_rate": 6.916426512968301e-06, | |
| "loss": 0.469, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.1413733410271206, | |
| "grad_norm": 0.3986364193440875, | |
| "learning_rate": 7.060518731988473e-06, | |
| "loss": 0.4664, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.14425851125216388, | |
| "grad_norm": 0.3894372801996192, | |
| "learning_rate": 7.204610951008646e-06, | |
| "loss": 0.4596, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.14714368147720716, | |
| "grad_norm": 0.3822307947159079, | |
| "learning_rate": 7.348703170028819e-06, | |
| "loss": 0.4611, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.15002885170225044, | |
| "grad_norm": 0.3770508093447163, | |
| "learning_rate": 7.492795389048992e-06, | |
| "loss": 0.4513, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.1529140219272937, | |
| "grad_norm": 0.4040420054739231, | |
| "learning_rate": 7.636887608069165e-06, | |
| "loss": 0.4404, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.15579919215233698, | |
| "grad_norm": 0.3763215501882052, | |
| "learning_rate": 7.780979827089338e-06, | |
| "loss": 0.4484, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.15868436237738026, | |
| "grad_norm": 0.3823790521602452, | |
| "learning_rate": 7.92507204610951e-06, | |
| "loss": 0.4568, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.16156953260242354, | |
| "grad_norm": 0.3732843632883636, | |
| "learning_rate": 8.069164265129685e-06, | |
| "loss": 0.4389, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.16445470282746683, | |
| "grad_norm": 0.37767577343237896, | |
| "learning_rate": 8.213256484149856e-06, | |
| "loss": 0.4572, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.1673398730525101, | |
| "grad_norm": 0.3980275526344354, | |
| "learning_rate": 8.357348703170029e-06, | |
| "loss": 0.4625, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.1702250432775534, | |
| "grad_norm": 0.4046607258650182, | |
| "learning_rate": 8.501440922190203e-06, | |
| "loss": 0.488, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.17311021350259664, | |
| "grad_norm": 0.3759989273900334, | |
| "learning_rate": 8.645533141210375e-06, | |
| "loss": 0.4696, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.17599538372763993, | |
| "grad_norm": 0.3782703272325284, | |
| "learning_rate": 8.789625360230547e-06, | |
| "loss": 0.434, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.1788805539526832, | |
| "grad_norm": 0.3906713039092043, | |
| "learning_rate": 8.933717579250722e-06, | |
| "loss": 0.4565, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.1817657241777265, | |
| "grad_norm": 0.3805412690860938, | |
| "learning_rate": 9.077809798270895e-06, | |
| "loss": 0.448, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.18465089440276977, | |
| "grad_norm": 0.39402776095093084, | |
| "learning_rate": 9.221902017291067e-06, | |
| "loss": 0.4553, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.18753606462781305, | |
| "grad_norm": 0.45284451101047796, | |
| "learning_rate": 9.36599423631124e-06, | |
| "loss": 0.4447, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.1904212348528563, | |
| "grad_norm": 0.3873105480189736, | |
| "learning_rate": 9.510086455331413e-06, | |
| "loss": 0.4499, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.1933064050778996, | |
| "grad_norm": 0.40199355492029853, | |
| "learning_rate": 9.654178674351586e-06, | |
| "loss": 0.454, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.19619157530294287, | |
| "grad_norm": 0.38330387655003734, | |
| "learning_rate": 9.798270893371759e-06, | |
| "loss": 0.4484, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.19907674552798615, | |
| "grad_norm": 0.3790700593281963, | |
| "learning_rate": 9.942363112391931e-06, | |
| "loss": 0.4693, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.20196191575302944, | |
| "grad_norm": 0.4091985418093552, | |
| "learning_rate": 1.0086455331412104e-05, | |
| "loss": 0.4638, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.20484708597807272, | |
| "grad_norm": 0.3816167638203252, | |
| "learning_rate": 1.0230547550432277e-05, | |
| "loss": 0.4529, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.20773225620311597, | |
| "grad_norm": 0.3705496517450964, | |
| "learning_rate": 1.037463976945245e-05, | |
| "loss": 0.4491, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.21061742642815925, | |
| "grad_norm": 0.4096345206746822, | |
| "learning_rate": 1.0518731988472624e-05, | |
| "loss": 0.4769, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.21350259665320254, | |
| "grad_norm": 0.3828300212257734, | |
| "learning_rate": 1.0662824207492797e-05, | |
| "loss": 0.446, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.21638776687824582, | |
| "grad_norm": 0.3664642342578931, | |
| "learning_rate": 1.080691642651297e-05, | |
| "loss": 0.4629, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.2192729371032891, | |
| "grad_norm": 0.3701318700670979, | |
| "learning_rate": 1.0951008645533141e-05, | |
| "loss": 0.4454, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.22215810732833238, | |
| "grad_norm": 0.3765944116998475, | |
| "learning_rate": 1.1095100864553314e-05, | |
| "loss": 0.445, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.22504327755337564, | |
| "grad_norm": 0.35848655613640135, | |
| "learning_rate": 1.1239193083573488e-05, | |
| "loss": 0.4482, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.22792844777841892, | |
| "grad_norm": 0.3522345707962982, | |
| "learning_rate": 1.1383285302593661e-05, | |
| "loss": 0.4488, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.2308136180034622, | |
| "grad_norm": 0.42916841303669334, | |
| "learning_rate": 1.1527377521613834e-05, | |
| "loss": 0.4815, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.23369878822850548, | |
| "grad_norm": 0.39333826210382833, | |
| "learning_rate": 1.1671469740634007e-05, | |
| "loss": 0.4479, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.23658395845354876, | |
| "grad_norm": 0.395568573594799, | |
| "learning_rate": 1.1815561959654181e-05, | |
| "loss": 0.4375, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.23946912867859205, | |
| "grad_norm": 0.36804463353795325, | |
| "learning_rate": 1.1959654178674354e-05, | |
| "loss": 0.4544, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.2423542989036353, | |
| "grad_norm": 0.4057948653659005, | |
| "learning_rate": 1.2103746397694525e-05, | |
| "loss": 0.4489, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.24523946912867858, | |
| "grad_norm": 0.4042926809090149, | |
| "learning_rate": 1.2247838616714698e-05, | |
| "loss": 0.4432, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.24812463935372187, | |
| "grad_norm": 0.3746660432560446, | |
| "learning_rate": 1.239193083573487e-05, | |
| "loss": 0.454, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.2510098095787651, | |
| "grad_norm": 0.39348997561248156, | |
| "learning_rate": 1.2536023054755045e-05, | |
| "loss": 0.4647, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.25389497980380843, | |
| "grad_norm": 0.3872848928403898, | |
| "learning_rate": 1.2680115273775218e-05, | |
| "loss": 0.4535, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.2567801500288517, | |
| "grad_norm": 0.3813359635416143, | |
| "learning_rate": 1.282420749279539e-05, | |
| "loss": 0.4508, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.259665320253895, | |
| "grad_norm": 0.40132785614459726, | |
| "learning_rate": 1.2968299711815563e-05, | |
| "loss": 0.4549, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.26255049047893825, | |
| "grad_norm": 0.3674608489689301, | |
| "learning_rate": 1.3112391930835735e-05, | |
| "loss": 0.4519, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.26543566070398156, | |
| "grad_norm": 0.3672381265982974, | |
| "learning_rate": 1.3256484149855909e-05, | |
| "loss": 0.4504, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.2683208309290248, | |
| "grad_norm": 0.41466539433993804, | |
| "learning_rate": 1.3400576368876082e-05, | |
| "loss": 0.4509, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.27120600115406807, | |
| "grad_norm": 0.37606450867708474, | |
| "learning_rate": 1.3544668587896255e-05, | |
| "loss": 0.4646, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.2740911713791114, | |
| "grad_norm": 0.3868252718025835, | |
| "learning_rate": 1.3688760806916427e-05, | |
| "loss": 0.4395, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.27697634160415463, | |
| "grad_norm": 0.36956730373024904, | |
| "learning_rate": 1.3832853025936602e-05, | |
| "loss": 0.4421, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.27986151182919794, | |
| "grad_norm": 0.3703399882804596, | |
| "learning_rate": 1.3976945244956775e-05, | |
| "loss": 0.4288, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.2827466820542412, | |
| "grad_norm": 0.4006912338102629, | |
| "learning_rate": 1.4121037463976946e-05, | |
| "loss": 0.4721, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.2856318522792845, | |
| "grad_norm": 0.3443154071927776, | |
| "learning_rate": 1.4265129682997119e-05, | |
| "loss": 0.4322, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.28851702250432776, | |
| "grad_norm": 0.37915946639466674, | |
| "learning_rate": 1.4409221902017291e-05, | |
| "loss": 0.4264, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.291402192729371, | |
| "grad_norm": 0.38032216321312473, | |
| "learning_rate": 1.4553314121037466e-05, | |
| "loss": 0.4408, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.2942873629544143, | |
| "grad_norm": 0.40009286504791697, | |
| "learning_rate": 1.4697406340057639e-05, | |
| "loss": 0.4616, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.2971725331794576, | |
| "grad_norm": 0.35910974753646935, | |
| "learning_rate": 1.4841498559077811e-05, | |
| "loss": 0.4306, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.3000577034045009, | |
| "grad_norm": 0.37319150994773553, | |
| "learning_rate": 1.4985590778097984e-05, | |
| "loss": 0.4497, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.30294287362954414, | |
| "grad_norm": 0.3637719085871802, | |
| "learning_rate": 1.5129682997118155e-05, | |
| "loss": 0.4551, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.3058280438545874, | |
| "grad_norm": 0.3575259753574912, | |
| "learning_rate": 1.527377521613833e-05, | |
| "loss": 0.447, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.3087132140796307, | |
| "grad_norm": 0.3596026864789941, | |
| "learning_rate": 1.54178674351585e-05, | |
| "loss": 0.4642, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.31159838430467396, | |
| "grad_norm": 0.3613487147969066, | |
| "learning_rate": 1.5561959654178675e-05, | |
| "loss": 0.4384, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.31448355452971727, | |
| "grad_norm": 0.5248264510992934, | |
| "learning_rate": 1.570605187319885e-05, | |
| "loss": 0.4427, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.3173687247547605, | |
| "grad_norm": 0.3648196713346363, | |
| "learning_rate": 1.585014409221902e-05, | |
| "loss": 0.4649, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.32025389497980383, | |
| "grad_norm": 0.40066795102765307, | |
| "learning_rate": 1.5994236311239196e-05, | |
| "loss": 0.439, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.3231390652048471, | |
| "grad_norm": 0.33951584959069336, | |
| "learning_rate": 1.613832853025937e-05, | |
| "loss": 0.4432, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.32602423542989034, | |
| "grad_norm": 0.3533117973734279, | |
| "learning_rate": 1.6282420749279538e-05, | |
| "loss": 0.4335, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.32890940565493365, | |
| "grad_norm": 0.37863111987341347, | |
| "learning_rate": 1.6426512968299712e-05, | |
| "loss": 0.428, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.3317945758799769, | |
| "grad_norm": 0.4346675097845832, | |
| "learning_rate": 1.6570605187319887e-05, | |
| "loss": 0.449, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.3346797461050202, | |
| "grad_norm": 0.37817008201494773, | |
| "learning_rate": 1.6714697406340058e-05, | |
| "loss": 0.4433, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.33756491633006347, | |
| "grad_norm": 0.3808322285847144, | |
| "learning_rate": 1.6858789625360232e-05, | |
| "loss": 0.4466, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.3404500865551068, | |
| "grad_norm": 0.37923635129260924, | |
| "learning_rate": 1.7002881844380407e-05, | |
| "loss": 0.4476, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.34333525678015003, | |
| "grad_norm": 0.3823099849897589, | |
| "learning_rate": 1.7146974063400578e-05, | |
| "loss": 0.4323, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.3462204270051933, | |
| "grad_norm": 0.3818824327579753, | |
| "learning_rate": 1.729106628242075e-05, | |
| "loss": 0.4375, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.3491055972302366, | |
| "grad_norm": 0.37210993394094566, | |
| "learning_rate": 1.7435158501440924e-05, | |
| "loss": 0.4305, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.35199076745527985, | |
| "grad_norm": 0.3465073067639098, | |
| "learning_rate": 1.7579250720461095e-05, | |
| "loss": 0.4255, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.35487593768032316, | |
| "grad_norm": 0.38378004625397016, | |
| "learning_rate": 1.772334293948127e-05, | |
| "loss": 0.4324, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.3577611079053664, | |
| "grad_norm": 0.37432578456970733, | |
| "learning_rate": 1.7867435158501444e-05, | |
| "loss": 0.4543, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.36064627813040967, | |
| "grad_norm": 0.3341493702083905, | |
| "learning_rate": 1.8011527377521615e-05, | |
| "loss": 0.4501, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.363531448355453, | |
| "grad_norm": 0.3618443471925012, | |
| "learning_rate": 1.815561959654179e-05, | |
| "loss": 0.4387, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.36641661858049623, | |
| "grad_norm": 0.3493534570885224, | |
| "learning_rate": 1.829971181556196e-05, | |
| "loss": 0.4483, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.36930178880553954, | |
| "grad_norm": 0.37866953974278367, | |
| "learning_rate": 1.8443804034582135e-05, | |
| "loss": 0.4289, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.3721869590305828, | |
| "grad_norm": 0.37227424777091433, | |
| "learning_rate": 1.8587896253602306e-05, | |
| "loss": 0.4306, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.3750721292556261, | |
| "grad_norm": 0.4014872351370534, | |
| "learning_rate": 1.873198847262248e-05, | |
| "loss": 0.4414, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.37795729948066936, | |
| "grad_norm": 0.3724847659964483, | |
| "learning_rate": 1.887608069164265e-05, | |
| "loss": 0.4446, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.3808424697057126, | |
| "grad_norm": 0.3783020221193775, | |
| "learning_rate": 1.9020172910662826e-05, | |
| "loss": 0.4608, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.3837276399307559, | |
| "grad_norm": 0.4307776328223203, | |
| "learning_rate": 1.9164265129683e-05, | |
| "loss": 0.4343, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.3866128101557992, | |
| "grad_norm": 0.37291676085751585, | |
| "learning_rate": 1.930835734870317e-05, | |
| "loss": 0.4263, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.3894979803808425, | |
| "grad_norm": 0.3765824744528829, | |
| "learning_rate": 1.9452449567723343e-05, | |
| "loss": 0.4258, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.39238315060588574, | |
| "grad_norm": 0.354531651779117, | |
| "learning_rate": 1.9596541786743517e-05, | |
| "loss": 0.4504, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.395268320830929, | |
| "grad_norm": 0.3695419929782549, | |
| "learning_rate": 1.974063400576369e-05, | |
| "loss": 0.4663, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.3981534910559723, | |
| "grad_norm": 0.3752799655893121, | |
| "learning_rate": 1.9884726224783863e-05, | |
| "loss": 0.4519, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.40103866128101556, | |
| "grad_norm": 0.3743552771364968, | |
| "learning_rate": 1.9999998731825553e-05, | |
| "loss": 0.4466, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.4039238315060589, | |
| "grad_norm": 0.3690677017067688, | |
| "learning_rate": 1.9999954345753522e-05, | |
| "loss": 0.4672, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.4068090017311021, | |
| "grad_norm": 0.3521337986777979, | |
| "learning_rate": 1.9999846551280566e-05, | |
| "loss": 0.4546, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.40969417195614544, | |
| "grad_norm": 0.371125205175818, | |
| "learning_rate": 1.99996753490902e-05, | |
| "loss": 0.4476, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.4125793421811887, | |
| "grad_norm": 0.3870027577034139, | |
| "learning_rate": 1.999944074026799e-05, | |
| "loss": 0.4528, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.41546451240623195, | |
| "grad_norm": 0.3694599366137213, | |
| "learning_rate": 1.999914272630156e-05, | |
| "loss": 0.4435, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.41834968263127525, | |
| "grad_norm": 0.39229103038261187, | |
| "learning_rate": 1.999878130908058e-05, | |
| "loss": 0.4624, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.4212348528563185, | |
| "grad_norm": 0.37197227629586244, | |
| "learning_rate": 1.999835649089675e-05, | |
| "loss": 0.4392, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.4241200230813618, | |
| "grad_norm": 0.36658551403514933, | |
| "learning_rate": 1.9997868274443787e-05, | |
| "loss": 0.4479, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.4270051933064051, | |
| "grad_norm": 0.38507395642784864, | |
| "learning_rate": 1.9997316662817403e-05, | |
| "loss": 0.4448, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.4298903635314484, | |
| "grad_norm": 0.36930658832139956, | |
| "learning_rate": 1.9996701659515293e-05, | |
| "loss": 0.4338, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.43277553375649164, | |
| "grad_norm": 0.3415004408124957, | |
| "learning_rate": 1.999602326843712e-05, | |
| "loss": 0.4571, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.4356607039815349, | |
| "grad_norm": 0.3616392339336591, | |
| "learning_rate": 1.999528149388447e-05, | |
| "loss": 0.4364, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.4385458742065782, | |
| "grad_norm": 0.3614440903798367, | |
| "learning_rate": 1.999447634056084e-05, | |
| "loss": 0.442, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.44143104443162146, | |
| "grad_norm": 0.3475594464185632, | |
| "learning_rate": 1.9993607813571595e-05, | |
| "loss": 0.4302, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.44431621465666477, | |
| "grad_norm": 0.38907218952119343, | |
| "learning_rate": 1.9992675918423958e-05, | |
| "loss": 0.4475, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.447201384881708, | |
| "grad_norm": 0.3532869868641272, | |
| "learning_rate": 1.999168066102695e-05, | |
| "loss": 0.4486, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.4500865551067513, | |
| "grad_norm": 0.37651731293442015, | |
| "learning_rate": 1.9990622047691372e-05, | |
| "loss": 0.4326, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.4529717253317946, | |
| "grad_norm": 0.3748786150498863, | |
| "learning_rate": 1.998950008512975e-05, | |
| "loss": 0.4474, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.45585689555683784, | |
| "grad_norm": 0.39902509681503917, | |
| "learning_rate": 1.9988314780456305e-05, | |
| "loss": 0.4568, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.45874206578188115, | |
| "grad_norm": 0.3884381984087952, | |
| "learning_rate": 1.998706614118689e-05, | |
| "loss": 0.4568, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.4616272360069244, | |
| "grad_norm": 0.36054442749164206, | |
| "learning_rate": 1.9985754175238972e-05, | |
| "loss": 0.4429, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.4645124062319677, | |
| "grad_norm": 0.3527339288196345, | |
| "learning_rate": 1.998437889093155e-05, | |
| "loss": 0.4437, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.46739757645701097, | |
| "grad_norm": 0.3737014283866458, | |
| "learning_rate": 1.9982940296985124e-05, | |
| "loss": 0.4469, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.4702827466820542, | |
| "grad_norm": 0.3836320689851165, | |
| "learning_rate": 1.998143840252163e-05, | |
| "loss": 0.4391, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.47316791690709753, | |
| "grad_norm": 0.36567112690649783, | |
| "learning_rate": 1.997987321706438e-05, | |
| "loss": 0.4414, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.4760530871321408, | |
| "grad_norm": 0.37388859283028064, | |
| "learning_rate": 1.9978244750538017e-05, | |
| "loss": 0.4408, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.4789382573571841, | |
| "grad_norm": 0.35288917666405584, | |
| "learning_rate": 1.9976553013268428e-05, | |
| "loss": 0.4432, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.48182342758222735, | |
| "grad_norm": 0.35467525608999173, | |
| "learning_rate": 1.9974798015982704e-05, | |
| "loss": 0.4319, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.4847085978072706, | |
| "grad_norm": 0.36246807797824815, | |
| "learning_rate": 1.9972979769809044e-05, | |
| "loss": 0.4421, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.4875937680323139, | |
| "grad_norm": 0.38809255131157183, | |
| "learning_rate": 1.9971098286276716e-05, | |
| "loss": 0.438, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.49047893825735717, | |
| "grad_norm": 0.35620554170658003, | |
| "learning_rate": 1.996915357731596e-05, | |
| "loss": 0.4361, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.4933641084824005, | |
| "grad_norm": 0.37319706651769974, | |
| "learning_rate": 1.996714565525792e-05, | |
| "loss": 0.4273, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.49624927870744373, | |
| "grad_norm": 0.3561234532855242, | |
| "learning_rate": 1.996507453283456e-05, | |
| "loss": 0.4405, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.49913444893248704, | |
| "grad_norm": 0.3599439467426677, | |
| "learning_rate": 1.9962940223178607e-05, | |
| "loss": 0.4209, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.5020196191575302, | |
| "grad_norm": 0.3708191889641566, | |
| "learning_rate": 1.9960742739823433e-05, | |
| "loss": 0.4375, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.5049047893825735, | |
| "grad_norm": 0.371986303129069, | |
| "learning_rate": 1.9958482096702997e-05, | |
| "loss": 0.44, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.5077899596076169, | |
| "grad_norm": 0.37613060788442565, | |
| "learning_rate": 1.995615830815173e-05, | |
| "loss": 0.444, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.5106751298326602, | |
| "grad_norm": 0.35285659183130846, | |
| "learning_rate": 1.9953771388904484e-05, | |
| "loss": 0.4483, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.5135603000577034, | |
| "grad_norm": 0.3744385315940648, | |
| "learning_rate": 1.9951321354096395e-05, | |
| "loss": 0.4307, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.5164454702827467, | |
| "grad_norm": 0.3527685970235997, | |
| "learning_rate": 1.9948808219262813e-05, | |
| "loss": 0.4352, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.51933064050779, | |
| "grad_norm": 0.3629898148956746, | |
| "learning_rate": 1.9946232000339192e-05, | |
| "loss": 0.4325, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.5222158107328332, | |
| "grad_norm": 0.33230153297658693, | |
| "learning_rate": 1.994359271366101e-05, | |
| "loss": 0.4358, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.5251009809578765, | |
| "grad_norm": 0.36872688475457055, | |
| "learning_rate": 1.9940890375963634e-05, | |
| "loss": 0.4389, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.5279861511829198, | |
| "grad_norm": 0.3500504418913126, | |
| "learning_rate": 1.9938125004382226e-05, | |
| "loss": 0.4368, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.5308713214079631, | |
| "grad_norm": 0.38076969493807267, | |
| "learning_rate": 1.9935296616451654e-05, | |
| "loss": 0.4224, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.5337564916330063, | |
| "grad_norm": 0.38809106802534876, | |
| "learning_rate": 1.9932405230106352e-05, | |
| "loss": 0.458, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.5366416618580496, | |
| "grad_norm": 0.34830821561529673, | |
| "learning_rate": 1.992945086368022e-05, | |
| "loss": 0.4427, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.5395268320830929, | |
| "grad_norm": 0.36348521739778694, | |
| "learning_rate": 1.9926433535906514e-05, | |
| "loss": 0.4332, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.5424120023081361, | |
| "grad_norm": 0.3516690545778123, | |
| "learning_rate": 1.992335326591771e-05, | |
| "loss": 0.4493, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.5452971725331794, | |
| "grad_norm": 0.3622130880303389, | |
| "learning_rate": 1.99202100732454e-05, | |
| "loss": 0.4399, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.5481823427582228, | |
| "grad_norm": 0.3340397775809456, | |
| "learning_rate": 1.9917003977820154e-05, | |
| "loss": 0.4155, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.5510675129832661, | |
| "grad_norm": 0.3373549627806811, | |
| "learning_rate": 1.9913734999971402e-05, | |
| "loss": 0.4343, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.5539526832083093, | |
| "grad_norm": 0.34945896571648644, | |
| "learning_rate": 1.9910403160427308e-05, | |
| "loss": 0.4423, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.5568378534333526, | |
| "grad_norm": 0.3477598210882681, | |
| "learning_rate": 1.990700848031463e-05, | |
| "loss": 0.4328, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.5597230236583959, | |
| "grad_norm": 0.3728272868907162, | |
| "learning_rate": 1.990355098115859e-05, | |
| "loss": 0.4407, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.5626081938834391, | |
| "grad_norm": 0.36540010757747865, | |
| "learning_rate": 1.990003068488273e-05, | |
| "loss": 0.4254, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.5654933641084824, | |
| "grad_norm": 0.3491338839333341, | |
| "learning_rate": 1.9896447613808802e-05, | |
| "loss": 0.4333, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.5683785343335257, | |
| "grad_norm": 0.3453766808693869, | |
| "learning_rate": 1.9892801790656575e-05, | |
| "loss": 0.4505, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.571263704558569, | |
| "grad_norm": 0.3663489965490133, | |
| "learning_rate": 1.9889093238543746e-05, | |
| "loss": 0.4308, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.5741488747836122, | |
| "grad_norm": 0.3679457848990051, | |
| "learning_rate": 1.9885321980985758e-05, | |
| "loss": 0.4594, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.5770340450086555, | |
| "grad_norm": 0.3780068047990749, | |
| "learning_rate": 1.988148804189566e-05, | |
| "loss": 0.4255, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.5799192152336988, | |
| "grad_norm": 0.36727910455394897, | |
| "learning_rate": 1.987759144558395e-05, | |
| "loss": 0.4378, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.582804385458742, | |
| "grad_norm": 0.36166322845028065, | |
| "learning_rate": 1.9873632216758448e-05, | |
| "loss": 0.4445, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.5856895556837853, | |
| "grad_norm": 0.3412780788439251, | |
| "learning_rate": 1.9869610380524098e-05, | |
| "loss": 0.4285, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.5885747259088286, | |
| "grad_norm": 0.3642868390810403, | |
| "learning_rate": 1.986552596238284e-05, | |
| "loss": 0.426, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.5914598961338718, | |
| "grad_norm": 0.3513636822125392, | |
| "learning_rate": 1.986137898823343e-05, | |
| "loss": 0.4689, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.5943450663589152, | |
| "grad_norm": 0.37061082820118274, | |
| "learning_rate": 1.985716948437129e-05, | |
| "loss": 0.4396, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.5972302365839585, | |
| "grad_norm": 0.35578918227279527, | |
| "learning_rate": 1.9852897477488335e-05, | |
| "loss": 0.4347, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.6001154068090018, | |
| "grad_norm": 0.37785556076696936, | |
| "learning_rate": 1.9848562994672797e-05, | |
| "loss": 0.4272, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.603000577034045, | |
| "grad_norm": 0.36978022207552363, | |
| "learning_rate": 1.984416606340907e-05, | |
| "loss": 0.4326, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.6058857472590883, | |
| "grad_norm": 0.3791504839723979, | |
| "learning_rate": 1.9839706711577514e-05, | |
| "loss": 0.4587, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.6087709174841316, | |
| "grad_norm": 0.3710313865819546, | |
| "learning_rate": 1.98351849674543e-05, | |
| "loss": 0.457, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.6116560877091748, | |
| "grad_norm": 0.35231701107663654, | |
| "learning_rate": 1.9830600859711207e-05, | |
| "loss": 0.4144, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.6145412579342181, | |
| "grad_norm": 0.3615898765366417, | |
| "learning_rate": 1.982595441741547e-05, | |
| "loss": 0.4401, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.6174264281592614, | |
| "grad_norm": 0.35977723996205, | |
| "learning_rate": 1.9821245670029566e-05, | |
| "loss": 0.4309, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.6203115983843047, | |
| "grad_norm": 0.36669507423796843, | |
| "learning_rate": 1.9816474647411047e-05, | |
| "loss": 0.4383, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.6231967686093479, | |
| "grad_norm": 0.35392661751275434, | |
| "learning_rate": 1.9811641379812342e-05, | |
| "loss": 0.446, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.6260819388343912, | |
| "grad_norm": 0.3409833418281117, | |
| "learning_rate": 1.9806745897880573e-05, | |
| "loss": 0.4367, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.6289671090594345, | |
| "grad_norm": 0.37150843192119876, | |
| "learning_rate": 1.9801788232657343e-05, | |
| "loss": 0.4527, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.6318522792844777, | |
| "grad_norm": 0.36569227925348624, | |
| "learning_rate": 1.9796768415578564e-05, | |
| "loss": 0.4436, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.634737449509521, | |
| "grad_norm": 0.35016739834472266, | |
| "learning_rate": 1.9791686478474235e-05, | |
| "loss": 0.4368, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.6376226197345644, | |
| "grad_norm": 0.36715520059954826, | |
| "learning_rate": 1.978654245356825e-05, | |
| "loss": 0.4354, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.6405077899596077, | |
| "grad_norm": 0.3840761403084782, | |
| "learning_rate": 1.9781336373478207e-05, | |
| "loss": 0.4523, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.6433929601846509, | |
| "grad_norm": 0.3154658430275838, | |
| "learning_rate": 1.9776068271215167e-05, | |
| "loss": 0.4406, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.6462781304096942, | |
| "grad_norm": 0.3772719849710032, | |
| "learning_rate": 1.9770738180183485e-05, | |
| "loss": 0.4344, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.6491633006347375, | |
| "grad_norm": 0.3381456770156173, | |
| "learning_rate": 1.9765346134180567e-05, | |
| "loss": 0.4324, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.6520484708597807, | |
| "grad_norm": 0.3498102184043816, | |
| "learning_rate": 1.975989216739667e-05, | |
| "loss": 0.4294, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.654933641084824, | |
| "grad_norm": 0.3449183171225279, | |
| "learning_rate": 1.9754376314414686e-05, | |
| "loss": 0.4139, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.6578188113098673, | |
| "grad_norm": 0.3541093995239644, | |
| "learning_rate": 1.9748798610209912e-05, | |
| "loss": 0.4295, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.6607039815349106, | |
| "grad_norm": 0.36096527367909803, | |
| "learning_rate": 1.9743159090149844e-05, | |
| "loss": 0.4605, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.6635891517599538, | |
| "grad_norm": 0.334233191996567, | |
| "learning_rate": 1.9737457789993936e-05, | |
| "loss": 0.4368, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.6664743219849971, | |
| "grad_norm": 0.35974161365839685, | |
| "learning_rate": 1.973169474589339e-05, | |
| "loss": 0.428, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.6693594922100404, | |
| "grad_norm": 0.3568705187506466, | |
| "learning_rate": 1.9725869994390912e-05, | |
| "loss": 0.4417, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.6722446624350836, | |
| "grad_norm": 0.33809479002108944, | |
| "learning_rate": 1.971998357242049e-05, | |
| "loss": 0.4413, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.6751298326601269, | |
| "grad_norm": 0.3502193821967476, | |
| "learning_rate": 1.971403551730715e-05, | |
| "loss": 0.4564, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.6780150028851702, | |
| "grad_norm": 0.35976087067679435, | |
| "learning_rate": 1.9708025866766735e-05, | |
| "loss": 0.4263, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.6809001731102136, | |
| "grad_norm": 0.34825014902921947, | |
| "learning_rate": 1.9701954658905652e-05, | |
| "loss": 0.4324, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.6837853433352568, | |
| "grad_norm": 0.3515962597578983, | |
| "learning_rate": 1.9695821932220632e-05, | |
| "loss": 0.4347, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.6866705135603001, | |
| "grad_norm": 0.3518698997384743, | |
| "learning_rate": 1.9689627725598495e-05, | |
| "loss": 0.4141, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.6895556837853434, | |
| "grad_norm": 0.3406825005378319, | |
| "learning_rate": 1.9683372078315892e-05, | |
| "loss": 0.4294, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.6924408540103866, | |
| "grad_norm": 0.3555369233516162, | |
| "learning_rate": 1.9677055030039067e-05, | |
| "loss": 0.4389, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.6953260242354299, | |
| "grad_norm": 0.4609971974027847, | |
| "learning_rate": 1.967067662082359e-05, | |
| "loss": 0.4479, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.6982111944604732, | |
| "grad_norm": 0.3533102697284043, | |
| "learning_rate": 1.966423689111412e-05, | |
| "loss": 0.4594, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.7010963646855164, | |
| "grad_norm": 0.41163003146138266, | |
| "learning_rate": 1.965773588174414e-05, | |
| "loss": 0.4211, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.7039815349105597, | |
| "grad_norm": 0.36355249316504007, | |
| "learning_rate": 1.9651173633935702e-05, | |
| "loss": 0.4326, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.706866705135603, | |
| "grad_norm": 0.361512663082872, | |
| "learning_rate": 1.964455018929916e-05, | |
| "loss": 0.4605, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.7097518753606463, | |
| "grad_norm": 0.3395334945396585, | |
| "learning_rate": 1.9637865589832902e-05, | |
| "loss": 0.4552, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.7126370455856895, | |
| "grad_norm": 0.3358166664046763, | |
| "learning_rate": 1.963111987792311e-05, | |
| "loss": 0.4418, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.7155222158107328, | |
| "grad_norm": 0.35952927832201437, | |
| "learning_rate": 1.9624313096343445e-05, | |
| "loss": 0.4528, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.7184073860357761, | |
| "grad_norm": 0.3516252150445922, | |
| "learning_rate": 1.961744528825483e-05, | |
| "loss": 0.4444, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.7212925562608193, | |
| "grad_norm": 0.37486390370155487, | |
| "learning_rate": 1.961051649720513e-05, | |
| "loss": 0.4405, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.7241777264858626, | |
| "grad_norm": 0.347168899689223, | |
| "learning_rate": 1.96035267671289e-05, | |
| "loss": 0.4197, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.727062896710906, | |
| "grad_norm": 0.3553448619523881, | |
| "learning_rate": 1.9596476142347114e-05, | |
| "loss": 0.4273, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.7299480669359493, | |
| "grad_norm": 0.3747743065569822, | |
| "learning_rate": 1.958936466756685e-05, | |
| "loss": 0.4318, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.7328332371609925, | |
| "grad_norm": 0.36946705338179187, | |
| "learning_rate": 1.9582192387881037e-05, | |
| "loss": 0.439, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.7357184073860358, | |
| "grad_norm": 0.3535008834153477, | |
| "learning_rate": 1.9574959348768173e-05, | |
| "loss": 0.4215, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.7386035776110791, | |
| "grad_norm": 0.3528743884422751, | |
| "learning_rate": 1.9567665596092e-05, | |
| "loss": 0.4451, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.7414887478361223, | |
| "grad_norm": 0.36361502852826977, | |
| "learning_rate": 1.9560311176101253e-05, | |
| "loss": 0.4488, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.7443739180611656, | |
| "grad_norm": 0.3393383444681419, | |
| "learning_rate": 1.9552896135429346e-05, | |
| "loss": 0.4602, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.7472590882862089, | |
| "grad_norm": 0.3504869096195613, | |
| "learning_rate": 1.9545420521094082e-05, | |
| "loss": 0.4213, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.7501442585112522, | |
| "grad_norm": 0.33568870619524627, | |
| "learning_rate": 1.953788438049735e-05, | |
| "loss": 0.4159, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.7530294287362954, | |
| "grad_norm": 0.349599888573523, | |
| "learning_rate": 1.953028776142483e-05, | |
| "loss": 0.4245, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.7559145989613387, | |
| "grad_norm": 0.36044335412116235, | |
| "learning_rate": 1.9522630712045695e-05, | |
| "loss": 0.4253, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.758799769186382, | |
| "grad_norm": 0.33120277929093034, | |
| "learning_rate": 1.9514913280912284e-05, | |
| "loss": 0.4196, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.7616849394114252, | |
| "grad_norm": 0.3556004564806594, | |
| "learning_rate": 1.9507135516959824e-05, | |
| "loss": 0.4319, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.7645701096364685, | |
| "grad_norm": 0.357624065192376, | |
| "learning_rate": 1.949929746950609e-05, | |
| "loss": 0.4464, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.7674552798615119, | |
| "grad_norm": 0.3429194439395415, | |
| "learning_rate": 1.9491399188251123e-05, | |
| "loss": 0.4384, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.7703404500865552, | |
| "grad_norm": 0.35807803811180317, | |
| "learning_rate": 1.9483440723276887e-05, | |
| "loss": 0.4274, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.7732256203115984, | |
| "grad_norm": 0.3516485111182853, | |
| "learning_rate": 1.9475422125046962e-05, | |
| "loss": 0.4314, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.7761107905366417, | |
| "grad_norm": 0.3448017580215496, | |
| "learning_rate": 1.9467343444406234e-05, | |
| "loss": 0.4305, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.778995960761685, | |
| "grad_norm": 0.33719191373319685, | |
| "learning_rate": 1.9459204732580557e-05, | |
| "loss": 0.4209, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.7818811309867282, | |
| "grad_norm": 0.357112111873714, | |
| "learning_rate": 1.9451006041176437e-05, | |
| "loss": 0.4614, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.7847663012117715, | |
| "grad_norm": 0.3638827997486545, | |
| "learning_rate": 1.9442747422180704e-05, | |
| "loss": 0.4392, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.7876514714368148, | |
| "grad_norm": 0.3779229355078506, | |
| "learning_rate": 1.9434428927960177e-05, | |
| "loss": 0.4175, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.790536641661858, | |
| "grad_norm": 0.33870833738392225, | |
| "learning_rate": 1.9426050611261343e-05, | |
| "loss": 0.4378, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.7934218118869013, | |
| "grad_norm": 0.3407658218095889, | |
| "learning_rate": 1.9417612525210007e-05, | |
| "loss": 0.4244, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.7963069821119446, | |
| "grad_norm": 0.3587456861468758, | |
| "learning_rate": 1.9409114723310967e-05, | |
| "loss": 0.4296, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.7991921523369879, | |
| "grad_norm": 0.3526548403605731, | |
| "learning_rate": 1.9400557259447673e-05, | |
| "loss": 0.4334, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.8020773225620311, | |
| "grad_norm": 0.35126996708797126, | |
| "learning_rate": 1.9391940187881885e-05, | |
| "loss": 0.4455, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.8049624927870744, | |
| "grad_norm": 0.35018120101424355, | |
| "learning_rate": 1.9383263563253322e-05, | |
| "loss": 0.4432, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.8078476630121177, | |
| "grad_norm": 0.3375686338902311, | |
| "learning_rate": 1.937452744057933e-05, | |
| "loss": 0.4453, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.8107328332371609, | |
| "grad_norm": 0.34947963167148055, | |
| "learning_rate": 1.936573187525451e-05, | |
| "loss": 0.4392, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.8136180034622043, | |
| "grad_norm": 0.329777575614968, | |
| "learning_rate": 1.9356876923050395e-05, | |
| "loss": 0.4554, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.8165031736872476, | |
| "grad_norm": 0.3481917738275054, | |
| "learning_rate": 1.9347962640115078e-05, | |
| "loss": 0.4172, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.8193883439122909, | |
| "grad_norm": 0.3488032875220557, | |
| "learning_rate": 1.933898908297286e-05, | |
| "loss": 0.4471, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.8222735141373341, | |
| "grad_norm": 0.35259907680989855, | |
| "learning_rate": 1.9329956308523886e-05, | |
| "loss": 0.4265, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.8251586843623774, | |
| "grad_norm": 0.36772785974509603, | |
| "learning_rate": 1.9320864374043805e-05, | |
| "loss": 0.4346, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.8280438545874207, | |
| "grad_norm": 0.33906704838185237, | |
| "learning_rate": 1.9311713337183375e-05, | |
| "loss": 0.4222, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.8309290248124639, | |
| "grad_norm": 0.3197548387537902, | |
| "learning_rate": 1.9302503255968127e-05, | |
| "loss": 0.4242, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.8338141950375072, | |
| "grad_norm": 0.3790505826213643, | |
| "learning_rate": 1.929323418879798e-05, | |
| "loss": 0.444, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.8366993652625505, | |
| "grad_norm": 0.356548976452809, | |
| "learning_rate": 1.928390619444687e-05, | |
| "loss": 0.4389, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.8395845354875938, | |
| "grad_norm": 0.3238309837351344, | |
| "learning_rate": 1.9274519332062393e-05, | |
| "loss": 0.4428, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.842469705712637, | |
| "grad_norm": 0.3328195979317483, | |
| "learning_rate": 1.9265073661165407e-05, | |
| "loss": 0.4418, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.8453548759376803, | |
| "grad_norm": 0.34611626618940344, | |
| "learning_rate": 1.9255569241649685e-05, | |
| "loss": 0.4311, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 0.8482400461627236, | |
| "grad_norm": 0.34500455296149757, | |
| "learning_rate": 1.9246006133781496e-05, | |
| "loss": 0.4307, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.8511252163877668, | |
| "grad_norm": 0.3473701533278901, | |
| "learning_rate": 1.9236384398199262e-05, | |
| "loss": 0.4484, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.8540103866128101, | |
| "grad_norm": 0.3409397737914092, | |
| "learning_rate": 1.9226704095913143e-05, | |
| "loss": 0.4247, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.8568955568378535, | |
| "grad_norm": 0.36709716826008937, | |
| "learning_rate": 1.9216965288304675e-05, | |
| "loss": 0.4454, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.8597807270628968, | |
| "grad_norm": 0.32825230009567846, | |
| "learning_rate": 1.9207168037126352e-05, | |
| "loss": 0.4246, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.86266589728794, | |
| "grad_norm": 0.33313005166969667, | |
| "learning_rate": 1.919731240450127e-05, | |
| "loss": 0.4204, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 0.8655510675129833, | |
| "grad_norm": 0.34672496584019735, | |
| "learning_rate": 1.91873984529227e-05, | |
| "loss": 0.4469, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.8684362377380266, | |
| "grad_norm": 0.3342264609614132, | |
| "learning_rate": 1.917742624525371e-05, | |
| "loss": 0.4231, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 0.8713214079630698, | |
| "grad_norm": 0.3612492143077761, | |
| "learning_rate": 1.9167395844726763e-05, | |
| "loss": 0.4454, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.8742065781881131, | |
| "grad_norm": 0.3407789646703214, | |
| "learning_rate": 1.915730731494332e-05, | |
| "loss": 0.4293, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.8770917484131564, | |
| "grad_norm": 0.33592086055664455, | |
| "learning_rate": 1.914716071987342e-05, | |
| "loss": 0.4358, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.8799769186381996, | |
| "grad_norm": 0.36425558440653516, | |
| "learning_rate": 1.91369561238553e-05, | |
| "loss": 0.4378, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.8828620888632429, | |
| "grad_norm": 0.3522122082887795, | |
| "learning_rate": 1.912669359159496e-05, | |
| "loss": 0.4295, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.8857472590882862, | |
| "grad_norm": 0.324930415843359, | |
| "learning_rate": 1.911637318816578e-05, | |
| "loss": 0.4251, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 0.8886324293133295, | |
| "grad_norm": 0.3466989166716833, | |
| "learning_rate": 1.9105994979008083e-05, | |
| "loss": 0.4328, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.8915175995383727, | |
| "grad_norm": 0.3603001072615171, | |
| "learning_rate": 1.9095559029928735e-05, | |
| "loss": 0.4331, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.894402769763416, | |
| "grad_norm": 0.35998678640671994, | |
| "learning_rate": 1.9085065407100723e-05, | |
| "loss": 0.4247, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.8972879399884593, | |
| "grad_norm": 0.3353610544242477, | |
| "learning_rate": 1.9074514177062734e-05, | |
| "loss": 0.4365, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 0.9001731102135025, | |
| "grad_norm": 0.3362819633369662, | |
| "learning_rate": 1.9063905406718735e-05, | |
| "loss": 0.4387, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.9030582804385459, | |
| "grad_norm": 0.35146022780606057, | |
| "learning_rate": 1.9053239163337553e-05, | |
| "loss": 0.4294, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 0.9059434506635892, | |
| "grad_norm": 0.3334044241812438, | |
| "learning_rate": 1.9042515514552438e-05, | |
| "loss": 0.4297, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.9088286208886325, | |
| "grad_norm": 0.33641137519229225, | |
| "learning_rate": 1.9031734528360642e-05, | |
| "loss": 0.42, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.9117137911136757, | |
| "grad_norm": 0.37493485136716287, | |
| "learning_rate": 1.9020896273122987e-05, | |
| "loss": 0.4254, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.914598961338719, | |
| "grad_norm": 0.3466700992777568, | |
| "learning_rate": 1.901000081756343e-05, | |
| "loss": 0.4203, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 0.9174841315637623, | |
| "grad_norm": 0.3601041654222663, | |
| "learning_rate": 1.899904823076863e-05, | |
| "loss": 0.4293, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.9203693017888055, | |
| "grad_norm": 0.5717625955327809, | |
| "learning_rate": 1.8988038582187506e-05, | |
| "loss": 0.4331, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 0.9232544720138488, | |
| "grad_norm": 0.3563731456994365, | |
| "learning_rate": 1.8976971941630792e-05, | |
| "loss": 0.4434, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.9261396422388921, | |
| "grad_norm": 0.3133483522354351, | |
| "learning_rate": 1.896584837927061e-05, | |
| "loss": 0.4381, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.9290248124639354, | |
| "grad_norm": 0.3602400529793251, | |
| "learning_rate": 1.8954667965640013e-05, | |
| "loss": 0.4295, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.9319099826889786, | |
| "grad_norm": 0.3660387727678179, | |
| "learning_rate": 1.894343077163254e-05, | |
| "loss": 0.4137, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 0.9347951529140219, | |
| "grad_norm": 0.33502056662496815, | |
| "learning_rate": 1.8932136868501763e-05, | |
| "loss": 0.431, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.9376803231390652, | |
| "grad_norm": 0.3519580845164796, | |
| "learning_rate": 1.8920786327860848e-05, | |
| "loss": 0.4382, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.9405654933641084, | |
| "grad_norm": 0.3609318959364458, | |
| "learning_rate": 1.8909379221682083e-05, | |
| "loss": 0.4263, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.9434506635891518, | |
| "grad_norm": 0.3470987740772583, | |
| "learning_rate": 1.8897915622296436e-05, | |
| "loss": 0.436, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.9463358338141951, | |
| "grad_norm": 0.3524435689745886, | |
| "learning_rate": 1.8886395602393087e-05, | |
| "loss": 0.4364, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.9492210040392384, | |
| "grad_norm": 0.3438549232804042, | |
| "learning_rate": 1.887481923501898e-05, | |
| "loss": 0.4457, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 0.9521061742642816, | |
| "grad_norm": 0.3359346965205838, | |
| "learning_rate": 1.8863186593578336e-05, | |
| "loss": 0.4489, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.9549913444893249, | |
| "grad_norm": 0.33509833712084897, | |
| "learning_rate": 1.8851497751832216e-05, | |
| "loss": 0.4291, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 0.9578765147143682, | |
| "grad_norm": 0.34585796301858324, | |
| "learning_rate": 1.8839752783898033e-05, | |
| "loss": 0.4449, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.9607616849394114, | |
| "grad_norm": 0.38171505764397407, | |
| "learning_rate": 1.8827951764249093e-05, | |
| "loss": 0.4482, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.9636468551644547, | |
| "grad_norm": 0.37008446077548557, | |
| "learning_rate": 1.8816094767714113e-05, | |
| "loss": 0.441, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.966532025389498, | |
| "grad_norm": 0.36975245636292803, | |
| "learning_rate": 1.880418186947676e-05, | |
| "loss": 0.4411, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.9694171956145412, | |
| "grad_norm": 0.36040770935696576, | |
| "learning_rate": 1.8792213145075156e-05, | |
| "loss": 0.4279, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.9723023658395845, | |
| "grad_norm": 0.3555551748715796, | |
| "learning_rate": 1.8780188670401415e-05, | |
| "loss": 0.4201, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 0.9751875360646278, | |
| "grad_norm": 0.34389008179144037, | |
| "learning_rate": 1.876810852170116e-05, | |
| "loss": 0.4659, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.9780727062896711, | |
| "grad_norm": 0.34681456919275633, | |
| "learning_rate": 1.875597277557303e-05, | |
| "loss": 0.4214, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.9809578765147143, | |
| "grad_norm": 0.33612247065522033, | |
| "learning_rate": 1.8743781508968206e-05, | |
| "loss": 0.4322, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.9838430467397576, | |
| "grad_norm": 0.35053266185577925, | |
| "learning_rate": 1.8731534799189905e-05, | |
| "loss": 0.4338, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 0.986728216964801, | |
| "grad_norm": 0.3554276718676979, | |
| "learning_rate": 1.8719232723892916e-05, | |
| "loss": 0.4387, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.9896133871898442, | |
| "grad_norm": 0.3439024647253949, | |
| "learning_rate": 1.8706875361083088e-05, | |
| "loss": 0.4161, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 0.9924985574148875, | |
| "grad_norm": 0.32640748056119073, | |
| "learning_rate": 1.869446278911683e-05, | |
| "loss": 0.4205, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.9953837276399308, | |
| "grad_norm": 0.3673852819098302, | |
| "learning_rate": 1.8681995086700654e-05, | |
| "loss": 0.4264, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.9982688978649741, | |
| "grad_norm": 0.3414077938012464, | |
| "learning_rate": 1.866947233289061e-05, | |
| "loss": 0.4295, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.4312867522239685, | |
| "eval_runtime": 0.6755, | |
| "eval_samples_per_second": 113.987, | |
| "eval_steps_per_second": 2.961, | |
| "step": 1733 | |
| }, | |
| { | |
| "epoch": 1.0011540680900173, | |
| "grad_norm": 0.3719885030349497, | |
| "learning_rate": 1.865689460709185e-05, | |
| "loss": 0.4085, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 1.0040392383150605, | |
| "grad_norm": 0.3653013563481502, | |
| "learning_rate": 1.8644261989058082e-05, | |
| "loss": 0.3663, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.006924408540104, | |
| "grad_norm": 0.4249244844285388, | |
| "learning_rate": 1.8631574558891087e-05, | |
| "loss": 0.3636, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 1.009809578765147, | |
| "grad_norm": 0.4065943346576567, | |
| "learning_rate": 1.8618832397040192e-05, | |
| "loss": 0.3602, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.0126947489901905, | |
| "grad_norm": 0.3839053642256101, | |
| "learning_rate": 1.860603558430178e-05, | |
| "loss": 0.3793, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 1.0155799192152337, | |
| "grad_norm": 0.36995484185295535, | |
| "learning_rate": 1.859318420181877e-05, | |
| "loss": 0.3438, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.018465089440277, | |
| "grad_norm": 0.3848314010331738, | |
| "learning_rate": 1.8580278331080085e-05, | |
| "loss": 0.3859, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 1.0213502596653203, | |
| "grad_norm": 0.3963349177249421, | |
| "learning_rate": 1.8567318053920166e-05, | |
| "loss": 0.3769, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.0242354298903635, | |
| "grad_norm": 0.37297944843138187, | |
| "learning_rate": 1.8554303452518436e-05, | |
| "loss": 0.3579, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 1.0271206001154067, | |
| "grad_norm": 0.3904119687309658, | |
| "learning_rate": 1.854123460939877e-05, | |
| "loss": 0.3715, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.0300057703404502, | |
| "grad_norm": 0.3788697016458385, | |
| "learning_rate": 1.8528111607429e-05, | |
| "loss": 0.3728, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 1.0328909405654934, | |
| "grad_norm": 0.37738365383900435, | |
| "learning_rate": 1.851493452982036e-05, | |
| "loss": 0.3611, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.0357761107905366, | |
| "grad_norm": 0.39677380812493784, | |
| "learning_rate": 1.8501703460126973e-05, | |
| "loss": 0.3661, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 1.03866128101558, | |
| "grad_norm": 0.39895561200913643, | |
| "learning_rate": 1.8488418482245317e-05, | |
| "loss": 0.3699, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.0415464512406232, | |
| "grad_norm": 0.3962653634072917, | |
| "learning_rate": 1.84750796804137e-05, | |
| "loss": 0.3513, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 1.0444316214656664, | |
| "grad_norm": 0.4045522991037393, | |
| "learning_rate": 1.8461687139211712e-05, | |
| "loss": 0.3621, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.0473167916907098, | |
| "grad_norm": 0.37889291444959144, | |
| "learning_rate": 1.844824094355971e-05, | |
| "loss": 0.3629, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 1.050201961915753, | |
| "grad_norm": 0.36255794337001546, | |
| "learning_rate": 1.843474117871826e-05, | |
| "loss": 0.356, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.0530871321407964, | |
| "grad_norm": 0.36607303065989255, | |
| "learning_rate": 1.84211879302876e-05, | |
| "loss": 0.3481, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 1.0559723023658396, | |
| "grad_norm": 0.38272410808497676, | |
| "learning_rate": 1.84075812842071e-05, | |
| "loss": 0.3509, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.0588574725908828, | |
| "grad_norm": 0.359466522831376, | |
| "learning_rate": 1.8393921326754724e-05, | |
| "loss": 0.3691, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 1.0617426428159262, | |
| "grad_norm": 0.3758374522906589, | |
| "learning_rate": 1.8380208144546473e-05, | |
| "loss": 0.3718, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.0646278130409694, | |
| "grad_norm": 0.3762140791163445, | |
| "learning_rate": 1.836644182453584e-05, | |
| "loss": 0.3777, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 1.0675129832660126, | |
| "grad_norm": 0.4061862324473009, | |
| "learning_rate": 1.835262245401326e-05, | |
| "loss": 0.3618, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.070398153491056, | |
| "grad_norm": 0.361301220342186, | |
| "learning_rate": 1.833875012060555e-05, | |
| "loss": 0.3974, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 1.0732833237160992, | |
| "grad_norm": 0.3769202277748975, | |
| "learning_rate": 1.8324824912275355e-05, | |
| "loss": 0.3495, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.0761684939411424, | |
| "grad_norm": 0.3605517647936932, | |
| "learning_rate": 1.8310846917320602e-05, | |
| "loss": 0.3481, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 1.0790536641661859, | |
| "grad_norm": 0.384557592260759, | |
| "learning_rate": 1.8296816224373926e-05, | |
| "loss": 0.3431, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.081938834391229, | |
| "grad_norm": 0.39969729027596174, | |
| "learning_rate": 1.828273292240211e-05, | |
| "loss": 0.361, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 1.0848240046162723, | |
| "grad_norm": 0.3893084314526289, | |
| "learning_rate": 1.8268597100705534e-05, | |
| "loss": 0.3533, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.0877091748413157, | |
| "grad_norm": 0.39190402847282024, | |
| "learning_rate": 1.8254408848917587e-05, | |
| "loss": 0.3528, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 1.0905943450663589, | |
| "grad_norm": 0.3991376675237472, | |
| "learning_rate": 1.824016825700412e-05, | |
| "loss": 0.3547, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.0934795152914023, | |
| "grad_norm": 0.3647535186718755, | |
| "learning_rate": 1.8225875415262852e-05, | |
| "loss": 0.3752, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 1.0963646855164455, | |
| "grad_norm": 0.36492225390564653, | |
| "learning_rate": 1.8211530414322834e-05, | |
| "loss": 0.3705, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.0992498557414887, | |
| "grad_norm": 0.38796798500765634, | |
| "learning_rate": 1.8197133345143828e-05, | |
| "loss": 0.3691, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 1.1021350259665321, | |
| "grad_norm": 0.3827359128376877, | |
| "learning_rate": 1.8182684299015764e-05, | |
| "loss": 0.3689, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.1050201961915753, | |
| "grad_norm": 0.38752794638536536, | |
| "learning_rate": 1.816818336755816e-05, | |
| "loss": 0.3567, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 1.1079053664166185, | |
| "grad_norm": 0.38630104069610516, | |
| "learning_rate": 1.815363064271952e-05, | |
| "loss": 0.3483, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.110790536641662, | |
| "grad_norm": 0.37210418058912814, | |
| "learning_rate": 1.813902621677676e-05, | |
| "loss": 0.3543, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 1.1136757068667051, | |
| "grad_norm": 0.37713275848642197, | |
| "learning_rate": 1.812437018233464e-05, | |
| "loss": 0.3531, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.1165608770917483, | |
| "grad_norm": 0.36232403860634815, | |
| "learning_rate": 1.8109662632325152e-05, | |
| "loss": 0.3658, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 1.1194460473167918, | |
| "grad_norm": 0.3466007749649004, | |
| "learning_rate": 1.809490366000694e-05, | |
| "loss": 0.3717, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.122331217541835, | |
| "grad_norm": 0.37512115226860376, | |
| "learning_rate": 1.8080093358964727e-05, | |
| "loss": 0.3684, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 1.1252163877668782, | |
| "grad_norm": 0.4257633969210461, | |
| "learning_rate": 1.806523182310868e-05, | |
| "loss": 0.365, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.1281015579919216, | |
| "grad_norm": 0.3907894746639846, | |
| "learning_rate": 1.8050319146673856e-05, | |
| "loss": 0.3539, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 1.1309867282169648, | |
| "grad_norm": 0.3868414285848145, | |
| "learning_rate": 1.8035355424219586e-05, | |
| "loss": 0.3619, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.133871898442008, | |
| "grad_norm": 0.391737905658893, | |
| "learning_rate": 1.8020340750628872e-05, | |
| "loss": 0.3556, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 1.1367570686670514, | |
| "grad_norm": 0.39546663065808646, | |
| "learning_rate": 1.800527522110779e-05, | |
| "loss": 0.3608, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.1396422388920946, | |
| "grad_norm": 0.3982732465681356, | |
| "learning_rate": 1.7990158931184892e-05, | |
| "loss": 0.3553, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 1.1425274091171378, | |
| "grad_norm": 0.3738040441095041, | |
| "learning_rate": 1.79749919767106e-05, | |
| "loss": 0.3599, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.1454125793421812, | |
| "grad_norm": 0.36487443451311957, | |
| "learning_rate": 1.7959774453856576e-05, | |
| "loss": 0.3521, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 1.1482977495672244, | |
| "grad_norm": 0.40333453889614557, | |
| "learning_rate": 1.794450645911514e-05, | |
| "loss": 0.3651, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 1.1511829197922678, | |
| "grad_norm": 0.5753350465511803, | |
| "learning_rate": 1.7929188089298653e-05, | |
| "loss": 0.369, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 1.154068090017311, | |
| "grad_norm": 0.3727451234003791, | |
| "learning_rate": 1.791381944153889e-05, | |
| "loss": 0.3536, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.1569532602423542, | |
| "grad_norm": 0.3787668346391338, | |
| "learning_rate": 1.789840061328643e-05, | |
| "loss": 0.3765, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 1.1598384304673977, | |
| "grad_norm": 0.36284152770529254, | |
| "learning_rate": 1.788293170231005e-05, | |
| "loss": 0.3455, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 1.1627236006924409, | |
| "grad_norm": 0.38508608974804587, | |
| "learning_rate": 1.7867412806696087e-05, | |
| "loss": 0.3716, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 1.165608770917484, | |
| "grad_norm": 0.3959297634423943, | |
| "learning_rate": 1.785184402484782e-05, | |
| "loss": 0.3715, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 1.1684939411425275, | |
| "grad_norm": 0.4224675659582224, | |
| "learning_rate": 1.7836225455484865e-05, | |
| "loss": 0.3844, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 1.1713791113675707, | |
| "grad_norm": 0.36797685694286536, | |
| "learning_rate": 1.7820557197642513e-05, | |
| "loss": 0.3691, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 1.1742642815926139, | |
| "grad_norm": 0.3831267176121955, | |
| "learning_rate": 1.7804839350671145e-05, | |
| "loss": 0.3571, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 1.1771494518176573, | |
| "grad_norm": 0.3905138892178264, | |
| "learning_rate": 1.778907201423557e-05, | |
| "loss": 0.3636, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 1.1800346220427005, | |
| "grad_norm": 0.3912936170187795, | |
| "learning_rate": 1.7773255288314398e-05, | |
| "loss": 0.3574, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 1.182919792267744, | |
| "grad_norm": 0.3754522878062524, | |
| "learning_rate": 1.775738927319942e-05, | |
| "loss": 0.3874, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.185804962492787, | |
| "grad_norm": 0.36965375437322023, | |
| "learning_rate": 1.774147406949496e-05, | |
| "loss": 0.3717, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 1.1886901327178303, | |
| "grad_norm": 0.4220256804649789, | |
| "learning_rate": 1.7725509778117242e-05, | |
| "loss": 0.3541, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 1.1915753029428737, | |
| "grad_norm": 0.38295286541947393, | |
| "learning_rate": 1.770949650029374e-05, | |
| "loss": 0.3574, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 1.194460473167917, | |
| "grad_norm": 0.37402840183729646, | |
| "learning_rate": 1.7693434337562557e-05, | |
| "loss": 0.3738, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 1.1973456433929601, | |
| "grad_norm": 0.385808424420751, | |
| "learning_rate": 1.7677323391771766e-05, | |
| "loss": 0.3646, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 1.2002308136180035, | |
| "grad_norm": 0.37207777332282577, | |
| "learning_rate": 1.766116376507876e-05, | |
| "loss": 0.3699, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 1.2031159838430467, | |
| "grad_norm": 0.39725920392186914, | |
| "learning_rate": 1.764495555994962e-05, | |
| "loss": 0.3736, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 1.20600115406809, | |
| "grad_norm": 0.3885483852856526, | |
| "learning_rate": 1.7628698879158444e-05, | |
| "loss": 0.3493, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 1.2088863242931334, | |
| "grad_norm": 0.3779946551242946, | |
| "learning_rate": 1.7612393825786723e-05, | |
| "loss": 0.3758, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 1.2117714945181766, | |
| "grad_norm": 0.3625482575463983, | |
| "learning_rate": 1.7596040503222665e-05, | |
| "loss": 0.3745, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 1.2146566647432198, | |
| "grad_norm": 0.38330347138439946, | |
| "learning_rate": 1.7579639015160545e-05, | |
| "loss": 0.362, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 1.2175418349682632, | |
| "grad_norm": 0.4177686333809965, | |
| "learning_rate": 1.7563189465600047e-05, | |
| "loss": 0.3699, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 1.2204270051933064, | |
| "grad_norm": 0.3707075705845332, | |
| "learning_rate": 1.7546691958845617e-05, | |
| "loss": 0.3485, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 1.2233121754183496, | |
| "grad_norm": 0.37675133161938845, | |
| "learning_rate": 1.7530146599505782e-05, | |
| "loss": 0.3738, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 1.226197345643393, | |
| "grad_norm": 0.4073716130331381, | |
| "learning_rate": 1.75135534924925e-05, | |
| "loss": 0.3726, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 1.2290825158684362, | |
| "grad_norm": 0.377632182741206, | |
| "learning_rate": 1.7496912743020487e-05, | |
| "loss": 0.3606, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 1.2319676860934794, | |
| "grad_norm": 0.41027591746654574, | |
| "learning_rate": 1.748022445660656e-05, | |
| "loss": 0.3677, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 1.2348528563185228, | |
| "grad_norm": 0.454127691618852, | |
| "learning_rate": 1.7463488739068952e-05, | |
| "loss": 0.356, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 1.237738026543566, | |
| "grad_norm": 0.38902494056071213, | |
| "learning_rate": 1.744670569652666e-05, | |
| "loss": 0.356, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 1.2406231967686094, | |
| "grad_norm": 0.3547365448603461, | |
| "learning_rate": 1.742987543539876e-05, | |
| "loss": 0.3717, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 1.2435083669936526, | |
| "grad_norm": 0.37046329915555815, | |
| "learning_rate": 1.741299806240373e-05, | |
| "loss": 0.3675, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 1.2463935372186958, | |
| "grad_norm": 0.3869703816603672, | |
| "learning_rate": 1.7396073684558788e-05, | |
| "loss": 0.3665, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 1.2492787074437393, | |
| "grad_norm": 0.38188698969658114, | |
| "learning_rate": 1.737910240917919e-05, | |
| "loss": 0.3624, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 1.2521638776687825, | |
| "grad_norm": 0.37535923832323237, | |
| "learning_rate": 1.7362084343877576e-05, | |
| "loss": 0.368, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 1.2550490478938257, | |
| "grad_norm": 0.36714065996796413, | |
| "learning_rate": 1.7345019596563268e-05, | |
| "loss": 0.3667, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 1.257934218118869, | |
| "grad_norm": 0.4157188737740098, | |
| "learning_rate": 1.732790827544159e-05, | |
| "loss": 0.3737, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 1.2608193883439123, | |
| "grad_norm": 0.3811681622112457, | |
| "learning_rate": 1.7310750489013198e-05, | |
| "loss": 0.3544, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 1.2637045585689557, | |
| "grad_norm": 0.3753050362248373, | |
| "learning_rate": 1.729354634607336e-05, | |
| "loss": 0.3664, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 1.266589728793999, | |
| "grad_norm": 0.36855745458054423, | |
| "learning_rate": 1.7276295955711295e-05, | |
| "loss": 0.3637, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 1.269474899019042, | |
| "grad_norm": 0.3650539628680322, | |
| "learning_rate": 1.725899942730947e-05, | |
| "loss": 0.3645, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 1.2723600692440855, | |
| "grad_norm": 0.3831096440319005, | |
| "learning_rate": 1.724165687054291e-05, | |
| "loss": 0.3618, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 1.2752452394691287, | |
| "grad_norm": 0.3603410399079337, | |
| "learning_rate": 1.7224268395378493e-05, | |
| "loss": 0.3453, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 1.278130409694172, | |
| "grad_norm": 0.37269959302348715, | |
| "learning_rate": 1.720683411207426e-05, | |
| "loss": 0.3576, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 1.2810155799192153, | |
| "grad_norm": 0.38173221937480023, | |
| "learning_rate": 1.718935413117872e-05, | |
| "loss": 0.371, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 1.2839007501442585, | |
| "grad_norm": 0.3722436971668028, | |
| "learning_rate": 1.7171828563530146e-05, | |
| "loss": 0.3663, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 1.2867859203693017, | |
| "grad_norm": 0.39837683099640625, | |
| "learning_rate": 1.7154257520255864e-05, | |
| "loss": 0.3745, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 1.2896710905943451, | |
| "grad_norm": 0.41009144762310795, | |
| "learning_rate": 1.713664111277156e-05, | |
| "loss": 0.3666, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 1.2925562608193883, | |
| "grad_norm": 0.36668782585448506, | |
| "learning_rate": 1.711897945278057e-05, | |
| "loss": 0.3738, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 1.2954414310444315, | |
| "grad_norm": 0.41098358203090174, | |
| "learning_rate": 1.710127265227316e-05, | |
| "loss": 0.3668, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 1.298326601269475, | |
| "grad_norm": 0.3862089819354528, | |
| "learning_rate": 1.7083520823525836e-05, | |
| "loss": 0.3661, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 1.3012117714945182, | |
| "grad_norm": 0.4033977142486018, | |
| "learning_rate": 1.706572407910062e-05, | |
| "loss": 0.3688, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 1.3040969417195614, | |
| "grad_norm": 0.4003446173157873, | |
| "learning_rate": 1.7047882531844338e-05, | |
| "loss": 0.377, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 1.3069821119446048, | |
| "grad_norm": 0.3702354075920004, | |
| "learning_rate": 1.7029996294887904e-05, | |
| "loss": 0.3774, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 1.309867282169648, | |
| "grad_norm": 0.3915858903728003, | |
| "learning_rate": 1.7012065481645604e-05, | |
| "loss": 0.3671, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 1.3127524523946912, | |
| "grad_norm": 0.3721772452216554, | |
| "learning_rate": 1.699409020581438e-05, | |
| "loss": 0.3553, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 1.3156376226197346, | |
| "grad_norm": 0.3726078669899795, | |
| "learning_rate": 1.6976070581373098e-05, | |
| "loss": 0.3781, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 1.3185227928447778, | |
| "grad_norm": 0.3851091572746381, | |
| "learning_rate": 1.6958006722581838e-05, | |
| "loss": 0.3764, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 1.321407963069821, | |
| "grad_norm": 0.3999997138224476, | |
| "learning_rate": 1.693989874398116e-05, | |
| "loss": 0.3719, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 1.3242931332948644, | |
| "grad_norm": 0.36767006912934236, | |
| "learning_rate": 1.6921746760391387e-05, | |
| "loss": 0.3748, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 1.3271783035199076, | |
| "grad_norm": 0.3730179310503956, | |
| "learning_rate": 1.6903550886911873e-05, | |
| "loss": 0.3608, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 1.3300634737449508, | |
| "grad_norm": 0.3672274323141715, | |
| "learning_rate": 1.688531123892027e-05, | |
| "loss": 0.3443, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 1.3329486439699942, | |
| "grad_norm": 0.36049125372252444, | |
| "learning_rate": 1.6867027932071786e-05, | |
| "loss": 0.3606, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 1.3358338141950374, | |
| "grad_norm": 0.37209341340715907, | |
| "learning_rate": 1.684870108229849e-05, | |
| "loss": 0.3866, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 1.3387189844200809, | |
| "grad_norm": 0.4041341082410308, | |
| "learning_rate": 1.6830330805808524e-05, | |
| "loss": 0.3552, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 1.341604154645124, | |
| "grad_norm": 0.44070233485279253, | |
| "learning_rate": 1.681191721908541e-05, | |
| "loss": 0.3641, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 1.3444893248701673, | |
| "grad_norm": 0.3782903985125648, | |
| "learning_rate": 1.6793460438887294e-05, | |
| "loss": 0.3702, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 1.3473744950952107, | |
| "grad_norm": 0.38093409058463656, | |
| "learning_rate": 1.67749605822462e-05, | |
| "loss": 0.3738, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 1.3502596653202539, | |
| "grad_norm": 0.4076843344147931, | |
| "learning_rate": 1.6756417766467294e-05, | |
| "loss": 0.3713, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 1.3531448355452973, | |
| "grad_norm": 0.4114981370603978, | |
| "learning_rate": 1.6737832109128142e-05, | |
| "loss": 0.3626, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 1.3560300057703405, | |
| "grad_norm": 0.4070630768126986, | |
| "learning_rate": 1.6719203728077963e-05, | |
| "loss": 0.3579, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 1.3589151759953837, | |
| "grad_norm": 0.3980167515551806, | |
| "learning_rate": 1.670053274143689e-05, | |
| "loss": 0.3688, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 1.3618003462204271, | |
| "grad_norm": 0.4117026797519244, | |
| "learning_rate": 1.6681819267595193e-05, | |
| "loss": 0.3631, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 1.3646855164454703, | |
| "grad_norm": 0.3794143818195765, | |
| "learning_rate": 1.6663063425212564e-05, | |
| "loss": 0.3658, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 1.3675706866705135, | |
| "grad_norm": 0.4389143548454321, | |
| "learning_rate": 1.6644265333217347e-05, | |
| "loss": 0.3622, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 1.370455856895557, | |
| "grad_norm": 0.34975549009666657, | |
| "learning_rate": 1.6625425110805774e-05, | |
| "loss": 0.3636, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 1.3733410271206001, | |
| "grad_norm": 0.4000657679860908, | |
| "learning_rate": 1.660654287744123e-05, | |
| "loss": 0.3655, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 1.3762261973456433, | |
| "grad_norm": 0.39327391261319994, | |
| "learning_rate": 1.6587618752853484e-05, | |
| "loss": 0.3673, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 1.3791113675706868, | |
| "grad_norm": 0.3803854498603796, | |
| "learning_rate": 1.656865285703794e-05, | |
| "loss": 0.3644, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 1.38199653779573, | |
| "grad_norm": 0.36271341863389506, | |
| "learning_rate": 1.6549645310254848e-05, | |
| "loss": 0.3597, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 1.3848817080207732, | |
| "grad_norm": 0.39017089645929526, | |
| "learning_rate": 1.653059623302858e-05, | |
| "loss": 0.3584, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 1.3877668782458166, | |
| "grad_norm": 0.377459677055801, | |
| "learning_rate": 1.6511505746146837e-05, | |
| "loss": 0.3603, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 1.3906520484708598, | |
| "grad_norm": 0.36253902214360645, | |
| "learning_rate": 1.6492373970659897e-05, | |
| "loss": 0.3622, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 1.393537218695903, | |
| "grad_norm": 0.3792755236074643, | |
| "learning_rate": 1.647320102787984e-05, | |
| "loss": 0.3579, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 1.3964223889209464, | |
| "grad_norm": 0.4106370547161479, | |
| "learning_rate": 1.645398703937979e-05, | |
| "loss": 0.3548, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 1.3993075591459896, | |
| "grad_norm": 0.3961446879313167, | |
| "learning_rate": 1.6434732126993137e-05, | |
| "loss": 0.3788, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 1.4021927293710328, | |
| "grad_norm": 0.41444403209474506, | |
| "learning_rate": 1.6415436412812742e-05, | |
| "loss": 0.3935, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 1.4050778995960762, | |
| "grad_norm": 0.45443169110667836, | |
| "learning_rate": 1.6396100019190216e-05, | |
| "loss": 0.3672, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 1.4079630698211194, | |
| "grad_norm": 0.3704577288597062, | |
| "learning_rate": 1.6376723068735096e-05, | |
| "loss": 0.3614, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 1.4108482400461626, | |
| "grad_norm": 0.37885398698326, | |
| "learning_rate": 1.6357305684314083e-05, | |
| "loss": 0.3473, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 1.413733410271206, | |
| "grad_norm": 0.41962586363127974, | |
| "learning_rate": 1.6337847989050276e-05, | |
| "loss": 0.3658, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 1.4166185804962492, | |
| "grad_norm": 0.3851208552370083, | |
| "learning_rate": 1.6318350106322363e-05, | |
| "loss": 0.3699, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 1.4195037507212924, | |
| "grad_norm": 0.3548157462238281, | |
| "learning_rate": 1.6298812159763868e-05, | |
| "loss": 0.3681, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 1.4223889209463358, | |
| "grad_norm": 0.37651463846589284, | |
| "learning_rate": 1.6279234273262357e-05, | |
| "loss": 0.3516, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 1.425274091171379, | |
| "grad_norm": 0.35900318539287046, | |
| "learning_rate": 1.6259616570958638e-05, | |
| "loss": 0.3704, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 1.4281592613964225, | |
| "grad_norm": 0.4024381546192383, | |
| "learning_rate": 1.6239959177246e-05, | |
| "loss": 0.3638, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 1.4310444316214657, | |
| "grad_norm": 0.398944280818137, | |
| "learning_rate": 1.6220262216769393e-05, | |
| "loss": 0.3813, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 1.4339296018465089, | |
| "grad_norm": 0.3831395961609494, | |
| "learning_rate": 1.6200525814424666e-05, | |
| "loss": 0.3693, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 1.4368147720715523, | |
| "grad_norm": 0.3661492085432, | |
| "learning_rate": 1.6180750095357766e-05, | |
| "loss": 0.3776, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 1.4396999422965955, | |
| "grad_norm": 0.36586438881277866, | |
| "learning_rate": 1.6160935184963937e-05, | |
| "loss": 0.3503, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 1.442585112521639, | |
| "grad_norm": 0.3784608295775924, | |
| "learning_rate": 1.614108120888693e-05, | |
| "loss": 0.3715, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.445470282746682, | |
| "grad_norm": 0.364821310782025, | |
| "learning_rate": 1.6121188293018198e-05, | |
| "loss": 0.3774, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 1.4483554529717253, | |
| "grad_norm": 0.38118287581212157, | |
| "learning_rate": 1.6101256563496133e-05, | |
| "loss": 0.3608, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 1.4512406231967687, | |
| "grad_norm": 0.37900358798651623, | |
| "learning_rate": 1.608128614670521e-05, | |
| "loss": 0.3889, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 1.454125793421812, | |
| "grad_norm": 0.40251780895398115, | |
| "learning_rate": 1.6061277169275227e-05, | |
| "loss": 0.3606, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 1.4570109636468551, | |
| "grad_norm": 0.38327828759400334, | |
| "learning_rate": 1.6041229758080496e-05, | |
| "loss": 0.3635, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 1.4598961338718985, | |
| "grad_norm": 0.37246590219746933, | |
| "learning_rate": 1.6021144040239017e-05, | |
| "loss": 0.3766, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 1.4627813040969417, | |
| "grad_norm": 0.3833944550693614, | |
| "learning_rate": 1.60010201431117e-05, | |
| "loss": 0.3612, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 1.465666474321985, | |
| "grad_norm": 0.36363911288467893, | |
| "learning_rate": 1.598085819430155e-05, | |
| "loss": 0.3711, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 1.4685516445470284, | |
| "grad_norm": 0.3588573344819394, | |
| "learning_rate": 1.596065832165283e-05, | |
| "loss": 0.3726, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 1.4714368147720716, | |
| "grad_norm": 0.3705849995740623, | |
| "learning_rate": 1.5940420653250293e-05, | |
| "loss": 0.3642, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 1.4743219849971148, | |
| "grad_norm": 0.38341287260652185, | |
| "learning_rate": 1.5920145317418346e-05, | |
| "loss": 0.3714, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 1.4772071552221582, | |
| "grad_norm": 0.38734703763461453, | |
| "learning_rate": 1.589983244272024e-05, | |
| "loss": 0.3836, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 1.4800923254472014, | |
| "grad_norm": 0.36011456145391224, | |
| "learning_rate": 1.5879482157957245e-05, | |
| "loss": 0.3669, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 1.4829774956722446, | |
| "grad_norm": 0.3784776076526531, | |
| "learning_rate": 1.5859094592167858e-05, | |
| "loss": 0.3634, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 1.485862665897288, | |
| "grad_norm": 0.38334066789201254, | |
| "learning_rate": 1.5838669874626963e-05, | |
| "loss": 0.3836, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 1.4887478361223312, | |
| "grad_norm": 0.3769831368967705, | |
| "learning_rate": 1.5818208134845022e-05, | |
| "loss": 0.3592, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 1.4916330063473744, | |
| "grad_norm": 0.411783030147672, | |
| "learning_rate": 1.579770950256724e-05, | |
| "loss": 0.3794, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 1.4945181765724178, | |
| "grad_norm": 0.5027841057723603, | |
| "learning_rate": 1.577717410777276e-05, | |
| "loss": 0.3601, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 1.497403346797461, | |
| "grad_norm": 0.38774582541265035, | |
| "learning_rate": 1.5756602080673843e-05, | |
| "loss": 0.3753, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 1.5002885170225042, | |
| "grad_norm": 0.37739115096678616, | |
| "learning_rate": 1.5735993551715006e-05, | |
| "loss": 0.3737, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 1.5031736872475476, | |
| "grad_norm": 0.3902916999812962, | |
| "learning_rate": 1.571534865157224e-05, | |
| "loss": 0.3632, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 1.5060588574725908, | |
| "grad_norm": 0.3868041357607236, | |
| "learning_rate": 1.5694667511152143e-05, | |
| "loss": 0.3809, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 1.508944027697634, | |
| "grad_norm": 0.37704578010786705, | |
| "learning_rate": 1.5673950261591133e-05, | |
| "loss": 0.3694, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 1.5118291979226774, | |
| "grad_norm": 0.3607652817962323, | |
| "learning_rate": 1.5653197034254564e-05, | |
| "loss": 0.3519, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 1.5147143681477209, | |
| "grad_norm": 0.37320845213700304, | |
| "learning_rate": 1.563240796073594e-05, | |
| "loss": 0.3606, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 1.5175995383727638, | |
| "grad_norm": 0.43098838699804737, | |
| "learning_rate": 1.5611583172856045e-05, | |
| "loss": 0.3741, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 1.5204847085978073, | |
| "grad_norm": 0.3812237058754953, | |
| "learning_rate": 1.5590722802662143e-05, | |
| "loss": 0.3486, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 1.5233698788228507, | |
| "grad_norm": 0.3799133932625459, | |
| "learning_rate": 1.55698269824271e-05, | |
| "loss": 0.3657, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 1.5262550490478937, | |
| "grad_norm": 0.34887334582855944, | |
| "learning_rate": 1.5548895844648583e-05, | |
| "loss": 0.3656, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 1.529140219272937, | |
| "grad_norm": 0.3798697034141352, | |
| "learning_rate": 1.552792952204819e-05, | |
| "loss": 0.3714, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 1.5320253894979805, | |
| "grad_norm": 0.3597061950949025, | |
| "learning_rate": 1.550692814757063e-05, | |
| "loss": 0.3651, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 1.5349105597230237, | |
| "grad_norm": 0.39911942641093723, | |
| "learning_rate": 1.5485891854382868e-05, | |
| "loss": 0.3662, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 1.537795729948067, | |
| "grad_norm": 0.4048874005019976, | |
| "learning_rate": 1.5464820775873285e-05, | |
| "loss": 0.378, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 1.5406809001731103, | |
| "grad_norm": 0.3614850352345454, | |
| "learning_rate": 1.544371504565082e-05, | |
| "loss": 0.3511, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 1.5435660703981535, | |
| "grad_norm": 0.3852211514058087, | |
| "learning_rate": 1.5422574797544166e-05, | |
| "loss": 0.3571, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 1.5464512406231967, | |
| "grad_norm": 0.36466553350789727, | |
| "learning_rate": 1.5401400165600848e-05, | |
| "loss": 0.3654, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 1.5493364108482401, | |
| "grad_norm": 0.4640557855511497, | |
| "learning_rate": 1.5380191284086442e-05, | |
| "loss": 0.3663, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 1.5522215810732833, | |
| "grad_norm": 0.40382378120892326, | |
| "learning_rate": 1.5358948287483688e-05, | |
| "loss": 0.395, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 1.5551067512983265, | |
| "grad_norm": 0.34945555100262243, | |
| "learning_rate": 1.533767131049164e-05, | |
| "loss": 0.3507, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 1.55799192152337, | |
| "grad_norm": 0.4017093497659257, | |
| "learning_rate": 1.5316360488024838e-05, | |
| "loss": 0.3666, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 1.5608770917484132, | |
| "grad_norm": 0.36824354949459775, | |
| "learning_rate": 1.5295015955212397e-05, | |
| "loss": 0.3629, | |
| "step": 2705 | |
| }, | |
| { | |
| "epoch": 1.5637622619734564, | |
| "grad_norm": 0.37118819752646043, | |
| "learning_rate": 1.5273637847397216e-05, | |
| "loss": 0.3619, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 1.5666474321984998, | |
| "grad_norm": 0.380235084250298, | |
| "learning_rate": 1.5252226300135075e-05, | |
| "loss": 0.3627, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 1.569532602423543, | |
| "grad_norm": 0.35479734123182555, | |
| "learning_rate": 1.5230781449193787e-05, | |
| "loss": 0.3499, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 1.5724177726485862, | |
| "grad_norm": 0.37537141405956453, | |
| "learning_rate": 1.5209303430552353e-05, | |
| "loss": 0.3689, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 1.5753029428736296, | |
| "grad_norm": 0.38314418131207195, | |
| "learning_rate": 1.5187792380400066e-05, | |
| "loss": 0.3568, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 1.5781881130986728, | |
| "grad_norm": 0.4015489085451092, | |
| "learning_rate": 1.5166248435135687e-05, | |
| "loss": 0.3584, | |
| "step": 2735 | |
| }, | |
| { | |
| "epoch": 1.581073283323716, | |
| "grad_norm": 0.413370518767082, | |
| "learning_rate": 1.5144671731366537e-05, | |
| "loss": 0.3627, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 1.5839584535487594, | |
| "grad_norm": 0.3855856313069642, | |
| "learning_rate": 1.5123062405907689e-05, | |
| "loss": 0.3576, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 1.5868436237738026, | |
| "grad_norm": 0.40248976719527124, | |
| "learning_rate": 1.510142059578103e-05, | |
| "loss": 0.371, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 1.5897287939988458, | |
| "grad_norm": 0.3896121946823667, | |
| "learning_rate": 1.5079746438214452e-05, | |
| "loss": 0.3569, | |
| "step": 2755 | |
| }, | |
| { | |
| "epoch": 1.5926139642238892, | |
| "grad_norm": 0.3601348358274979, | |
| "learning_rate": 1.5058040070640947e-05, | |
| "loss": 0.3689, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 1.5954991344489324, | |
| "grad_norm": 0.35986769610870906, | |
| "learning_rate": 1.5036301630697747e-05, | |
| "loss": 0.3555, | |
| "step": 2765 | |
| }, | |
| { | |
| "epoch": 1.5983843046739756, | |
| "grad_norm": 0.36475176711784657, | |
| "learning_rate": 1.5014531256225459e-05, | |
| "loss": 0.3605, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 1.601269474899019, | |
| "grad_norm": 0.37930345881140104, | |
| "learning_rate": 1.499272908526717e-05, | |
| "loss": 0.3728, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 1.6041546451240625, | |
| "grad_norm": 0.38456807910427965, | |
| "learning_rate": 1.4970895256067593e-05, | |
| "loss": 0.3736, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 1.6070398153491054, | |
| "grad_norm": 0.4121207010530726, | |
| "learning_rate": 1.4949029907072179e-05, | |
| "loss": 0.3855, | |
| "step": 2785 | |
| }, | |
| { | |
| "epoch": 1.6099249855741489, | |
| "grad_norm": 0.36551396208539605, | |
| "learning_rate": 1.4927133176926245e-05, | |
| "loss": 0.3662, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 1.6128101557991923, | |
| "grad_norm": 0.3716986354984255, | |
| "learning_rate": 1.4905205204474088e-05, | |
| "loss": 0.3782, | |
| "step": 2795 | |
| }, | |
| { | |
| "epoch": 1.6156953260242353, | |
| "grad_norm": 0.6240855472300652, | |
| "learning_rate": 1.488324612875811e-05, | |
| "loss": 0.3464, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 1.6185804962492787, | |
| "grad_norm": 0.33838650586561203, | |
| "learning_rate": 1.4861256089017934e-05, | |
| "loss": 0.3571, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 1.621465666474322, | |
| "grad_norm": 0.3772401322980198, | |
| "learning_rate": 1.483923522468952e-05, | |
| "loss": 0.3582, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 1.6243508366993653, | |
| "grad_norm": 0.3686319357750962, | |
| "learning_rate": 1.4817183675404292e-05, | |
| "loss": 0.3705, | |
| "step": 2815 | |
| }, | |
| { | |
| "epoch": 1.6272360069244085, | |
| "grad_norm": 0.3968174151919166, | |
| "learning_rate": 1.4795101580988228e-05, | |
| "loss": 0.3806, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 1.630121177149452, | |
| "grad_norm": 0.3902887780136523, | |
| "learning_rate": 1.4772989081460997e-05, | |
| "loss": 0.3626, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 1.6330063473744951, | |
| "grad_norm": 0.3746341856550175, | |
| "learning_rate": 1.4750846317035071e-05, | |
| "loss": 0.3672, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 1.6358915175995383, | |
| "grad_norm": 0.3994993753443447, | |
| "learning_rate": 1.4728673428114808e-05, | |
| "loss": 0.3688, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 1.6387766878245817, | |
| "grad_norm": 0.37287704407061595, | |
| "learning_rate": 1.4706470555295605e-05, | |
| "loss": 0.3554, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 1.641661858049625, | |
| "grad_norm": 0.37701343446505425, | |
| "learning_rate": 1.4684237839362963e-05, | |
| "loss": 0.3654, | |
| "step": 2845 | |
| }, | |
| { | |
| "epoch": 1.6445470282746681, | |
| "grad_norm": 0.38168823604260177, | |
| "learning_rate": 1.4661975421291623e-05, | |
| "loss": 0.3586, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 1.6474321984997116, | |
| "grad_norm": 0.39975859055427077, | |
| "learning_rate": 1.4639683442244663e-05, | |
| "loss": 0.3716, | |
| "step": 2855 | |
| }, | |
| { | |
| "epoch": 1.6503173687247548, | |
| "grad_norm": 0.37090442126154205, | |
| "learning_rate": 1.4617362043572607e-05, | |
| "loss": 0.3588, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 1.653202538949798, | |
| "grad_norm": 0.3519117177291488, | |
| "learning_rate": 1.4595011366812512e-05, | |
| "loss": 0.3663, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 1.6560877091748414, | |
| "grad_norm": 0.3737325432685507, | |
| "learning_rate": 1.4572631553687103e-05, | |
| "loss": 0.3582, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 1.6589728793998846, | |
| "grad_norm": 0.38870437566944893, | |
| "learning_rate": 1.4550222746103835e-05, | |
| "loss": 0.351, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 1.6618580496249278, | |
| "grad_norm": 0.3617260019248044, | |
| "learning_rate": 1.4527785086154027e-05, | |
| "loss": 0.3696, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 1.6647432198499712, | |
| "grad_norm": 0.38332448310467415, | |
| "learning_rate": 1.4505318716111943e-05, | |
| "loss": 0.373, | |
| "step": 2885 | |
| }, | |
| { | |
| "epoch": 1.6676283900750144, | |
| "grad_norm": 0.3747344452694502, | |
| "learning_rate": 1.448282377843389e-05, | |
| "loss": 0.3664, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 1.6705135603000576, | |
| "grad_norm": 0.36202726723998613, | |
| "learning_rate": 1.4460300415757323e-05, | |
| "loss": 0.3766, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 1.673398730525101, | |
| "grad_norm": 0.37923473785013656, | |
| "learning_rate": 1.4437748770899936e-05, | |
| "loss": 0.3814, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 1.6762839007501442, | |
| "grad_norm": 0.3873960582212302, | |
| "learning_rate": 1.4415168986858754e-05, | |
| "loss": 0.3745, | |
| "step": 2905 | |
| }, | |
| { | |
| "epoch": 1.6791690709751874, | |
| "grad_norm": 0.4013398358287003, | |
| "learning_rate": 1.4392561206809232e-05, | |
| "loss": 0.3798, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 1.6820542412002308, | |
| "grad_norm": 0.40802241994691185, | |
| "learning_rate": 1.4369925574104344e-05, | |
| "loss": 0.3586, | |
| "step": 2915 | |
| }, | |
| { | |
| "epoch": 1.684939411425274, | |
| "grad_norm": 0.4010573684279247, | |
| "learning_rate": 1.4347262232273671e-05, | |
| "loss": 0.3578, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 1.6878245816503172, | |
| "grad_norm": 0.3609006547431779, | |
| "learning_rate": 1.4324571325022496e-05, | |
| "loss": 0.3592, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 1.6907097518753607, | |
| "grad_norm": 0.3682631162288877, | |
| "learning_rate": 1.4301852996230889e-05, | |
| "loss": 0.3611, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 1.693594922100404, | |
| "grad_norm": 0.39391349541100773, | |
| "learning_rate": 1.42791073899528e-05, | |
| "loss": 0.3549, | |
| "step": 2935 | |
| }, | |
| { | |
| "epoch": 1.696480092325447, | |
| "grad_norm": 0.44233833212347734, | |
| "learning_rate": 1.4256334650415135e-05, | |
| "loss": 0.3537, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 1.6993652625504905, | |
| "grad_norm": 0.3869468573057741, | |
| "learning_rate": 1.4233534922016859e-05, | |
| "loss": 0.37, | |
| "step": 2945 | |
| }, | |
| { | |
| "epoch": 1.702250432775534, | |
| "grad_norm": 0.37586357463502135, | |
| "learning_rate": 1.4210708349328062e-05, | |
| "loss": 0.3675, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 1.7051356030005769, | |
| "grad_norm": 0.3903279762819204, | |
| "learning_rate": 1.418785507708905e-05, | |
| "loss": 0.369, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 1.7080207732256203, | |
| "grad_norm": 0.3969910656440696, | |
| "learning_rate": 1.4164975250209432e-05, | |
| "loss": 0.3694, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 1.7109059434506637, | |
| "grad_norm": 0.3556902520239883, | |
| "learning_rate": 1.4142069013767192e-05, | |
| "loss": 0.3504, | |
| "step": 2965 | |
| }, | |
| { | |
| "epoch": 1.713791113675707, | |
| "grad_norm": 0.3583546127221291, | |
| "learning_rate": 1.4119136513007776e-05, | |
| "loss": 0.3681, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 1.71667628390075, | |
| "grad_norm": 0.3968433520239081, | |
| "learning_rate": 1.409617789334317e-05, | |
| "loss": 0.3752, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 1.7195614541257935, | |
| "grad_norm": 0.3898401819324146, | |
| "learning_rate": 1.4073193300350973e-05, | |
| "loss": 0.3698, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 1.7224466243508367, | |
| "grad_norm": 0.3698983985858203, | |
| "learning_rate": 1.405018287977348e-05, | |
| "loss": 0.3521, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 1.72533179457588, | |
| "grad_norm": 0.4816180292439013, | |
| "learning_rate": 1.402714677751675e-05, | |
| "loss": 0.3604, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 1.7282169648009233, | |
| "grad_norm": 0.4104869915376292, | |
| "learning_rate": 1.40040851396497e-05, | |
| "loss": 0.3799, | |
| "step": 2995 | |
| }, | |
| { | |
| "epoch": 1.7311021350259665, | |
| "grad_norm": 0.3917852021920299, | |
| "learning_rate": 1.3980998112403146e-05, | |
| "loss": 0.3584, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.7339873052510097, | |
| "grad_norm": 0.3910325407989454, | |
| "learning_rate": 1.3957885842168908e-05, | |
| "loss": 0.3579, | |
| "step": 3005 | |
| }, | |
| { | |
| "epoch": 1.7368724754760532, | |
| "grad_norm": 0.35526389545543485, | |
| "learning_rate": 1.3934748475498867e-05, | |
| "loss": 0.3588, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 1.7397576457010964, | |
| "grad_norm": 0.39546625910444144, | |
| "learning_rate": 1.3911586159104032e-05, | |
| "loss": 0.3476, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 1.7426428159261396, | |
| "grad_norm": 0.369055887907894, | |
| "learning_rate": 1.3888399039853618e-05, | |
| "loss": 0.352, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 1.745527986151183, | |
| "grad_norm": 0.35801953491056676, | |
| "learning_rate": 1.3865187264774114e-05, | |
| "loss": 0.3554, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 1.7484131563762262, | |
| "grad_norm": 0.3531901383810597, | |
| "learning_rate": 1.3841950981048342e-05, | |
| "loss": 0.3522, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 1.7512983266012694, | |
| "grad_norm": 0.39607234863789387, | |
| "learning_rate": 1.3818690336014543e-05, | |
| "loss": 0.3762, | |
| "step": 3035 | |
| }, | |
| { | |
| "epoch": 1.7541834968263128, | |
| "grad_norm": 0.38127586298657323, | |
| "learning_rate": 1.3795405477165406e-05, | |
| "loss": 0.3563, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 1.757068667051356, | |
| "grad_norm": 0.43662892822050353, | |
| "learning_rate": 1.3772096552147187e-05, | |
| "loss": 0.352, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 1.7599538372763992, | |
| "grad_norm": 0.3889378354940805, | |
| "learning_rate": 1.3748763708758717e-05, | |
| "loss": 0.3686, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 1.7628390075014426, | |
| "grad_norm": 0.3972446044944282, | |
| "learning_rate": 1.3725407094950506e-05, | |
| "loss": 0.3621, | |
| "step": 3055 | |
| }, | |
| { | |
| "epoch": 1.7657241777264858, | |
| "grad_norm": 0.40962998535223355, | |
| "learning_rate": 1.3702026858823781e-05, | |
| "loss": 0.3876, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 1.768609347951529, | |
| "grad_norm": 0.35961773766923205, | |
| "learning_rate": 1.3678623148629555e-05, | |
| "loss": 0.3736, | |
| "step": 3065 | |
| }, | |
| { | |
| "epoch": 1.7714945181765724, | |
| "grad_norm": 0.38471972852000913, | |
| "learning_rate": 1.3655196112767695e-05, | |
| "loss": 0.3491, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 1.7743796884016156, | |
| "grad_norm": 0.36648434186274914, | |
| "learning_rate": 1.363174589978596e-05, | |
| "loss": 0.3593, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 1.7772648586266588, | |
| "grad_norm": 0.3872125897463841, | |
| "learning_rate": 1.3608272658379087e-05, | |
| "loss": 0.3646, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 1.7801500288517023, | |
| "grad_norm": 0.3740044921465578, | |
| "learning_rate": 1.358477653738782e-05, | |
| "loss": 0.3517, | |
| "step": 3085 | |
| }, | |
| { | |
| "epoch": 1.7830351990767457, | |
| "grad_norm": 0.38429954155990775, | |
| "learning_rate": 1.3561257685797992e-05, | |
| "loss": 0.3655, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 1.7859203693017887, | |
| "grad_norm": 0.3478740415125172, | |
| "learning_rate": 1.3537716252739552e-05, | |
| "loss": 0.3675, | |
| "step": 3095 | |
| }, | |
| { | |
| "epoch": 1.788805539526832, | |
| "grad_norm": 0.3597910283031001, | |
| "learning_rate": 1.3514152387485651e-05, | |
| "loss": 0.3489, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 1.7916907097518755, | |
| "grad_norm": 0.41193186738188275, | |
| "learning_rate": 1.3490566239451672e-05, | |
| "loss": 0.3818, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 1.7945758799769185, | |
| "grad_norm": 0.33623886878783116, | |
| "learning_rate": 1.3466957958194284e-05, | |
| "loss": 0.3618, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 1.797461050201962, | |
| "grad_norm": 0.35667893854991767, | |
| "learning_rate": 1.344332769341052e-05, | |
| "loss": 0.3558, | |
| "step": 3115 | |
| }, | |
| { | |
| "epoch": 1.8003462204270053, | |
| "grad_norm": 0.39258025755310366, | |
| "learning_rate": 1.3419675594936778e-05, | |
| "loss": 0.3477, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 1.8032313906520485, | |
| "grad_norm": 0.3782720017741446, | |
| "learning_rate": 1.339600181274793e-05, | |
| "loss": 0.3735, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 1.8061165608770917, | |
| "grad_norm": 0.4035057736179016, | |
| "learning_rate": 1.3372306496956324e-05, | |
| "loss": 0.3845, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 1.8090017311021351, | |
| "grad_norm": 0.3763574258043012, | |
| "learning_rate": 1.3348589797810854e-05, | |
| "loss": 0.3602, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 1.8118869013271783, | |
| "grad_norm": 0.35992628716248726, | |
| "learning_rate": 1.3324851865696e-05, | |
| "loss": 0.3792, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 1.8147720715522215, | |
| "grad_norm": 0.37804380856317177, | |
| "learning_rate": 1.3301092851130886e-05, | |
| "loss": 0.3657, | |
| "step": 3145 | |
| }, | |
| { | |
| "epoch": 1.817657241777265, | |
| "grad_norm": 0.3598537111947173, | |
| "learning_rate": 1.3277312904768314e-05, | |
| "loss": 0.3721, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 1.8205424120023082, | |
| "grad_norm": 0.34265734973552825, | |
| "learning_rate": 1.32535121773938e-05, | |
| "loss": 0.3673, | |
| "step": 3155 | |
| }, | |
| { | |
| "epoch": 1.8234275822273514, | |
| "grad_norm": 0.40992539604640754, | |
| "learning_rate": 1.3229690819924653e-05, | |
| "loss": 0.3738, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 1.8263127524523948, | |
| "grad_norm": 0.39170242560206425, | |
| "learning_rate": 1.3205848983408968e-05, | |
| "loss": 0.3725, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 1.829197922677438, | |
| "grad_norm": 0.4135698812295932, | |
| "learning_rate": 1.3181986819024717e-05, | |
| "loss": 0.3765, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 1.8320830929024812, | |
| "grad_norm": 0.39775337228906954, | |
| "learning_rate": 1.3158104478078753e-05, | |
| "loss": 0.3648, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 1.8349682631275246, | |
| "grad_norm": 0.4025320364227362, | |
| "learning_rate": 1.3134202112005876e-05, | |
| "loss": 0.3679, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 1.8378534333525678, | |
| "grad_norm": 0.39435133301085734, | |
| "learning_rate": 1.311027987236786e-05, | |
| "loss": 0.3766, | |
| "step": 3185 | |
| }, | |
| { | |
| "epoch": 1.840738603577611, | |
| "grad_norm": 0.3791471283767141, | |
| "learning_rate": 1.3086337910852487e-05, | |
| "loss": 0.3867, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 1.8436237738026544, | |
| "grad_norm": 0.37370201756820587, | |
| "learning_rate": 1.306237637927261e-05, | |
| "loss": 0.3669, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 1.8465089440276976, | |
| "grad_norm": 0.38003239897050844, | |
| "learning_rate": 1.3038395429565156e-05, | |
| "loss": 0.3597, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.8493941142527408, | |
| "grad_norm": 0.37745120795042336, | |
| "learning_rate": 1.3014395213790187e-05, | |
| "loss": 0.3666, | |
| "step": 3205 | |
| }, | |
| { | |
| "epoch": 1.8522792844777842, | |
| "grad_norm": 0.369849417274224, | |
| "learning_rate": 1.299037588412993e-05, | |
| "loss": 0.3731, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 1.8551644547028274, | |
| "grad_norm": 0.37108268900622904, | |
| "learning_rate": 1.2966337592887807e-05, | |
| "loss": 0.3619, | |
| "step": 3215 | |
| }, | |
| { | |
| "epoch": 1.8580496249278706, | |
| "grad_norm": 0.3603718565169327, | |
| "learning_rate": 1.2942280492487478e-05, | |
| "loss": 0.3711, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 1.860934795152914, | |
| "grad_norm": 0.38800493629769506, | |
| "learning_rate": 1.291820473547186e-05, | |
| "loss": 0.3672, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 1.8638199653779572, | |
| "grad_norm": 0.3778795059973765, | |
| "learning_rate": 1.2894110474502182e-05, | |
| "loss": 0.3665, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 1.8667051356030004, | |
| "grad_norm": 0.40908012961452245, | |
| "learning_rate": 1.2869997862356994e-05, | |
| "loss": 0.3639, | |
| "step": 3235 | |
| }, | |
| { | |
| "epoch": 1.8695903058280439, | |
| "grad_norm": 0.36111567245322745, | |
| "learning_rate": 1.2845867051931214e-05, | |
| "loss": 0.3679, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 1.8724754760530873, | |
| "grad_norm": 0.4126916975430262, | |
| "learning_rate": 1.2821718196235147e-05, | |
| "loss": 0.3577, | |
| "step": 3245 | |
| }, | |
| { | |
| "epoch": 1.8753606462781303, | |
| "grad_norm": 0.3896170917861394, | |
| "learning_rate": 1.2797551448393527e-05, | |
| "loss": 0.3533, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 1.8782458165031737, | |
| "grad_norm": 0.3918073404798387, | |
| "learning_rate": 1.2773366961644537e-05, | |
| "loss": 0.3693, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 1.881130986728217, | |
| "grad_norm": 0.36560888725527363, | |
| "learning_rate": 1.274916488933884e-05, | |
| "loss": 0.3698, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 1.88401615695326, | |
| "grad_norm": 0.35904154235729674, | |
| "learning_rate": 1.2724945384938608e-05, | |
| "loss": 0.3621, | |
| "step": 3265 | |
| }, | |
| { | |
| "epoch": 1.8869013271783035, | |
| "grad_norm": 0.3712234770214718, | |
| "learning_rate": 1.2700708602016545e-05, | |
| "loss": 0.3596, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 1.889786497403347, | |
| "grad_norm": 0.35515855306613603, | |
| "learning_rate": 1.2676454694254918e-05, | |
| "loss": 0.3436, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 1.8926716676283901, | |
| "grad_norm": 0.38038321952993087, | |
| "learning_rate": 1.265218381544458e-05, | |
| "loss": 0.3598, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 1.8955568378534333, | |
| "grad_norm": 0.38405790753936503, | |
| "learning_rate": 1.2627896119483998e-05, | |
| "loss": 0.3613, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 1.8984420080784767, | |
| "grad_norm": 0.35411934799722744, | |
| "learning_rate": 1.2603591760378266e-05, | |
| "loss": 0.3621, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 1.90132717830352, | |
| "grad_norm": 0.3833342471723807, | |
| "learning_rate": 1.2579270892238147e-05, | |
| "loss": 0.3688, | |
| "step": 3295 | |
| }, | |
| { | |
| "epoch": 1.9042123485285631, | |
| "grad_norm": 0.3783988493179844, | |
| "learning_rate": 1.2554933669279076e-05, | |
| "loss": 0.3566, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 1.9070975187536066, | |
| "grad_norm": 0.3849340817581559, | |
| "learning_rate": 1.2530580245820202e-05, | |
| "loss": 0.3664, | |
| "step": 3305 | |
| }, | |
| { | |
| "epoch": 1.9099826889786498, | |
| "grad_norm": 0.36371725434046853, | |
| "learning_rate": 1.250621077628339e-05, | |
| "loss": 0.3622, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 1.912867859203693, | |
| "grad_norm": 0.3741399204987393, | |
| "learning_rate": 1.2481825415192255e-05, | |
| "loss": 0.3381, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 1.9157530294287364, | |
| "grad_norm": 0.4669077205458527, | |
| "learning_rate": 1.245742431717118e-05, | |
| "loss": 0.3823, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 1.9186381996537796, | |
| "grad_norm": 0.35803705351389503, | |
| "learning_rate": 1.2433007636944331e-05, | |
| "loss": 0.351, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 1.9215233698788228, | |
| "grad_norm": 0.3701949796075695, | |
| "learning_rate": 1.240857552933468e-05, | |
| "loss": 0.3458, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 1.9244085401038662, | |
| "grad_norm": 0.3867704015823223, | |
| "learning_rate": 1.2384128149263023e-05, | |
| "loss": 0.372, | |
| "step": 3335 | |
| }, | |
| { | |
| "epoch": 1.9272937103289094, | |
| "grad_norm": 0.34734943471086316, | |
| "learning_rate": 1.2359665651746995e-05, | |
| "loss": 0.3587, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 1.9301788805539526, | |
| "grad_norm": 0.3456302062945545, | |
| "learning_rate": 1.2335188191900087e-05, | |
| "loss": 0.3639, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 1.933064050778996, | |
| "grad_norm": 0.37432101569615994, | |
| "learning_rate": 1.231069592493067e-05, | |
| "loss": 0.3567, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 1.9359492210040392, | |
| "grad_norm": 0.38495008542439035, | |
| "learning_rate": 1.2286189006140998e-05, | |
| "loss": 0.355, | |
| "step": 3355 | |
| }, | |
| { | |
| "epoch": 1.9388343912290824, | |
| "grad_norm": 0.3603027711889298, | |
| "learning_rate": 1.2261667590926232e-05, | |
| "loss": 0.3643, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 1.9417195614541258, | |
| "grad_norm": 0.3801200430528795, | |
| "learning_rate": 1.223713183477346e-05, | |
| "loss": 0.3547, | |
| "step": 3365 | |
| }, | |
| { | |
| "epoch": 1.944604731679169, | |
| "grad_norm": 0.3801873136595747, | |
| "learning_rate": 1.2212581893260694e-05, | |
| "loss": 0.3674, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 1.9474899019042122, | |
| "grad_norm": 0.3724561589479728, | |
| "learning_rate": 1.2188017922055901e-05, | |
| "loss": 0.3466, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 1.9503750721292556, | |
| "grad_norm": 0.38382927060326005, | |
| "learning_rate": 1.2163440076916005e-05, | |
| "loss": 0.3736, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 1.953260242354299, | |
| "grad_norm": 0.40285349364914064, | |
| "learning_rate": 1.2138848513685902e-05, | |
| "loss": 0.3518, | |
| "step": 3385 | |
| }, | |
| { | |
| "epoch": 1.956145412579342, | |
| "grad_norm": 0.39374554584138266, | |
| "learning_rate": 1.2114243388297478e-05, | |
| "loss": 0.3553, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 1.9590305828043855, | |
| "grad_norm": 0.3733836670088024, | |
| "learning_rate": 1.2089624856768603e-05, | |
| "loss": 0.3782, | |
| "step": 3395 | |
| }, | |
| { | |
| "epoch": 1.9619157530294289, | |
| "grad_norm": 0.3512295476922515, | |
| "learning_rate": 1.2064993075202172e-05, | |
| "loss": 0.352, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 1.9648009232544719, | |
| "grad_norm": 0.35755975140067886, | |
| "learning_rate": 1.2040348199785074e-05, | |
| "loss": 0.3614, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 1.9676860934795153, | |
| "grad_norm": 0.3541512834606944, | |
| "learning_rate": 1.2015690386787245e-05, | |
| "loss": 0.359, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 1.9705712637045587, | |
| "grad_norm": 0.3774486665711377, | |
| "learning_rate": 1.1991019792560648e-05, | |
| "loss": 0.3658, | |
| "step": 3415 | |
| }, | |
| { | |
| "epoch": 1.9734564339296017, | |
| "grad_norm": 0.38262821666740215, | |
| "learning_rate": 1.1966336573538287e-05, | |
| "loss": 0.3763, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 1.976341604154645, | |
| "grad_norm": 0.3903065644133358, | |
| "learning_rate": 1.1941640886233224e-05, | |
| "loss": 0.3589, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 1.9792267743796885, | |
| "grad_norm": 0.3579664105437974, | |
| "learning_rate": 1.1916932887237571e-05, | |
| "loss": 0.3594, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 1.9821119446047317, | |
| "grad_norm": 0.385483963575225, | |
| "learning_rate": 1.1892212733221523e-05, | |
| "loss": 0.3738, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 1.984997114829775, | |
| "grad_norm": 0.37602191182512, | |
| "learning_rate": 1.1867480580932336e-05, | |
| "loss": 0.3539, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 1.9878822850548183, | |
| "grad_norm": 0.40221670992777864, | |
| "learning_rate": 1.1842736587193349e-05, | |
| "loss": 0.3618, | |
| "step": 3445 | |
| }, | |
| { | |
| "epoch": 1.9907674552798615, | |
| "grad_norm": 0.3945792439685852, | |
| "learning_rate": 1.1817980908902981e-05, | |
| "loss": 0.3617, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 1.9936526255049047, | |
| "grad_norm": 0.3638645756888353, | |
| "learning_rate": 1.1793213703033755e-05, | |
| "loss": 0.3621, | |
| "step": 3455 | |
| }, | |
| { | |
| "epoch": 1.9965377957299482, | |
| "grad_norm": 0.383149019555126, | |
| "learning_rate": 1.1768435126631276e-05, | |
| "loss": 0.3596, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 1.9994229659549914, | |
| "grad_norm": 0.380227166238423, | |
| "learning_rate": 1.1743645336813248e-05, | |
| "loss": 0.3638, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.4331684410572052, | |
| "eval_runtime": 0.5935, | |
| "eval_samples_per_second": 129.744, | |
| "eval_steps_per_second": 3.37, | |
| "step": 3466 | |
| }, | |
| { | |
| "epoch": 2.0023081361800346, | |
| "grad_norm": 0.41074392868999404, | |
| "learning_rate": 1.1718844490768488e-05, | |
| "loss": 0.3104, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 2.005193306405078, | |
| "grad_norm": 0.3928008512339745, | |
| "learning_rate": 1.1694032745755906e-05, | |
| "loss": 0.261, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 2.008078476630121, | |
| "grad_norm": 0.3931528202345155, | |
| "learning_rate": 1.1669210259103535e-05, | |
| "loss": 0.287, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 2.0109636468551644, | |
| "grad_norm": 0.3869283477561703, | |
| "learning_rate": 1.164437718820751e-05, | |
| "loss": 0.2683, | |
| "step": 3485 | |
| }, | |
| { | |
| "epoch": 2.013848817080208, | |
| "grad_norm": 0.3991121860772099, | |
| "learning_rate": 1.1619533690531076e-05, | |
| "loss": 0.2695, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 2.016733987305251, | |
| "grad_norm": 0.4025523043283219, | |
| "learning_rate": 1.1594679923603606e-05, | |
| "loss": 0.2662, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 2.019619157530294, | |
| "grad_norm": 0.4441663382272395, | |
| "learning_rate": 1.1569816045019575e-05, | |
| "loss": 0.2663, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 2.0225043277553376, | |
| "grad_norm": 0.4158755932657419, | |
| "learning_rate": 1.1544942212437586e-05, | |
| "loss": 0.2701, | |
| "step": 3505 | |
| }, | |
| { | |
| "epoch": 2.025389497980381, | |
| "grad_norm": 0.4324043740872268, | |
| "learning_rate": 1.152005858357935e-05, | |
| "loss": 0.2751, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 2.028274668205424, | |
| "grad_norm": 0.41576677704649545, | |
| "learning_rate": 1.1495165316228704e-05, | |
| "loss": 0.2837, | |
| "step": 3515 | |
| }, | |
| { | |
| "epoch": 2.0311598384304674, | |
| "grad_norm": 0.4261462623114962, | |
| "learning_rate": 1.1470262568230593e-05, | |
| "loss": 0.2778, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 2.034045008655511, | |
| "grad_norm": 0.4222222859454043, | |
| "learning_rate": 1.1445350497490077e-05, | |
| "loss": 0.2663, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 2.036930178880554, | |
| "grad_norm": 0.4149492406466183, | |
| "learning_rate": 1.1420429261971342e-05, | |
| "loss": 0.2615, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 2.0398153491055973, | |
| "grad_norm": 0.4325594095101094, | |
| "learning_rate": 1.139549901969667e-05, | |
| "loss": 0.276, | |
| "step": 3535 | |
| }, | |
| { | |
| "epoch": 2.0427005193306407, | |
| "grad_norm": 0.4066105524550755, | |
| "learning_rate": 1.137055992874547e-05, | |
| "loss": 0.2749, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 2.0455856895556837, | |
| "grad_norm": 0.4091251188163908, | |
| "learning_rate": 1.1345612147253247e-05, | |
| "loss": 0.2662, | |
| "step": 3545 | |
| }, | |
| { | |
| "epoch": 2.048470859780727, | |
| "grad_norm": 0.3948683403145309, | |
| "learning_rate": 1.1320655833410621e-05, | |
| "loss": 0.2813, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 2.0513560300057705, | |
| "grad_norm": 0.41481838952861994, | |
| "learning_rate": 1.1295691145462306e-05, | |
| "loss": 0.2863, | |
| "step": 3555 | |
| }, | |
| { | |
| "epoch": 2.0542412002308135, | |
| "grad_norm": 0.4032363652268982, | |
| "learning_rate": 1.1270718241706118e-05, | |
| "loss": 0.2565, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 2.057126370455857, | |
| "grad_norm": 0.48615644869007646, | |
| "learning_rate": 1.1245737280491973e-05, | |
| "loss": 0.2856, | |
| "step": 3565 | |
| }, | |
| { | |
| "epoch": 2.0600115406809003, | |
| "grad_norm": 0.4219271094611073, | |
| "learning_rate": 1.122074842022087e-05, | |
| "loss": 0.2719, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 2.0628967109059433, | |
| "grad_norm": 0.41257650256952855, | |
| "learning_rate": 1.1195751819343906e-05, | |
| "loss": 0.2745, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 2.0657818811309867, | |
| "grad_norm": 0.428723721941346, | |
| "learning_rate": 1.1170747636361245e-05, | |
| "loss": 0.2779, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 2.06866705135603, | |
| "grad_norm": 0.40354050541964437, | |
| "learning_rate": 1.1145736029821148e-05, | |
| "loss": 0.283, | |
| "step": 3585 | |
| }, | |
| { | |
| "epoch": 2.071552221581073, | |
| "grad_norm": 0.43472195003969, | |
| "learning_rate": 1.1120717158318928e-05, | |
| "loss": 0.281, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 2.0744373918061165, | |
| "grad_norm": 0.4123476736485213, | |
| "learning_rate": 1.1095691180495973e-05, | |
| "loss": 0.2798, | |
| "step": 3595 | |
| }, | |
| { | |
| "epoch": 2.07732256203116, | |
| "grad_norm": 0.4150389828010794, | |
| "learning_rate": 1.1070658255038733e-05, | |
| "loss": 0.2871, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 2.080207732256203, | |
| "grad_norm": 0.4712720490312681, | |
| "learning_rate": 1.1045618540677712e-05, | |
| "loss": 0.2815, | |
| "step": 3605 | |
| }, | |
| { | |
| "epoch": 2.0830929024812463, | |
| "grad_norm": 0.5081431777602221, | |
| "learning_rate": 1.1020572196186457e-05, | |
| "loss": 0.263, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 2.0859780727062898, | |
| "grad_norm": 0.4549015236527131, | |
| "learning_rate": 1.0995519380380557e-05, | |
| "loss": 0.2717, | |
| "step": 3615 | |
| }, | |
| { | |
| "epoch": 2.0888632429313327, | |
| "grad_norm": 0.5033902066650326, | |
| "learning_rate": 1.097046025211664e-05, | |
| "loss": 0.2813, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 2.091748413156376, | |
| "grad_norm": 0.40826808731777586, | |
| "learning_rate": 1.0945394970291354e-05, | |
| "loss": 0.2607, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 2.0946335833814196, | |
| "grad_norm": 0.4593185041491404, | |
| "learning_rate": 1.0920323693840368e-05, | |
| "loss": 0.28, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 2.0975187536064626, | |
| "grad_norm": 0.4374169481716498, | |
| "learning_rate": 1.0895246581737366e-05, | |
| "loss": 0.2968, | |
| "step": 3635 | |
| }, | |
| { | |
| "epoch": 2.100403923831506, | |
| "grad_norm": 0.4281118948367996, | |
| "learning_rate": 1.087016379299303e-05, | |
| "loss": 0.2877, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 2.1032890940565494, | |
| "grad_norm": 0.45375611915195996, | |
| "learning_rate": 1.0845075486654037e-05, | |
| "loss": 0.261, | |
| "step": 3645 | |
| }, | |
| { | |
| "epoch": 2.106174264281593, | |
| "grad_norm": 0.4267355049796937, | |
| "learning_rate": 1.0819981821802053e-05, | |
| "loss": 0.2786, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 2.109059434506636, | |
| "grad_norm": 0.4376492488121602, | |
| "learning_rate": 1.0794882957552722e-05, | |
| "loss": 0.2766, | |
| "step": 3655 | |
| }, | |
| { | |
| "epoch": 2.111944604731679, | |
| "grad_norm": 0.45029828413170214, | |
| "learning_rate": 1.0769779053054651e-05, | |
| "loss": 0.283, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 2.1148297749567226, | |
| "grad_norm": 0.43071600979778896, | |
| "learning_rate": 1.0744670267488417e-05, | |
| "loss": 0.2848, | |
| "step": 3665 | |
| }, | |
| { | |
| "epoch": 2.1177149451817656, | |
| "grad_norm": 0.3970843751114558, | |
| "learning_rate": 1.071955676006554e-05, | |
| "loss": 0.2762, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 2.120600115406809, | |
| "grad_norm": 0.38320271641559517, | |
| "learning_rate": 1.0694438690027475e-05, | |
| "loss": 0.2819, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 2.1234852856318525, | |
| "grad_norm": 0.4529791521783765, | |
| "learning_rate": 1.0669316216644623e-05, | |
| "loss": 0.266, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 2.1263704558568954, | |
| "grad_norm": 0.42862083344692714, | |
| "learning_rate": 1.0644189499215292e-05, | |
| "loss": 0.2765, | |
| "step": 3685 | |
| }, | |
| { | |
| "epoch": 2.129255626081939, | |
| "grad_norm": 0.43384197829762733, | |
| "learning_rate": 1.061905869706471e-05, | |
| "loss": 0.3061, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 2.1321407963069823, | |
| "grad_norm": 0.4224346600569041, | |
| "learning_rate": 1.0593923969544e-05, | |
| "loss": 0.3003, | |
| "step": 3695 | |
| }, | |
| { | |
| "epoch": 2.1350259665320253, | |
| "grad_norm": 0.4332701906364038, | |
| "learning_rate": 1.056878547602918e-05, | |
| "loss": 0.2765, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 2.1379111367570687, | |
| "grad_norm": 0.4393240656228905, | |
| "learning_rate": 1.0543643375920145e-05, | |
| "loss": 0.2879, | |
| "step": 3705 | |
| }, | |
| { | |
| "epoch": 2.140796306982112, | |
| "grad_norm": 0.4036822061189937, | |
| "learning_rate": 1.0518497828639658e-05, | |
| "loss": 0.27, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 2.143681477207155, | |
| "grad_norm": 0.427794972829589, | |
| "learning_rate": 1.0493348993632337e-05, | |
| "loss": 0.2749, | |
| "step": 3715 | |
| }, | |
| { | |
| "epoch": 2.1465666474321985, | |
| "grad_norm": 0.39124739436748485, | |
| "learning_rate": 1.046819703036366e-05, | |
| "loss": 0.2718, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 2.149451817657242, | |
| "grad_norm": 0.4099521410733268, | |
| "learning_rate": 1.0443042098318927e-05, | |
| "loss": 0.2733, | |
| "step": 3725 | |
| }, | |
| { | |
| "epoch": 2.152336987882285, | |
| "grad_norm": 0.4225255601400179, | |
| "learning_rate": 1.0417884357002268e-05, | |
| "loss": 0.2773, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 2.1552221581073283, | |
| "grad_norm": 0.43412449015519566, | |
| "learning_rate": 1.039272396593563e-05, | |
| "loss": 0.2544, | |
| "step": 3735 | |
| }, | |
| { | |
| "epoch": 2.1581073283323717, | |
| "grad_norm": 0.414617071495182, | |
| "learning_rate": 1.0367561084657752e-05, | |
| "loss": 0.2672, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 2.1609924985574147, | |
| "grad_norm": 0.4451465025534693, | |
| "learning_rate": 1.0342395872723173e-05, | |
| "loss": 0.2597, | |
| "step": 3745 | |
| }, | |
| { | |
| "epoch": 2.163877668782458, | |
| "grad_norm": 0.4052440261347642, | |
| "learning_rate": 1.03172284897012e-05, | |
| "loss": 0.2758, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 2.1667628390075016, | |
| "grad_norm": 0.4393651404130225, | |
| "learning_rate": 1.0292059095174923e-05, | |
| "loss": 0.281, | |
| "step": 3755 | |
| }, | |
| { | |
| "epoch": 2.1696480092325445, | |
| "grad_norm": 0.43033068987713674, | |
| "learning_rate": 1.026688784874017e-05, | |
| "loss": 0.2839, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 2.172533179457588, | |
| "grad_norm": 0.4187797584558774, | |
| "learning_rate": 1.024171491000452e-05, | |
| "loss": 0.2955, | |
| "step": 3765 | |
| }, | |
| { | |
| "epoch": 2.1754183496826314, | |
| "grad_norm": 0.4306833753003216, | |
| "learning_rate": 1.021654043858628e-05, | |
| "loss": 0.268, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 2.1783035199076743, | |
| "grad_norm": 0.46354342504920265, | |
| "learning_rate": 1.0191364594113475e-05, | |
| "loss": 0.2753, | |
| "step": 3775 | |
| }, | |
| { | |
| "epoch": 2.1811886901327178, | |
| "grad_norm": 0.4667906762400645, | |
| "learning_rate": 1.0166187536222844e-05, | |
| "loss": 0.2826, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 2.184073860357761, | |
| "grad_norm": 0.4268396541209257, | |
| "learning_rate": 1.0141009424558803e-05, | |
| "loss": 0.2594, | |
| "step": 3785 | |
| }, | |
| { | |
| "epoch": 2.1869590305828046, | |
| "grad_norm": 0.42773226782347973, | |
| "learning_rate": 1.0115830418772474e-05, | |
| "loss": 0.2815, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 2.1898442008078476, | |
| "grad_norm": 0.42670690759544905, | |
| "learning_rate": 1.0090650678520624e-05, | |
| "loss": 0.2884, | |
| "step": 3795 | |
| }, | |
| { | |
| "epoch": 2.192729371032891, | |
| "grad_norm": 0.43803219635914387, | |
| "learning_rate": 1.0065470363464694e-05, | |
| "loss": 0.2789, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 2.1956145412579344, | |
| "grad_norm": 0.4799532890386973, | |
| "learning_rate": 1.0040289633269764e-05, | |
| "loss": 0.2945, | |
| "step": 3805 | |
| }, | |
| { | |
| "epoch": 2.1984997114829774, | |
| "grad_norm": 0.4207961621120341, | |
| "learning_rate": 1.0015108647603545e-05, | |
| "loss": 0.2665, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 2.201384881708021, | |
| "grad_norm": 0.43772623733868893, | |
| "learning_rate": 9.98992756613537e-06, | |
| "loss": 0.2791, | |
| "step": 3815 | |
| }, | |
| { | |
| "epoch": 2.2042700519330642, | |
| "grad_norm": 0.4246990837045649, | |
| "learning_rate": 9.96474654853518e-06, | |
| "loss": 0.2702, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 2.207155222158107, | |
| "grad_norm": 0.4214383786492429, | |
| "learning_rate": 9.93956575447251e-06, | |
| "loss": 0.2724, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 2.2100403923831506, | |
| "grad_norm": 0.42698570251672047, | |
| "learning_rate": 9.914385343615476e-06, | |
| "loss": 0.2714, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 2.212925562608194, | |
| "grad_norm": 0.4504000608499869, | |
| "learning_rate": 9.889205475629763e-06, | |
| "loss": 0.2754, | |
| "step": 3835 | |
| }, | |
| { | |
| "epoch": 2.215810732833237, | |
| "grad_norm": 0.4323598292397766, | |
| "learning_rate": 9.864026310177621e-06, | |
| "loss": 0.2624, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 2.2186959030582805, | |
| "grad_norm": 0.40922887326049673, | |
| "learning_rate": 9.83884800691684e-06, | |
| "loss": 0.2633, | |
| "step": 3845 | |
| }, | |
| { | |
| "epoch": 2.221581073283324, | |
| "grad_norm": 0.4357930415869159, | |
| "learning_rate": 9.813670725499745e-06, | |
| "loss": 0.2904, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 2.224466243508367, | |
| "grad_norm": 0.47764591829975994, | |
| "learning_rate": 9.788494625572177e-06, | |
| "loss": 0.2685, | |
| "step": 3855 | |
| }, | |
| { | |
| "epoch": 2.2273514137334103, | |
| "grad_norm": 0.4306543758497723, | |
| "learning_rate": 9.763319866772486e-06, | |
| "loss": 0.2757, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 2.2302365839584537, | |
| "grad_norm": 0.451979701648703, | |
| "learning_rate": 9.738146608730527e-06, | |
| "loss": 0.281, | |
| "step": 3865 | |
| }, | |
| { | |
| "epoch": 2.2331217541834967, | |
| "grad_norm": 0.39863721704621635, | |
| "learning_rate": 9.712975011066633e-06, | |
| "loss": 0.2651, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 2.23600692440854, | |
| "grad_norm": 0.4431702895497509, | |
| "learning_rate": 9.687805233390602e-06, | |
| "loss": 0.2875, | |
| "step": 3875 | |
| }, | |
| { | |
| "epoch": 2.2388920946335835, | |
| "grad_norm": 0.42463992002267936, | |
| "learning_rate": 9.662637435300702e-06, | |
| "loss": 0.2747, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 2.2417772648586265, | |
| "grad_norm": 0.4421016875554418, | |
| "learning_rate": 9.637471776382642e-06, | |
| "loss": 0.3014, | |
| "step": 3885 | |
| }, | |
| { | |
| "epoch": 2.24466243508367, | |
| "grad_norm": 0.4309775900999432, | |
| "learning_rate": 9.612308416208573e-06, | |
| "loss": 0.3009, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 2.2475476053087133, | |
| "grad_norm": 0.38604226765745275, | |
| "learning_rate": 9.587147514336068e-06, | |
| "loss": 0.2784, | |
| "step": 3895 | |
| }, | |
| { | |
| "epoch": 2.2504327755337563, | |
| "grad_norm": 0.4812969089447562, | |
| "learning_rate": 9.561989230307106e-06, | |
| "loss": 0.2768, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 2.2533179457587997, | |
| "grad_norm": 0.42601904264725454, | |
| "learning_rate": 9.536833723647073e-06, | |
| "loss": 0.2797, | |
| "step": 3905 | |
| }, | |
| { | |
| "epoch": 2.256203115983843, | |
| "grad_norm": 0.4357929404581439, | |
| "learning_rate": 9.51168115386374e-06, | |
| "loss": 0.2785, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 2.259088286208886, | |
| "grad_norm": 0.41501335099687836, | |
| "learning_rate": 9.486531680446266e-06, | |
| "loss": 0.2674, | |
| "step": 3915 | |
| }, | |
| { | |
| "epoch": 2.2619734564339296, | |
| "grad_norm": 0.4259900039138935, | |
| "learning_rate": 9.46138546286416e-06, | |
| "loss": 0.2659, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 2.264858626658973, | |
| "grad_norm": 0.4619671751542806, | |
| "learning_rate": 9.436242660566299e-06, | |
| "loss": 0.266, | |
| "step": 3925 | |
| }, | |
| { | |
| "epoch": 2.267743796884016, | |
| "grad_norm": 0.4284866042642329, | |
| "learning_rate": 9.411103432979895e-06, | |
| "loss": 0.277, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 2.2706289671090594, | |
| "grad_norm": 0.4405127132258486, | |
| "learning_rate": 9.385967939509503e-06, | |
| "loss": 0.2639, | |
| "step": 3935 | |
| }, | |
| { | |
| "epoch": 2.273514137334103, | |
| "grad_norm": 0.41142311165199374, | |
| "learning_rate": 9.360836339535997e-06, | |
| "loss": 0.2898, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 2.2763993075591458, | |
| "grad_norm": 0.41387485994040385, | |
| "learning_rate": 9.335708792415554e-06, | |
| "loss": 0.2831, | |
| "step": 3945 | |
| }, | |
| { | |
| "epoch": 2.279284477784189, | |
| "grad_norm": 0.4321573127781368, | |
| "learning_rate": 9.310585457478665e-06, | |
| "loss": 0.2667, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 2.2821696480092326, | |
| "grad_norm": 0.4184357007573235, | |
| "learning_rate": 9.285466494029103e-06, | |
| "loss": 0.2725, | |
| "step": 3955 | |
| }, | |
| { | |
| "epoch": 2.2850548182342756, | |
| "grad_norm": 0.4338603810054309, | |
| "learning_rate": 9.260352061342928e-06, | |
| "loss": 0.2788, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 2.287939988459319, | |
| "grad_norm": 0.5970591619349905, | |
| "learning_rate": 9.235242318667473e-06, | |
| "loss": 0.2679, | |
| "step": 3965 | |
| }, | |
| { | |
| "epoch": 2.2908251586843624, | |
| "grad_norm": 0.41404764849303466, | |
| "learning_rate": 9.21013742522032e-06, | |
| "loss": 0.2638, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 2.293710328909406, | |
| "grad_norm": 0.44951250598474296, | |
| "learning_rate": 9.185037540188319e-06, | |
| "loss": 0.2674, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 2.296595499134449, | |
| "grad_norm": 0.42037037135886823, | |
| "learning_rate": 9.159942822726539e-06, | |
| "loss": 0.2713, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 2.2994806693594922, | |
| "grad_norm": 0.42502356412209286, | |
| "learning_rate": 9.134853431957316e-06, | |
| "loss": 0.2648, | |
| "step": 3985 | |
| }, | |
| { | |
| "epoch": 2.3023658395845357, | |
| "grad_norm": 0.4216737901866948, | |
| "learning_rate": 9.109769526969179e-06, | |
| "loss": 0.2724, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 2.3052510098095786, | |
| "grad_norm": 0.4167455267572832, | |
| "learning_rate": 9.084691266815888e-06, | |
| "loss": 0.2766, | |
| "step": 3995 | |
| }, | |
| { | |
| "epoch": 2.308136180034622, | |
| "grad_norm": 0.45897408948064594, | |
| "learning_rate": 9.059618810515405e-06, | |
| "loss": 0.2781, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 2.3110213502596655, | |
| "grad_norm": 0.42036299610150296, | |
| "learning_rate": 9.034552317048895e-06, | |
| "loss": 0.2707, | |
| "step": 4005 | |
| }, | |
| { | |
| "epoch": 2.3139065204847085, | |
| "grad_norm": 0.4277783623427705, | |
| "learning_rate": 9.009491945359711e-06, | |
| "loss": 0.2832, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 2.316791690709752, | |
| "grad_norm": 0.421315487309371, | |
| "learning_rate": 8.984437854352384e-06, | |
| "loss": 0.2754, | |
| "step": 4015 | |
| }, | |
| { | |
| "epoch": 2.3196768609347953, | |
| "grad_norm": 0.4657256225278301, | |
| "learning_rate": 8.959390202891625e-06, | |
| "loss": 0.2912, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 2.3225620311598383, | |
| "grad_norm": 0.4272698374442868, | |
| "learning_rate": 8.93434914980131e-06, | |
| "loss": 0.2662, | |
| "step": 4025 | |
| }, | |
| { | |
| "epoch": 2.3254472013848817, | |
| "grad_norm": 0.4198417527704976, | |
| "learning_rate": 8.909314853863483e-06, | |
| "loss": 0.2727, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 2.328332371609925, | |
| "grad_norm": 0.5124921390099292, | |
| "learning_rate": 8.884287473817332e-06, | |
| "loss": 0.2708, | |
| "step": 4035 | |
| }, | |
| { | |
| "epoch": 2.331217541834968, | |
| "grad_norm": 0.42269744019953115, | |
| "learning_rate": 8.859267168358198e-06, | |
| "loss": 0.2745, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 2.3341027120600115, | |
| "grad_norm": 0.42318242166255127, | |
| "learning_rate": 8.834254096136561e-06, | |
| "loss": 0.2949, | |
| "step": 4045 | |
| }, | |
| { | |
| "epoch": 2.336987882285055, | |
| "grad_norm": 0.419851745283628, | |
| "learning_rate": 8.809248415757028e-06, | |
| "loss": 0.2795, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 2.339873052510098, | |
| "grad_norm": 0.44233002428351004, | |
| "learning_rate": 8.784250285777361e-06, | |
| "loss": 0.2719, | |
| "step": 4055 | |
| }, | |
| { | |
| "epoch": 2.3427582227351413, | |
| "grad_norm": 0.396083210912282, | |
| "learning_rate": 8.759259864707415e-06, | |
| "loss": 0.2697, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 2.3456433929601848, | |
| "grad_norm": 0.47094149055802703, | |
| "learning_rate": 8.73427731100818e-06, | |
| "loss": 0.2724, | |
| "step": 4065 | |
| }, | |
| { | |
| "epoch": 2.3485285631852277, | |
| "grad_norm": 0.42603022428645243, | |
| "learning_rate": 8.70930278309076e-06, | |
| "loss": 0.2654, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 2.351413733410271, | |
| "grad_norm": 0.4379509456011344, | |
| "learning_rate": 8.684336439315362e-06, | |
| "loss": 0.2786, | |
| "step": 4075 | |
| }, | |
| { | |
| "epoch": 2.3542989036353146, | |
| "grad_norm": 0.426972682822357, | |
| "learning_rate": 8.659378437990305e-06, | |
| "loss": 0.2852, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 2.357184073860358, | |
| "grad_norm": 0.42210417306175224, | |
| "learning_rate": 8.634428937371008e-06, | |
| "loss": 0.2777, | |
| "step": 4085 | |
| }, | |
| { | |
| "epoch": 2.360069244085401, | |
| "grad_norm": 0.4490344314229441, | |
| "learning_rate": 8.609488095658987e-06, | |
| "loss": 0.2732, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 2.3629544143104444, | |
| "grad_norm": 0.44262847306870384, | |
| "learning_rate": 8.584556071000845e-06, | |
| "loss": 0.2724, | |
| "step": 4095 | |
| }, | |
| { | |
| "epoch": 2.365839584535488, | |
| "grad_norm": 0.43681619099011737, | |
| "learning_rate": 8.559633021487298e-06, | |
| "loss": 0.291, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 2.368724754760531, | |
| "grad_norm": 0.669470605388525, | |
| "learning_rate": 8.53471910515213e-06, | |
| "loss": 0.29, | |
| "step": 4105 | |
| }, | |
| { | |
| "epoch": 2.371609924985574, | |
| "grad_norm": 0.4872795075157814, | |
| "learning_rate": 8.509814479971226e-06, | |
| "loss": 0.293, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 2.3744950952106176, | |
| "grad_norm": 0.4330590098492678, | |
| "learning_rate": 8.484919303861555e-06, | |
| "loss": 0.2794, | |
| "step": 4115 | |
| }, | |
| { | |
| "epoch": 2.3773802654356606, | |
| "grad_norm": 0.42534287206042853, | |
| "learning_rate": 8.460033734680158e-06, | |
| "loss": 0.2766, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 2.380265435660704, | |
| "grad_norm": 0.43174048146055444, | |
| "learning_rate": 8.43515793022318e-06, | |
| "loss": 0.2797, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 2.3831506058857475, | |
| "grad_norm": 0.4429583272831574, | |
| "learning_rate": 8.410292048224838e-06, | |
| "loss": 0.269, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 2.3860357761107904, | |
| "grad_norm": 0.41491168931306777, | |
| "learning_rate": 8.38543624635643e-06, | |
| "loss": 0.2771, | |
| "step": 4135 | |
| }, | |
| { | |
| "epoch": 2.388920946335834, | |
| "grad_norm": 0.4490843763025565, | |
| "learning_rate": 8.360590682225345e-06, | |
| "loss": 0.2775, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 2.3918061165608773, | |
| "grad_norm": 0.4162339442858775, | |
| "learning_rate": 8.335755513374044e-06, | |
| "loss": 0.2725, | |
| "step": 4145 | |
| }, | |
| { | |
| "epoch": 2.3946912867859202, | |
| "grad_norm": 0.3966357439538777, | |
| "learning_rate": 8.310930897279088e-06, | |
| "loss": 0.2824, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 2.3975764570109637, | |
| "grad_norm": 0.43243463473106125, | |
| "learning_rate": 8.286116991350113e-06, | |
| "loss": 0.273, | |
| "step": 4155 | |
| }, | |
| { | |
| "epoch": 2.400461627236007, | |
| "grad_norm": 0.40628353100856796, | |
| "learning_rate": 8.261313952928853e-06, | |
| "loss": 0.2697, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 2.40334679746105, | |
| "grad_norm": 0.41821368197595044, | |
| "learning_rate": 8.236521939288116e-06, | |
| "loss": 0.2722, | |
| "step": 4165 | |
| }, | |
| { | |
| "epoch": 2.4062319676860935, | |
| "grad_norm": 0.43315603245507267, | |
| "learning_rate": 8.211741107630817e-06, | |
| "loss": 0.2811, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 2.409117137911137, | |
| "grad_norm": 0.4192983792279987, | |
| "learning_rate": 8.186971615088967e-06, | |
| "loss": 0.27, | |
| "step": 4175 | |
| }, | |
| { | |
| "epoch": 2.41200230813618, | |
| "grad_norm": 0.4228869690317952, | |
| "learning_rate": 8.162213618722679e-06, | |
| "loss": 0.2936, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 2.4148874783612233, | |
| "grad_norm": 0.4264461059873532, | |
| "learning_rate": 8.137467275519156e-06, | |
| "loss": 0.255, | |
| "step": 4185 | |
| }, | |
| { | |
| "epoch": 2.4177726485862667, | |
| "grad_norm": 0.42183959270214744, | |
| "learning_rate": 8.112732742391722e-06, | |
| "loss": 0.265, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 2.4206578188113097, | |
| "grad_norm": 0.440409389917108, | |
| "learning_rate": 8.088010176178816e-06, | |
| "loss": 0.2895, | |
| "step": 4195 | |
| }, | |
| { | |
| "epoch": 2.423542989036353, | |
| "grad_norm": 0.4190921915499105, | |
| "learning_rate": 8.063299733642991e-06, | |
| "loss": 0.268, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 2.4264281592613965, | |
| "grad_norm": 0.4370598207669186, | |
| "learning_rate": 8.03860157146993e-06, | |
| "loss": 0.2764, | |
| "step": 4205 | |
| }, | |
| { | |
| "epoch": 2.4293133294864395, | |
| "grad_norm": 0.4607794338360012, | |
| "learning_rate": 8.013915846267439e-06, | |
| "loss": 0.2863, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 2.432198499711483, | |
| "grad_norm": 0.4378000785386685, | |
| "learning_rate": 7.98924271456447e-06, | |
| "loss": 0.2747, | |
| "step": 4215 | |
| }, | |
| { | |
| "epoch": 2.4350836699365264, | |
| "grad_norm": 0.4598949217946755, | |
| "learning_rate": 7.964582332810122e-06, | |
| "loss": 0.2756, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 2.4379688401615693, | |
| "grad_norm": 0.4305567337108496, | |
| "learning_rate": 7.939934857372646e-06, | |
| "loss": 0.2734, | |
| "step": 4225 | |
| }, | |
| { | |
| "epoch": 2.4408540103866128, | |
| "grad_norm": 0.42896614513275355, | |
| "learning_rate": 7.915300444538458e-06, | |
| "loss": 0.2704, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 2.443739180611656, | |
| "grad_norm": 0.4104247108771218, | |
| "learning_rate": 7.89067925051114e-06, | |
| "loss": 0.2683, | |
| "step": 4235 | |
| }, | |
| { | |
| "epoch": 2.446624350836699, | |
| "grad_norm": 0.42265110955709223, | |
| "learning_rate": 7.866071431410458e-06, | |
| "loss": 0.2834, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 2.4495095210617426, | |
| "grad_norm": 0.4025030251156565, | |
| "learning_rate": 7.841477143271374e-06, | |
| "loss": 0.2719, | |
| "step": 4245 | |
| }, | |
| { | |
| "epoch": 2.452394691286786, | |
| "grad_norm": 0.4347245013374062, | |
| "learning_rate": 7.816896542043048e-06, | |
| "loss": 0.2849, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 2.455279861511829, | |
| "grad_norm": 0.3970293249450069, | |
| "learning_rate": 7.792329783587853e-06, | |
| "loss": 0.281, | |
| "step": 4255 | |
| }, | |
| { | |
| "epoch": 2.4581650317368724, | |
| "grad_norm": 0.5089714627565546, | |
| "learning_rate": 7.767777023680384e-06, | |
| "loss": 0.2886, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 2.461050201961916, | |
| "grad_norm": 0.42036661292632566, | |
| "learning_rate": 7.743238418006476e-06, | |
| "loss": 0.274, | |
| "step": 4265 | |
| }, | |
| { | |
| "epoch": 2.463935372186959, | |
| "grad_norm": 0.41995706488380274, | |
| "learning_rate": 7.71871412216222e-06, | |
| "loss": 0.2834, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 2.466820542412002, | |
| "grad_norm": 0.405500972649185, | |
| "learning_rate": 7.694204291652962e-06, | |
| "loss": 0.2723, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 2.4697057126370456, | |
| "grad_norm": 0.47474988784176503, | |
| "learning_rate": 7.669709081892324e-06, | |
| "loss": 0.276, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 2.472590882862089, | |
| "grad_norm": 0.4208990452736956, | |
| "learning_rate": 7.645228648201225e-06, | |
| "loss": 0.2686, | |
| "step": 4285 | |
| }, | |
| { | |
| "epoch": 2.475476053087132, | |
| "grad_norm": 0.5125940516643855, | |
| "learning_rate": 7.62076314580689e-06, | |
| "loss": 0.2791, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 2.4783612233121755, | |
| "grad_norm": 0.43397963005803125, | |
| "learning_rate": 7.596312729841868e-06, | |
| "loss": 0.2714, | |
| "step": 4295 | |
| }, | |
| { | |
| "epoch": 2.481246393537219, | |
| "grad_norm": 0.43916382261638937, | |
| "learning_rate": 7.571877555343039e-06, | |
| "loss": 0.2953, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 2.484131563762262, | |
| "grad_norm": 0.42968851926015045, | |
| "learning_rate": 7.547457777250649e-06, | |
| "loss": 0.2915, | |
| "step": 4305 | |
| }, | |
| { | |
| "epoch": 2.4870167339873053, | |
| "grad_norm": 0.4334658833975296, | |
| "learning_rate": 7.523053550407308e-06, | |
| "loss": 0.2781, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 2.4899019042123487, | |
| "grad_norm": 0.442518451619476, | |
| "learning_rate": 7.498665029557028e-06, | |
| "loss": 0.2795, | |
| "step": 4315 | |
| }, | |
| { | |
| "epoch": 2.4927870744373917, | |
| "grad_norm": 0.4381590289442612, | |
| "learning_rate": 7.474292369344226e-06, | |
| "loss": 0.2641, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 2.495672244662435, | |
| "grad_norm": 0.44001237345403327, | |
| "learning_rate": 7.449935724312745e-06, | |
| "loss": 0.2688, | |
| "step": 4325 | |
| }, | |
| { | |
| "epoch": 2.4985574148874785, | |
| "grad_norm": 0.43261375909491934, | |
| "learning_rate": 7.42559524890488e-06, | |
| "loss": 0.2791, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 2.5014425851125215, | |
| "grad_norm": 0.41821164862411353, | |
| "learning_rate": 7.401271097460401e-06, | |
| "loss": 0.2753, | |
| "step": 4335 | |
| }, | |
| { | |
| "epoch": 2.504327755337565, | |
| "grad_norm": 0.4284156665016015, | |
| "learning_rate": 7.376963424215568e-06, | |
| "loss": 0.2916, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 2.5072129255626083, | |
| "grad_norm": 0.42595494161348285, | |
| "learning_rate": 7.352672383302153e-06, | |
| "loss": 0.2587, | |
| "step": 4345 | |
| }, | |
| { | |
| "epoch": 2.5100980957876513, | |
| "grad_norm": 0.4234233587645987, | |
| "learning_rate": 7.328398128746463e-06, | |
| "loss": 0.2723, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 2.5129832660126947, | |
| "grad_norm": 0.4206786678316223, | |
| "learning_rate": 7.304140814468365e-06, | |
| "loss": 0.2737, | |
| "step": 4355 | |
| }, | |
| { | |
| "epoch": 2.515868436237738, | |
| "grad_norm": 0.4421492786532555, | |
| "learning_rate": 7.27990059428032e-06, | |
| "loss": 0.2871, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 2.5187536064627816, | |
| "grad_norm": 0.4071794979099346, | |
| "learning_rate": 7.255677621886388e-06, | |
| "loss": 0.2669, | |
| "step": 4365 | |
| }, | |
| { | |
| "epoch": 2.5216387766878245, | |
| "grad_norm": 0.4085099532665915, | |
| "learning_rate": 7.23147205088126e-06, | |
| "loss": 0.2749, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 2.524523946912868, | |
| "grad_norm": 0.4540968719545943, | |
| "learning_rate": 7.207284034749294e-06, | |
| "loss": 0.2784, | |
| "step": 4375 | |
| }, | |
| { | |
| "epoch": 2.5274091171379114, | |
| "grad_norm": 0.43868950738012735, | |
| "learning_rate": 7.183113726863529e-06, | |
| "loss": 0.2817, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 2.5302942873629544, | |
| "grad_norm": 0.5603667780494723, | |
| "learning_rate": 7.15896128048473e-06, | |
| "loss": 0.2779, | |
| "step": 4385 | |
| }, | |
| { | |
| "epoch": 2.533179457587998, | |
| "grad_norm": 0.4643684780575487, | |
| "learning_rate": 7.134826848760392e-06, | |
| "loss": 0.2704, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 2.536064627813041, | |
| "grad_norm": 0.4311580857729655, | |
| "learning_rate": 7.110710584723785e-06, | |
| "loss": 0.2773, | |
| "step": 4395 | |
| }, | |
| { | |
| "epoch": 2.538949798038084, | |
| "grad_norm": 0.41665610893096905, | |
| "learning_rate": 7.086612641292985e-06, | |
| "loss": 0.2697, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 2.5418349682631276, | |
| "grad_norm": 0.4448150891591054, | |
| "learning_rate": 7.062533171269887e-06, | |
| "loss": 0.2679, | |
| "step": 4405 | |
| }, | |
| { | |
| "epoch": 2.544720138488171, | |
| "grad_norm": 0.4485899811392423, | |
| "learning_rate": 7.038472327339268e-06, | |
| "loss": 0.2624, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 2.547605308713214, | |
| "grad_norm": 0.4694083142624704, | |
| "learning_rate": 7.0144302620677804e-06, | |
| "loss": 0.2671, | |
| "step": 4415 | |
| }, | |
| { | |
| "epoch": 2.5504904789382574, | |
| "grad_norm": 0.44085510653638454, | |
| "learning_rate": 6.990407127903015e-06, | |
| "loss": 0.2741, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 2.553375649163301, | |
| "grad_norm": 0.4331688809785418, | |
| "learning_rate": 6.966403077172521e-06, | |
| "loss": 0.2784, | |
| "step": 4425 | |
| }, | |
| { | |
| "epoch": 2.556260819388344, | |
| "grad_norm": 0.42326748215143545, | |
| "learning_rate": 6.942418262082833e-06, | |
| "loss": 0.2803, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 2.5591459896133872, | |
| "grad_norm": 0.4579702510573031, | |
| "learning_rate": 6.91845283471853e-06, | |
| "loss": 0.2629, | |
| "step": 4435 | |
| }, | |
| { | |
| "epoch": 2.5620311598384307, | |
| "grad_norm": 0.41574752352195643, | |
| "learning_rate": 6.894506947041246e-06, | |
| "loss": 0.2606, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 2.5649163300634736, | |
| "grad_norm": 0.4365002129279704, | |
| "learning_rate": 6.870580750888722e-06, | |
| "loss": 0.2724, | |
| "step": 4445 | |
| }, | |
| { | |
| "epoch": 2.567801500288517, | |
| "grad_norm": 0.4399394455174853, | |
| "learning_rate": 6.846674397973825e-06, | |
| "loss": 0.2649, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 2.5706866705135605, | |
| "grad_norm": 0.447371975699018, | |
| "learning_rate": 6.822788039883621e-06, | |
| "loss": 0.2684, | |
| "step": 4455 | |
| }, | |
| { | |
| "epoch": 2.5735718407386035, | |
| "grad_norm": 0.3961735893626953, | |
| "learning_rate": 6.7989218280783686e-06, | |
| "loss": 0.2587, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 2.576457010963647, | |
| "grad_norm": 0.44985375940450545, | |
| "learning_rate": 6.775075913890597e-06, | |
| "loss": 0.2613, | |
| "step": 4465 | |
| }, | |
| { | |
| "epoch": 2.5793421811886903, | |
| "grad_norm": 0.4488817173284117, | |
| "learning_rate": 6.751250448524123e-06, | |
| "loss": 0.2814, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 2.5822273514137333, | |
| "grad_norm": 0.5139233580575341, | |
| "learning_rate": 6.727445583053096e-06, | |
| "loss": 0.2707, | |
| "step": 4475 | |
| }, | |
| { | |
| "epoch": 2.5851125216387767, | |
| "grad_norm": 0.4742477912373411, | |
| "learning_rate": 6.703661468421065e-06, | |
| "loss": 0.2971, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 2.58799769186382, | |
| "grad_norm": 0.4354163078529058, | |
| "learning_rate": 6.679898255439974e-06, | |
| "loss": 0.2796, | |
| "step": 4485 | |
| }, | |
| { | |
| "epoch": 2.590882862088863, | |
| "grad_norm": 0.42132555668081595, | |
| "learning_rate": 6.656156094789253e-06, | |
| "loss": 0.2844, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 2.5937680323139065, | |
| "grad_norm": 0.43721072545918765, | |
| "learning_rate": 6.632435137014823e-06, | |
| "loss": 0.2778, | |
| "step": 4495 | |
| }, | |
| { | |
| "epoch": 2.59665320253895, | |
| "grad_norm": 0.405910083382628, | |
| "learning_rate": 6.608735532528178e-06, | |
| "loss": 0.2977, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 2.599538372763993, | |
| "grad_norm": 0.41260341077774265, | |
| "learning_rate": 6.585057431605406e-06, | |
| "loss": 0.2686, | |
| "step": 4505 | |
| }, | |
| { | |
| "epoch": 2.6024235429890363, | |
| "grad_norm": 0.44802578482097516, | |
| "learning_rate": 6.561400984386243e-06, | |
| "loss": 0.2809, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 2.6053087132140798, | |
| "grad_norm": 0.4321051913017041, | |
| "learning_rate": 6.537766340873125e-06, | |
| "loss": 0.2842, | |
| "step": 4515 | |
| }, | |
| { | |
| "epoch": 2.6081938834391227, | |
| "grad_norm": 0.4775646961550856, | |
| "learning_rate": 6.514153650930228e-06, | |
| "loss": 0.2702, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 2.611079053664166, | |
| "grad_norm": 0.4330851918875865, | |
| "learning_rate": 6.4905630642825245e-06, | |
| "loss": 0.261, | |
| "step": 4525 | |
| }, | |
| { | |
| "epoch": 2.6139642238892096, | |
| "grad_norm": 0.46362229239200453, | |
| "learning_rate": 6.466994730514842e-06, | |
| "loss": 0.2815, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 2.6168493941142525, | |
| "grad_norm": 0.41520002516181664, | |
| "learning_rate": 6.443448799070895e-06, | |
| "loss": 0.2786, | |
| "step": 4535 | |
| }, | |
| { | |
| "epoch": 2.619734564339296, | |
| "grad_norm": 0.45005348838742215, | |
| "learning_rate": 6.419925419252353e-06, | |
| "loss": 0.2758, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 2.6226197345643394, | |
| "grad_norm": 0.56725726849741, | |
| "learning_rate": 6.39642474021788e-06, | |
| "loss": 0.2641, | |
| "step": 4545 | |
| }, | |
| { | |
| "epoch": 2.6255049047893824, | |
| "grad_norm": 0.49336912135631245, | |
| "learning_rate": 6.3729469109822115e-06, | |
| "loss": 0.2803, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 2.628390075014426, | |
| "grad_norm": 0.4407248177296093, | |
| "learning_rate": 6.349492080415187e-06, | |
| "loss": 0.2849, | |
| "step": 4555 | |
| }, | |
| { | |
| "epoch": 2.631275245239469, | |
| "grad_norm": 0.41465469511645586, | |
| "learning_rate": 6.326060397240814e-06, | |
| "loss": 0.275, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 2.634160415464512, | |
| "grad_norm": 0.4463897975975723, | |
| "learning_rate": 6.3026520100363276e-06, | |
| "loss": 0.2763, | |
| "step": 4565 | |
| }, | |
| { | |
| "epoch": 2.6370455856895556, | |
| "grad_norm": 0.4179248454832217, | |
| "learning_rate": 6.279267067231242e-06, | |
| "loss": 0.2759, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 2.639930755914599, | |
| "grad_norm": 0.4482183921648813, | |
| "learning_rate": 6.255905717106421e-06, | |
| "loss": 0.2809, | |
| "step": 4575 | |
| }, | |
| { | |
| "epoch": 2.642815926139642, | |
| "grad_norm": 0.45808207591686173, | |
| "learning_rate": 6.232568107793127e-06, | |
| "loss": 0.2891, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 2.6457010963646854, | |
| "grad_norm": 0.43079637132281384, | |
| "learning_rate": 6.209254387272085e-06, | |
| "loss": 0.2846, | |
| "step": 4585 | |
| }, | |
| { | |
| "epoch": 2.648586266589729, | |
| "grad_norm": 0.43896911522538273, | |
| "learning_rate": 6.185964703372538e-06, | |
| "loss": 0.2709, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 2.651471436814772, | |
| "grad_norm": 0.41937040243151874, | |
| "learning_rate": 6.162699203771324e-06, | |
| "loss": 0.2698, | |
| "step": 4595 | |
| }, | |
| { | |
| "epoch": 2.6543566070398152, | |
| "grad_norm": 0.436960743544877, | |
| "learning_rate": 6.139458035991932e-06, | |
| "loss": 0.2571, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 2.6572417772648587, | |
| "grad_norm": 0.43802501221211554, | |
| "learning_rate": 6.116241347403564e-06, | |
| "loss": 0.2845, | |
| "step": 4605 | |
| }, | |
| { | |
| "epoch": 2.6601269474899016, | |
| "grad_norm": 0.4586510827028596, | |
| "learning_rate": 6.093049285220198e-06, | |
| "loss": 0.2715, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 2.663012117714945, | |
| "grad_norm": 0.4206625935899696, | |
| "learning_rate": 6.069881996499664e-06, | |
| "loss": 0.2725, | |
| "step": 4615 | |
| }, | |
| { | |
| "epoch": 2.6658972879399885, | |
| "grad_norm": 0.4183330676839759, | |
| "learning_rate": 6.046739628142705e-06, | |
| "loss": 0.2732, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 2.668782458165032, | |
| "grad_norm": 0.4380908491163701, | |
| "learning_rate": 6.023622326892051e-06, | |
| "loss": 0.2763, | |
| "step": 4625 | |
| }, | |
| { | |
| "epoch": 2.671667628390075, | |
| "grad_norm": 0.44598178634221547, | |
| "learning_rate": 6.000530239331481e-06, | |
| "loss": 0.2794, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 2.6745527986151183, | |
| "grad_norm": 0.43851972482947554, | |
| "learning_rate": 5.977463511884898e-06, | |
| "loss": 0.2679, | |
| "step": 4635 | |
| }, | |
| { | |
| "epoch": 2.6774379688401617, | |
| "grad_norm": 0.4173251500861273, | |
| "learning_rate": 5.954422290815396e-06, | |
| "loss": 0.2757, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 2.6803231390652047, | |
| "grad_norm": 0.42797667265485484, | |
| "learning_rate": 5.931406722224344e-06, | |
| "loss": 0.2673, | |
| "step": 4645 | |
| }, | |
| { | |
| "epoch": 2.683208309290248, | |
| "grad_norm": 0.41916940809600545, | |
| "learning_rate": 5.908416952050453e-06, | |
| "loss": 0.2587, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 2.6860934795152915, | |
| "grad_norm": 0.40342454442056763, | |
| "learning_rate": 5.885453126068836e-06, | |
| "loss": 0.2575, | |
| "step": 4655 | |
| }, | |
| { | |
| "epoch": 2.6889786497403345, | |
| "grad_norm": 0.42304042640556994, | |
| "learning_rate": 5.862515389890111e-06, | |
| "loss": 0.2743, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 2.691863819965378, | |
| "grad_norm": 0.4459071254483481, | |
| "learning_rate": 5.839603888959455e-06, | |
| "loss": 0.2764, | |
| "step": 4665 | |
| }, | |
| { | |
| "epoch": 2.6947489901904214, | |
| "grad_norm": 0.41434533982836086, | |
| "learning_rate": 5.8167187685557045e-06, | |
| "loss": 0.2631, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 2.6976341604154648, | |
| "grad_norm": 0.44266182668844456, | |
| "learning_rate": 5.7938601737904e-06, | |
| "loss": 0.2709, | |
| "step": 4675 | |
| }, | |
| { | |
| "epoch": 2.7005193306405078, | |
| "grad_norm": 0.4225234227671778, | |
| "learning_rate": 5.771028249606906e-06, | |
| "loss": 0.2736, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 2.703404500865551, | |
| "grad_norm": 0.43317051916217664, | |
| "learning_rate": 5.74822314077946e-06, | |
| "loss": 0.2812, | |
| "step": 4685 | |
| }, | |
| { | |
| "epoch": 2.7062896710905946, | |
| "grad_norm": 0.4224513310713421, | |
| "learning_rate": 5.725444991912266e-06, | |
| "loss": 0.2657, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 2.7091748413156376, | |
| "grad_norm": 0.4314258773445562, | |
| "learning_rate": 5.702693947438589e-06, | |
| "loss": 0.271, | |
| "step": 4695 | |
| }, | |
| { | |
| "epoch": 2.712060011540681, | |
| "grad_norm": 0.4422321485260852, | |
| "learning_rate": 5.679970151619823e-06, | |
| "loss": 0.2711, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 2.7149451817657244, | |
| "grad_norm": 0.44035975780616426, | |
| "learning_rate": 5.6572737485445804e-06, | |
| "loss": 0.2874, | |
| "step": 4705 | |
| }, | |
| { | |
| "epoch": 2.7178303519907674, | |
| "grad_norm": 0.44924301832440483, | |
| "learning_rate": 5.634604882127771e-06, | |
| "loss": 0.2703, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 2.720715522215811, | |
| "grad_norm": 0.4107307992041668, | |
| "learning_rate": 5.611963696109726e-06, | |
| "loss": 0.2729, | |
| "step": 4715 | |
| }, | |
| { | |
| "epoch": 2.7236006924408542, | |
| "grad_norm": 0.4372366602675938, | |
| "learning_rate": 5.58935033405523e-06, | |
| "loss": 0.276, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 2.726485862665897, | |
| "grad_norm": 0.45334235025453323, | |
| "learning_rate": 5.5667649393526535e-06, | |
| "loss": 0.2801, | |
| "step": 4725 | |
| }, | |
| { | |
| "epoch": 2.7293710328909406, | |
| "grad_norm": 0.4201853295737695, | |
| "learning_rate": 5.544207655213021e-06, | |
| "loss": 0.2831, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 2.732256203115984, | |
| "grad_norm": 0.44159486930994, | |
| "learning_rate": 5.521678624669121e-06, | |
| "loss": 0.2644, | |
| "step": 4735 | |
| }, | |
| { | |
| "epoch": 2.735141373341027, | |
| "grad_norm": 0.45477978259519153, | |
| "learning_rate": 5.499177990574591e-06, | |
| "loss": 0.2856, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 2.7380265435660704, | |
| "grad_norm": 0.4329706475468657, | |
| "learning_rate": 5.4767058956029955e-06, | |
| "loss": 0.2757, | |
| "step": 4745 | |
| }, | |
| { | |
| "epoch": 2.740911713791114, | |
| "grad_norm": 0.46147021358014734, | |
| "learning_rate": 5.454262482246954e-06, | |
| "loss": 0.2816, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 2.743796884016157, | |
| "grad_norm": 0.4661777717367274, | |
| "learning_rate": 5.431847892817208e-06, | |
| "loss": 0.2715, | |
| "step": 4755 | |
| }, | |
| { | |
| "epoch": 2.7466820542412003, | |
| "grad_norm": 0.42606918880864486, | |
| "learning_rate": 5.40946226944173e-06, | |
| "loss": 0.2785, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 2.7495672244662437, | |
| "grad_norm": 0.44527079904346323, | |
| "learning_rate": 5.387105754064829e-06, | |
| "loss": 0.2776, | |
| "step": 4765 | |
| }, | |
| { | |
| "epoch": 2.7524523946912867, | |
| "grad_norm": 0.4377246196123211, | |
| "learning_rate": 5.364778488446247e-06, | |
| "loss": 0.2753, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 2.75533756491633, | |
| "grad_norm": 0.44386684339819055, | |
| "learning_rate": 5.342480614160247e-06, | |
| "loss": 0.2859, | |
| "step": 4775 | |
| }, | |
| { | |
| "epoch": 2.7582227351413735, | |
| "grad_norm": 0.4235638405047958, | |
| "learning_rate": 5.320212272594725e-06, | |
| "loss": 0.2811, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 2.7611079053664165, | |
| "grad_norm": 0.45140518863698265, | |
| "learning_rate": 5.297973604950321e-06, | |
| "loss": 0.2704, | |
| "step": 4785 | |
| }, | |
| { | |
| "epoch": 2.76399307559146, | |
| "grad_norm": 0.4079388286223738, | |
| "learning_rate": 5.275764752239523e-06, | |
| "loss": 0.2625, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 2.7668782458165033, | |
| "grad_norm": 0.41869006995290337, | |
| "learning_rate": 5.2535858552857474e-06, | |
| "loss": 0.2603, | |
| "step": 4795 | |
| }, | |
| { | |
| "epoch": 2.7697634160415463, | |
| "grad_norm": 0.4333081525779454, | |
| "learning_rate": 5.231437054722477e-06, | |
| "loss": 0.2908, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 2.7726485862665897, | |
| "grad_norm": 0.46442418499425026, | |
| "learning_rate": 5.209318490992355e-06, | |
| "loss": 0.2711, | |
| "step": 4805 | |
| }, | |
| { | |
| "epoch": 2.775533756491633, | |
| "grad_norm": 0.42235720398392995, | |
| "learning_rate": 5.187230304346304e-06, | |
| "loss": 0.2874, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 2.778418926716676, | |
| "grad_norm": 0.42546995948748306, | |
| "learning_rate": 5.1651726348426146e-06, | |
| "loss": 0.2845, | |
| "step": 4815 | |
| }, | |
| { | |
| "epoch": 2.7813040969417195, | |
| "grad_norm": 0.4458920282497016, | |
| "learning_rate": 5.143145622346089e-06, | |
| "loss": 0.2656, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 2.784189267166763, | |
| "grad_norm": 0.4140358163182573, | |
| "learning_rate": 5.121149406527123e-06, | |
| "loss": 0.2674, | |
| "step": 4825 | |
| }, | |
| { | |
| "epoch": 2.787074437391806, | |
| "grad_norm": 0.4300310193339041, | |
| "learning_rate": 5.099184126860837e-06, | |
| "loss": 0.266, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 2.7899596076168494, | |
| "grad_norm": 0.8717864481697682, | |
| "learning_rate": 5.077249922626197e-06, | |
| "loss": 0.2794, | |
| "step": 4835 | |
| }, | |
| { | |
| "epoch": 2.792844777841893, | |
| "grad_norm": 0.43179815522266113, | |
| "learning_rate": 5.055346932905125e-06, | |
| "loss": 0.2821, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 2.7957299480669358, | |
| "grad_norm": 0.438127032880629, | |
| "learning_rate": 5.033475296581606e-06, | |
| "loss": 0.2749, | |
| "step": 4845 | |
| }, | |
| { | |
| "epoch": 2.798615118291979, | |
| "grad_norm": 0.42678461004761115, | |
| "learning_rate": 5.011635152340816e-06, | |
| "loss": 0.2724, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 2.8015002885170226, | |
| "grad_norm": 0.4403224341518811, | |
| "learning_rate": 4.9898266386682534e-06, | |
| "loss": 0.2848, | |
| "step": 4855 | |
| }, | |
| { | |
| "epoch": 2.8043854587420656, | |
| "grad_norm": 0.4551147345010523, | |
| "learning_rate": 4.968049893848854e-06, | |
| "loss": 0.2637, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 2.807270628967109, | |
| "grad_norm": 0.43423817098138445, | |
| "learning_rate": 4.9463050559660985e-06, | |
| "loss": 0.268, | |
| "step": 4865 | |
| }, | |
| { | |
| "epoch": 2.8101557991921524, | |
| "grad_norm": 0.42455505455856984, | |
| "learning_rate": 4.924592262901154e-06, | |
| "loss": 0.2681, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 2.8130409694171954, | |
| "grad_norm": 0.4530758791142782, | |
| "learning_rate": 4.9029116523320035e-06, | |
| "loss": 0.268, | |
| "step": 4875 | |
| }, | |
| { | |
| "epoch": 2.815926139642239, | |
| "grad_norm": 0.4249843125062271, | |
| "learning_rate": 4.881263361732552e-06, | |
| "loss": 0.2753, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 2.8188113098672822, | |
| "grad_norm": 0.4472656624417374, | |
| "learning_rate": 4.859647528371784e-06, | |
| "loss": 0.283, | |
| "step": 4885 | |
| }, | |
| { | |
| "epoch": 2.821696480092325, | |
| "grad_norm": 0.42578297771011603, | |
| "learning_rate": 4.838064289312862e-06, | |
| "loss": 0.2898, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 2.8245816503173686, | |
| "grad_norm": 0.4436993117564527, | |
| "learning_rate": 4.816513781412282e-06, | |
| "loss": 0.2701, | |
| "step": 4895 | |
| }, | |
| { | |
| "epoch": 2.827466820542412, | |
| "grad_norm": 0.4621541772278322, | |
| "learning_rate": 4.7949961413189885e-06, | |
| "loss": 0.2771, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 2.830351990767455, | |
| "grad_norm": 0.4266188005044788, | |
| "learning_rate": 4.77351150547352e-06, | |
| "loss": 0.2803, | |
| "step": 4905 | |
| }, | |
| { | |
| "epoch": 2.8332371609924984, | |
| "grad_norm": 0.45663638676525325, | |
| "learning_rate": 4.752060010107145e-06, | |
| "loss": 0.2826, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 2.836122331217542, | |
| "grad_norm": 0.42586386429655265, | |
| "learning_rate": 4.730641791240981e-06, | |
| "loss": 0.2633, | |
| "step": 4915 | |
| }, | |
| { | |
| "epoch": 2.839007501442585, | |
| "grad_norm": 0.42216429186560445, | |
| "learning_rate": 4.709256984685147e-06, | |
| "loss": 0.2835, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 2.8418926716676283, | |
| "grad_norm": 0.45151376072343974, | |
| "learning_rate": 4.687905726037902e-06, | |
| "loss": 0.2716, | |
| "step": 4925 | |
| }, | |
| { | |
| "epoch": 2.8447778418926717, | |
| "grad_norm": 0.40227119407723355, | |
| "learning_rate": 4.666588150684786e-06, | |
| "loss": 0.2744, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 2.847663012117715, | |
| "grad_norm": 0.41851949935896066, | |
| "learning_rate": 4.6453043937977496e-06, | |
| "loss": 0.2771, | |
| "step": 4935 | |
| }, | |
| { | |
| "epoch": 2.850548182342758, | |
| "grad_norm": 0.40933746114854325, | |
| "learning_rate": 4.624054590334304e-06, | |
| "loss": 0.2689, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 2.8534333525678015, | |
| "grad_norm": 0.47564242491449826, | |
| "learning_rate": 4.6028388750366756e-06, | |
| "loss": 0.2809, | |
| "step": 4945 | |
| }, | |
| { | |
| "epoch": 2.856318522792845, | |
| "grad_norm": 0.43828857688753076, | |
| "learning_rate": 4.581657382430932e-06, | |
| "loss": 0.2853, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 2.859203693017888, | |
| "grad_norm": 0.43574357681542614, | |
| "learning_rate": 4.56051024682615e-06, | |
| "loss": 0.2756, | |
| "step": 4955 | |
| }, | |
| { | |
| "epoch": 2.8620888632429313, | |
| "grad_norm": 0.42537161903700743, | |
| "learning_rate": 4.53939760231354e-06, | |
| "loss": 0.2874, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 2.8649740334679747, | |
| "grad_norm": 0.4597603651479367, | |
| "learning_rate": 4.518319582765623e-06, | |
| "loss": 0.2892, | |
| "step": 4965 | |
| }, | |
| { | |
| "epoch": 2.8678592036930177, | |
| "grad_norm": 0.4446590811574243, | |
| "learning_rate": 4.497276321835357e-06, | |
| "loss": 0.2712, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 2.870744373918061, | |
| "grad_norm": 0.46299522122029557, | |
| "learning_rate": 4.476267952955295e-06, | |
| "loss": 0.3009, | |
| "step": 4975 | |
| }, | |
| { | |
| "epoch": 2.8736295441431046, | |
| "grad_norm": 0.4385843033290836, | |
| "learning_rate": 4.455294609336771e-06, | |
| "loss": 0.2755, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 2.876514714368148, | |
| "grad_norm": 0.44649836304692625, | |
| "learning_rate": 4.434356423968999e-06, | |
| "loss": 0.2837, | |
| "step": 4985 | |
| }, | |
| { | |
| "epoch": 2.879399884593191, | |
| "grad_norm": 0.4656079138932178, | |
| "learning_rate": 4.413453529618274e-06, | |
| "loss": 0.2896, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 2.8822850548182344, | |
| "grad_norm": 0.40090614598820945, | |
| "learning_rate": 4.3925860588271015e-06, | |
| "loss": 0.3039, | |
| "step": 4995 | |
| }, | |
| { | |
| "epoch": 2.885170225043278, | |
| "grad_norm": 0.429166127365773, | |
| "learning_rate": 4.371754143913398e-06, | |
| "loss": 0.2732, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 2.888055395268321, | |
| "grad_norm": 0.4028565372934939, | |
| "learning_rate": 4.350957916969598e-06, | |
| "loss": 0.2629, | |
| "step": 5005 | |
| }, | |
| { | |
| "epoch": 2.890940565493364, | |
| "grad_norm": 0.4626974059523127, | |
| "learning_rate": 4.330197509861851e-06, | |
| "loss": 0.2748, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 2.8938257357184076, | |
| "grad_norm": 0.4168915007613455, | |
| "learning_rate": 4.3094730542291875e-06, | |
| "loss": 0.2851, | |
| "step": 5015 | |
| }, | |
| { | |
| "epoch": 2.8967109059434506, | |
| "grad_norm": 0.4404850147613879, | |
| "learning_rate": 4.288784681482657e-06, | |
| "loss": 0.2633, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 2.899596076168494, | |
| "grad_norm": 0.4368263858822652, | |
| "learning_rate": 4.268132522804532e-06, | |
| "loss": 0.2795, | |
| "step": 5025 | |
| }, | |
| { | |
| "epoch": 2.9024812463935374, | |
| "grad_norm": 0.4518971041077726, | |
| "learning_rate": 4.247516709147437e-06, | |
| "loss": 0.2779, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 2.9053664166185804, | |
| "grad_norm": 0.42168757805620943, | |
| "learning_rate": 4.2269373712335535e-06, | |
| "loss": 0.2781, | |
| "step": 5035 | |
| }, | |
| { | |
| "epoch": 2.908251586843624, | |
| "grad_norm": 0.44301611235467, | |
| "learning_rate": 4.206394639553766e-06, | |
| "loss": 0.2844, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 2.9111367570686673, | |
| "grad_norm": 0.43608207255196907, | |
| "learning_rate": 4.185888644366841e-06, | |
| "loss": 0.2764, | |
| "step": 5045 | |
| }, | |
| { | |
| "epoch": 2.9140219272937102, | |
| "grad_norm": 0.43614919217568904, | |
| "learning_rate": 4.165419515698613e-06, | |
| "loss": 0.2686, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 2.9169070975187537, | |
| "grad_norm": 0.45163629978863207, | |
| "learning_rate": 4.14498738334115e-06, | |
| "loss": 0.2685, | |
| "step": 5055 | |
| }, | |
| { | |
| "epoch": 2.919792267743797, | |
| "grad_norm": 0.42566284160748424, | |
| "learning_rate": 4.1245923768519235e-06, | |
| "loss": 0.2624, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 2.92267743796884, | |
| "grad_norm": 0.42580290639250135, | |
| "learning_rate": 4.104234625552991e-06, | |
| "loss": 0.2656, | |
| "step": 5065 | |
| }, | |
| { | |
| "epoch": 2.9255626081938835, | |
| "grad_norm": 0.4508732247385924, | |
| "learning_rate": 4.083914258530202e-06, | |
| "loss": 0.2829, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 2.928447778418927, | |
| "grad_norm": 0.4230373908678959, | |
| "learning_rate": 4.063631404632336e-06, | |
| "loss": 0.283, | |
| "step": 5075 | |
| }, | |
| { | |
| "epoch": 2.93133294864397, | |
| "grad_norm": 0.4519630975638044, | |
| "learning_rate": 4.043386192470309e-06, | |
| "loss": 0.299, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 2.9342181188690133, | |
| "grad_norm": 0.46105467021109736, | |
| "learning_rate": 4.0231787504163684e-06, | |
| "loss": 0.2878, | |
| "step": 5085 | |
| }, | |
| { | |
| "epoch": 2.9371032890940567, | |
| "grad_norm": 0.4511914211925812, | |
| "learning_rate": 4.00300920660325e-06, | |
| "loss": 0.2778, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 2.9399884593190997, | |
| "grad_norm": 0.44702115253367053, | |
| "learning_rate": 3.982877688923396e-06, | |
| "loss": 0.2662, | |
| "step": 5095 | |
| }, | |
| { | |
| "epoch": 2.942873629544143, | |
| "grad_norm": 0.427417788950817, | |
| "learning_rate": 3.962784325028119e-06, | |
| "loss": 0.2713, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 2.9457587997691865, | |
| "grad_norm": 0.43897355082883444, | |
| "learning_rate": 3.942729242326814e-06, | |
| "loss": 0.2743, | |
| "step": 5105 | |
| }, | |
| { | |
| "epoch": 2.9486439699942295, | |
| "grad_norm": 0.4670512082769263, | |
| "learning_rate": 3.9227125679861286e-06, | |
| "loss": 0.284, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 2.951529140219273, | |
| "grad_norm": 0.4664395393065826, | |
| "learning_rate": 3.902734428929172e-06, | |
| "loss": 0.2769, | |
| "step": 5115 | |
| }, | |
| { | |
| "epoch": 2.9544143104443163, | |
| "grad_norm": 0.4670672506794604, | |
| "learning_rate": 3.882794951834711e-06, | |
| "loss": 0.2754, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 2.9572994806693593, | |
| "grad_norm": 0.43499585946337327, | |
| "learning_rate": 3.862894263136361e-06, | |
| "loss": 0.2688, | |
| "step": 5125 | |
| }, | |
| { | |
| "epoch": 2.9601846508944027, | |
| "grad_norm": 0.4471133717151604, | |
| "learning_rate": 3.8430324890217805e-06, | |
| "loss": 0.2685, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 2.963069821119446, | |
| "grad_norm": 0.4382618072040372, | |
| "learning_rate": 3.823209755431873e-06, | |
| "loss": 0.2809, | |
| "step": 5135 | |
| }, | |
| { | |
| "epoch": 2.965954991344489, | |
| "grad_norm": 0.4241285275096586, | |
| "learning_rate": 3.8034261880600034e-06, | |
| "loss": 0.27, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 2.9688401615695326, | |
| "grad_norm": 0.4091644110971949, | |
| "learning_rate": 3.78368191235118e-06, | |
| "loss": 0.2799, | |
| "step": 5145 | |
| }, | |
| { | |
| "epoch": 2.971725331794576, | |
| "grad_norm": 0.43102734170478396, | |
| "learning_rate": 3.76397705350127e-06, | |
| "loss": 0.2797, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 2.974610502019619, | |
| "grad_norm": 0.4370911603217546, | |
| "learning_rate": 3.7443117364561964e-06, | |
| "loss": 0.2701, | |
| "step": 5155 | |
| }, | |
| { | |
| "epoch": 2.9774956722446624, | |
| "grad_norm": 0.42916373087001025, | |
| "learning_rate": 3.7246860859111655e-06, | |
| "loss": 0.2734, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 2.980380842469706, | |
| "grad_norm": 0.4275858080889107, | |
| "learning_rate": 3.705100226309858e-06, | |
| "loss": 0.2677, | |
| "step": 5165 | |
| }, | |
| { | |
| "epoch": 2.983266012694749, | |
| "grad_norm": 0.4310089026075201, | |
| "learning_rate": 3.6855542818436397e-06, | |
| "loss": 0.2726, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 2.986151182919792, | |
| "grad_norm": 0.44889673891857773, | |
| "learning_rate": 3.6660483764507916e-06, | |
| "loss": 0.265, | |
| "step": 5175 | |
| }, | |
| { | |
| "epoch": 2.9890363531448356, | |
| "grad_norm": 0.4597978279604376, | |
| "learning_rate": 3.6465826338157007e-06, | |
| "loss": 0.2857, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 2.9919215233698786, | |
| "grad_norm": 0.4672195167157573, | |
| "learning_rate": 3.6271571773680893e-06, | |
| "loss": 0.2765, | |
| "step": 5185 | |
| }, | |
| { | |
| "epoch": 2.994806693594922, | |
| "grad_norm": 0.4409412220124302, | |
| "learning_rate": 3.6077721302822355e-06, | |
| "loss": 0.2826, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 2.9976918638199654, | |
| "grad_norm": 0.4448322757550867, | |
| "learning_rate": 3.588427615476189e-06, | |
| "loss": 0.2951, | |
| "step": 5195 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 0.4704112410545349, | |
| "eval_runtime": 0.5938, | |
| "eval_samples_per_second": 129.667, | |
| "eval_steps_per_second": 3.368, | |
| "step": 5199 | |
| }, | |
| { | |
| "epoch": 3.000577034045009, | |
| "grad_norm": 0.478834624205184, | |
| "learning_rate": 3.5691237556109794e-06, | |
| "loss": 0.2548, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 3.003462204270052, | |
| "grad_norm": 0.45264093142712225, | |
| "learning_rate": 3.549860673089852e-06, | |
| "loss": 0.2178, | |
| "step": 5205 | |
| }, | |
| { | |
| "epoch": 3.0063473744950953, | |
| "grad_norm": 0.42695857914733454, | |
| "learning_rate": 3.5306384900574943e-06, | |
| "loss": 0.2145, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 3.0092325447201387, | |
| "grad_norm": 0.4252664702307335, | |
| "learning_rate": 3.5114573283992536e-06, | |
| "loss": 0.2017, | |
| "step": 5215 | |
| }, | |
| { | |
| "epoch": 3.0121177149451817, | |
| "grad_norm": 0.4362013593591915, | |
| "learning_rate": 3.492317309740362e-06, | |
| "loss": 0.202, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 3.015002885170225, | |
| "grad_norm": 0.7918296216376329, | |
| "learning_rate": 3.473218555445166e-06, | |
| "loss": 0.2023, | |
| "step": 5225 | |
| }, | |
| { | |
| "epoch": 3.0178880553952685, | |
| "grad_norm": 0.4425689050623567, | |
| "learning_rate": 3.4541611866163748e-06, | |
| "loss": 0.2156, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 3.0207732256203115, | |
| "grad_norm": 0.434243728200696, | |
| "learning_rate": 3.435145324094259e-06, | |
| "loss": 0.1967, | |
| "step": 5235 | |
| }, | |
| { | |
| "epoch": 3.023658395845355, | |
| "grad_norm": 0.48550762220985233, | |
| "learning_rate": 3.4161710884559186e-06, | |
| "loss": 0.228, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 3.0265435660703983, | |
| "grad_norm": 0.4451374888019577, | |
| "learning_rate": 3.3972386000144975e-06, | |
| "loss": 0.1994, | |
| "step": 5245 | |
| }, | |
| { | |
| "epoch": 3.0294287362954413, | |
| "grad_norm": 0.4479982505677569, | |
| "learning_rate": 3.378347978818425e-06, | |
| "loss": 0.2064, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 3.0323139065204847, | |
| "grad_norm": 0.48420281819894295, | |
| "learning_rate": 3.359499344650651e-06, | |
| "loss": 0.2104, | |
| "step": 5255 | |
| }, | |
| { | |
| "epoch": 3.035199076745528, | |
| "grad_norm": 0.4674463960460507, | |
| "learning_rate": 3.3406928170278997e-06, | |
| "loss": 0.2107, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 3.038084246970571, | |
| "grad_norm": 0.42110517886246845, | |
| "learning_rate": 3.3219285151999037e-06, | |
| "loss": 0.1936, | |
| "step": 5265 | |
| }, | |
| { | |
| "epoch": 3.0409694171956145, | |
| "grad_norm": 0.4325902687189473, | |
| "learning_rate": 3.3032065581486393e-06, | |
| "loss": 0.2107, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 3.043854587420658, | |
| "grad_norm": 0.45814670201275987, | |
| "learning_rate": 3.2845270645875838e-06, | |
| "loss": 0.2073, | |
| "step": 5275 | |
| }, | |
| { | |
| "epoch": 3.046739757645701, | |
| "grad_norm": 0.4673351294536787, | |
| "learning_rate": 3.26589015296096e-06, | |
| "loss": 0.2141, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 3.0496249278707444, | |
| "grad_norm": 0.43997040330941384, | |
| "learning_rate": 3.2472959414429883e-06, | |
| "loss": 0.2108, | |
| "step": 5285 | |
| }, | |
| { | |
| "epoch": 3.0525100980957878, | |
| "grad_norm": 0.4862585478605358, | |
| "learning_rate": 3.228744547937125e-06, | |
| "loss": 0.2089, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 3.0553952683208307, | |
| "grad_norm": 0.4501337212448104, | |
| "learning_rate": 3.2102360900753237e-06, | |
| "loss": 0.216, | |
| "step": 5295 | |
| }, | |
| { | |
| "epoch": 3.058280438545874, | |
| "grad_norm": 0.43785695040034356, | |
| "learning_rate": 3.1917706852173003e-06, | |
| "loss": 0.1999, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 3.0611656087709176, | |
| "grad_norm": 0.45799695752548214, | |
| "learning_rate": 3.173348450449759e-06, | |
| "loss": 0.2034, | |
| "step": 5305 | |
| }, | |
| { | |
| "epoch": 3.0640507789959606, | |
| "grad_norm": 0.4896057098908631, | |
| "learning_rate": 3.1549695025856873e-06, | |
| "loss": 0.2138, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 3.066935949221004, | |
| "grad_norm": 0.43978051760808934, | |
| "learning_rate": 3.1366339581635785e-06, | |
| "loss": 0.211, | |
| "step": 5315 | |
| }, | |
| { | |
| "epoch": 3.0698211194460474, | |
| "grad_norm": 0.5408651817364915, | |
| "learning_rate": 3.118341933446726e-06, | |
| "loss": 0.2189, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 3.0727062896710904, | |
| "grad_norm": 0.47167963172935967, | |
| "learning_rate": 3.100093544422459e-06, | |
| "loss": 0.2087, | |
| "step": 5325 | |
| }, | |
| { | |
| "epoch": 3.075591459896134, | |
| "grad_norm": 0.45222235059141447, | |
| "learning_rate": 3.0818889068014167e-06, | |
| "loss": 0.2086, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 3.0784766301211772, | |
| "grad_norm": 0.46762858344658953, | |
| "learning_rate": 3.0637281360168348e-06, | |
| "loss": 0.2058, | |
| "step": 5335 | |
| }, | |
| { | |
| "epoch": 3.0813618003462206, | |
| "grad_norm": 0.4982055068360655, | |
| "learning_rate": 3.045611347223776e-06, | |
| "loss": 0.2168, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 3.0842469705712636, | |
| "grad_norm": 0.44825046777882716, | |
| "learning_rate": 3.0275386552984232e-06, | |
| "loss": 0.2089, | |
| "step": 5345 | |
| }, | |
| { | |
| "epoch": 3.087132140796307, | |
| "grad_norm": 0.4780597142010933, | |
| "learning_rate": 3.00951017483735e-06, | |
| "loss": 0.2, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 3.0900173110213505, | |
| "grad_norm": 0.4749892463930495, | |
| "learning_rate": 2.991526020156795e-06, | |
| "loss": 0.2082, | |
| "step": 5355 | |
| }, | |
| { | |
| "epoch": 3.0929024812463934, | |
| "grad_norm": 0.5316835135478458, | |
| "learning_rate": 2.9735863052919256e-06, | |
| "loss": 0.2117, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 3.095787651471437, | |
| "grad_norm": 0.48592697259547013, | |
| "learning_rate": 2.955691143996119e-06, | |
| "loss": 0.2189, | |
| "step": 5365 | |
| }, | |
| { | |
| "epoch": 3.0986728216964803, | |
| "grad_norm": 0.478325002856811, | |
| "learning_rate": 2.9378406497402577e-06, | |
| "loss": 0.2217, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 3.1015579919215233, | |
| "grad_norm": 0.47784013307657136, | |
| "learning_rate": 2.9200349357119805e-06, | |
| "loss": 0.2047, | |
| "step": 5375 | |
| }, | |
| { | |
| "epoch": 3.1044431621465667, | |
| "grad_norm": 0.4366920553845523, | |
| "learning_rate": 2.902274114814995e-06, | |
| "loss": 0.2052, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 3.10732833237161, | |
| "grad_norm": 0.48726638763350977, | |
| "learning_rate": 2.884558299668333e-06, | |
| "loss": 0.2135, | |
| "step": 5385 | |
| }, | |
| { | |
| "epoch": 3.110213502596653, | |
| "grad_norm": 0.4352029364038837, | |
| "learning_rate": 2.866887602605667e-06, | |
| "loss": 0.2146, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 3.1130986728216965, | |
| "grad_norm": 0.4618849547883958, | |
| "learning_rate": 2.8492621356745677e-06, | |
| "loss": 0.2007, | |
| "step": 5395 | |
| }, | |
| { | |
| "epoch": 3.11598384304674, | |
| "grad_norm": 0.5680788070054426, | |
| "learning_rate": 2.831682010635811e-06, | |
| "loss": 0.2254, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 3.118869013271783, | |
| "grad_norm": 0.4478067921317212, | |
| "learning_rate": 2.8141473389626706e-06, | |
| "loss": 0.2025, | |
| "step": 5405 | |
| }, | |
| { | |
| "epoch": 3.1217541834968263, | |
| "grad_norm": 0.4383072651557812, | |
| "learning_rate": 2.7966582318402046e-06, | |
| "loss": 0.2038, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 3.1246393537218697, | |
| "grad_norm": 0.4469251744492116, | |
| "learning_rate": 2.7792148001645505e-06, | |
| "loss": 0.2034, | |
| "step": 5415 | |
| }, | |
| { | |
| "epoch": 3.1275245239469127, | |
| "grad_norm": 0.45796076894000676, | |
| "learning_rate": 2.7618171545422156e-06, | |
| "loss": 0.2237, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 3.130409694171956, | |
| "grad_norm": 0.4732594923700523, | |
| "learning_rate": 2.744465405289406e-06, | |
| "loss": 0.2115, | |
| "step": 5425 | |
| }, | |
| { | |
| "epoch": 3.1332948643969996, | |
| "grad_norm": 0.47883717903500816, | |
| "learning_rate": 2.7271596624312845e-06, | |
| "loss": 0.2176, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 3.1361800346220425, | |
| "grad_norm": 0.47490798721155425, | |
| "learning_rate": 2.709900035701297e-06, | |
| "loss": 0.1962, | |
| "step": 5435 | |
| }, | |
| { | |
| "epoch": 3.139065204847086, | |
| "grad_norm": 0.47556110953605957, | |
| "learning_rate": 2.6926866345404846e-06, | |
| "loss": 0.2122, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 3.1419503750721294, | |
| "grad_norm": 0.4674071776337381, | |
| "learning_rate": 2.6755195680967607e-06, | |
| "loss": 0.1992, | |
| "step": 5445 | |
| }, | |
| { | |
| "epoch": 3.1448355452971724, | |
| "grad_norm": 0.474411421746897, | |
| "learning_rate": 2.658398945224253e-06, | |
| "loss": 0.1924, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 3.1477207155222158, | |
| "grad_norm": 0.47606141844860833, | |
| "learning_rate": 2.6413248744825837e-06, | |
| "loss": 0.2076, | |
| "step": 5455 | |
| }, | |
| { | |
| "epoch": 3.150605885747259, | |
| "grad_norm": 0.45905433780073146, | |
| "learning_rate": 2.624297464136204e-06, | |
| "loss": 0.2142, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 3.153491055972302, | |
| "grad_norm": 0.536638485210775, | |
| "learning_rate": 2.6073168221536906e-06, | |
| "loss": 0.219, | |
| "step": 5465 | |
| }, | |
| { | |
| "epoch": 3.1563762261973456, | |
| "grad_norm": 0.45678787981866464, | |
| "learning_rate": 2.5903830562070675e-06, | |
| "loss": 0.2036, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 3.159261396422389, | |
| "grad_norm": 0.4496673360192395, | |
| "learning_rate": 2.57349627367113e-06, | |
| "loss": 0.2092, | |
| "step": 5475 | |
| }, | |
| { | |
| "epoch": 3.162146566647432, | |
| "grad_norm": 0.4612630830131145, | |
| "learning_rate": 2.5566565816227585e-06, | |
| "loss": 0.204, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 3.1650317368724754, | |
| "grad_norm": 0.45145251818392435, | |
| "learning_rate": 2.5398640868402304e-06, | |
| "loss": 0.2056, | |
| "step": 5485 | |
| }, | |
| { | |
| "epoch": 3.167916907097519, | |
| "grad_norm": 0.49129615711916425, | |
| "learning_rate": 2.523118895802553e-06, | |
| "loss": 0.2069, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 3.170802077322562, | |
| "grad_norm": 0.5139622757706684, | |
| "learning_rate": 2.506421114688794e-06, | |
| "loss": 0.221, | |
| "step": 5495 | |
| }, | |
| { | |
| "epoch": 3.1736872475476052, | |
| "grad_norm": 0.48052040604024665, | |
| "learning_rate": 2.4897708493773997e-06, | |
| "loss": 0.1973, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 3.1765724177726486, | |
| "grad_norm": 0.49706524838406957, | |
| "learning_rate": 2.4731682054455174e-06, | |
| "loss": 0.2053, | |
| "step": 5505 | |
| }, | |
| { | |
| "epoch": 3.179457587997692, | |
| "grad_norm": 0.5133679294908896, | |
| "learning_rate": 2.456613288168337e-06, | |
| "loss": 0.2035, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 3.182342758222735, | |
| "grad_norm": 0.483475787236041, | |
| "learning_rate": 2.440106202518423e-06, | |
| "loss": 0.2082, | |
| "step": 5515 | |
| }, | |
| { | |
| "epoch": 3.1852279284477785, | |
| "grad_norm": 0.4815903117086154, | |
| "learning_rate": 2.423647053165046e-06, | |
| "loss": 0.2046, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 3.188113098672822, | |
| "grad_norm": 0.4753159403205944, | |
| "learning_rate": 2.4072359444735117e-06, | |
| "loss": 0.2116, | |
| "step": 5525 | |
| }, | |
| { | |
| "epoch": 3.190998268897865, | |
| "grad_norm": 0.5207748356365552, | |
| "learning_rate": 2.390872980504516e-06, | |
| "loss": 0.2078, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 3.1938834391229083, | |
| "grad_norm": 0.500796915089432, | |
| "learning_rate": 2.374558265013469e-06, | |
| "loss": 0.2013, | |
| "step": 5535 | |
| }, | |
| { | |
| "epoch": 3.1967686093479517, | |
| "grad_norm": 0.5192772858747745, | |
| "learning_rate": 2.358291901449842e-06, | |
| "loss": 0.2003, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 3.1996537795729947, | |
| "grad_norm": 0.4809730211239497, | |
| "learning_rate": 2.342073992956517e-06, | |
| "loss": 0.2098, | |
| "step": 5545 | |
| }, | |
| { | |
| "epoch": 3.202538949798038, | |
| "grad_norm": 0.4627046426816422, | |
| "learning_rate": 2.3259046423691333e-06, | |
| "loss": 0.2032, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 3.2054241200230815, | |
| "grad_norm": 0.5339618244022528, | |
| "learning_rate": 2.309783952215421e-06, | |
| "loss": 0.2075, | |
| "step": 5555 | |
| }, | |
| { | |
| "epoch": 3.2083092902481245, | |
| "grad_norm": 0.5160587961491079, | |
| "learning_rate": 2.293712024714564e-06, | |
| "loss": 0.2147, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 3.211194460473168, | |
| "grad_norm": 0.5541552433226901, | |
| "learning_rate": 2.2776889617765495e-06, | |
| "loss": 0.2367, | |
| "step": 5565 | |
| }, | |
| { | |
| "epoch": 3.2140796306982113, | |
| "grad_norm": 0.45574118435347694, | |
| "learning_rate": 2.2617148650015275e-06, | |
| "loss": 0.2104, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 3.2169648009232543, | |
| "grad_norm": 0.4714574839149813, | |
| "learning_rate": 2.2457898356791496e-06, | |
| "loss": 0.2154, | |
| "step": 5575 | |
| }, | |
| { | |
| "epoch": 3.2198499711482977, | |
| "grad_norm": 0.5065703016927723, | |
| "learning_rate": 2.2299139747879385e-06, | |
| "loss": 0.2045, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 3.222735141373341, | |
| "grad_norm": 0.46572520384835114, | |
| "learning_rate": 2.214087382994653e-06, | |
| "loss": 0.2368, | |
| "step": 5585 | |
| }, | |
| { | |
| "epoch": 3.225620311598384, | |
| "grad_norm": 0.4780383032384353, | |
| "learning_rate": 2.198310160653636e-06, | |
| "loss": 0.2041, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 3.2285054818234276, | |
| "grad_norm": 0.48268579681924456, | |
| "learning_rate": 2.1825824078061898e-06, | |
| "loss": 0.2057, | |
| "step": 5595 | |
| }, | |
| { | |
| "epoch": 3.231390652048471, | |
| "grad_norm": 0.48646053407264, | |
| "learning_rate": 2.1669042241799378e-06, | |
| "loss": 0.199, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 3.234275822273514, | |
| "grad_norm": 0.49576817208299834, | |
| "learning_rate": 2.1512757091881874e-06, | |
| "loss": 0.1971, | |
| "step": 5605 | |
| }, | |
| { | |
| "epoch": 3.2371609924985574, | |
| "grad_norm": 0.4841025834202475, | |
| "learning_rate": 2.1356969619293023e-06, | |
| "loss": 0.2227, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 3.240046162723601, | |
| "grad_norm": 0.5512557293571464, | |
| "learning_rate": 2.1201680811860815e-06, | |
| "loss": 0.1993, | |
| "step": 5615 | |
| }, | |
| { | |
| "epoch": 3.2429313329486438, | |
| "grad_norm": 0.49443124657159826, | |
| "learning_rate": 2.1046891654251266e-06, | |
| "loss": 0.207, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 3.245816503173687, | |
| "grad_norm": 0.502803434696454, | |
| "learning_rate": 2.089260312796213e-06, | |
| "loss": 0.2037, | |
| "step": 5625 | |
| }, | |
| { | |
| "epoch": 3.2487016733987306, | |
| "grad_norm": 0.4761464874997464, | |
| "learning_rate": 2.0738816211316716e-06, | |
| "loss": 0.1969, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 3.251586843623774, | |
| "grad_norm": 0.5250059943014331, | |
| "learning_rate": 2.0585531879457743e-06, | |
| "loss": 0.1998, | |
| "step": 5635 | |
| }, | |
| { | |
| "epoch": 3.254472013848817, | |
| "grad_norm": 0.4534253902795417, | |
| "learning_rate": 2.043275110434113e-06, | |
| "loss": 0.2057, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 3.2573571840738604, | |
| "grad_norm": 0.456942792070371, | |
| "learning_rate": 2.0280474854729727e-06, | |
| "loss": 0.2014, | |
| "step": 5645 | |
| }, | |
| { | |
| "epoch": 3.260242354298904, | |
| "grad_norm": 0.49132441725377424, | |
| "learning_rate": 2.0128704096187258e-06, | |
| "loss": 0.214, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 3.263127524523947, | |
| "grad_norm": 0.43559057805097934, | |
| "learning_rate": 1.9977439791072296e-06, | |
| "loss": 0.2025, | |
| "step": 5655 | |
| }, | |
| { | |
| "epoch": 3.2660126947489903, | |
| "grad_norm": 0.5032263108222483, | |
| "learning_rate": 1.9826682898531923e-06, | |
| "loss": 0.2036, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 3.2688978649740337, | |
| "grad_norm": 0.5166100007288694, | |
| "learning_rate": 1.967643437449591e-06, | |
| "loss": 0.2039, | |
| "step": 5665 | |
| }, | |
| { | |
| "epoch": 3.2717830351990767, | |
| "grad_norm": 0.5625840783677031, | |
| "learning_rate": 1.9526695171670428e-06, | |
| "loss": 0.2186, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 3.27466820542412, | |
| "grad_norm": 0.4771204637584591, | |
| "learning_rate": 1.937746623953218e-06, | |
| "loss": 0.2105, | |
| "step": 5675 | |
| }, | |
| { | |
| "epoch": 3.2775533756491635, | |
| "grad_norm": 0.5120534992728352, | |
| "learning_rate": 1.9228748524322283e-06, | |
| "loss": 0.1962, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 3.2804385458742065, | |
| "grad_norm": 0.5018775730767316, | |
| "learning_rate": 1.908054296904023e-06, | |
| "loss": 0.2109, | |
| "step": 5685 | |
| }, | |
| { | |
| "epoch": 3.28332371609925, | |
| "grad_norm": 0.4876912669877195, | |
| "learning_rate": 1.8932850513438139e-06, | |
| "loss": 0.2126, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 3.2862088863242933, | |
| "grad_norm": 0.4691715888051167, | |
| "learning_rate": 1.8785672094014484e-06, | |
| "loss": 0.2016, | |
| "step": 5695 | |
| }, | |
| { | |
| "epoch": 3.2890940565493363, | |
| "grad_norm": 0.48546733091545335, | |
| "learning_rate": 1.8639008644008317e-06, | |
| "loss": 0.2139, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 3.2919792267743797, | |
| "grad_norm": 0.48053507834963044, | |
| "learning_rate": 1.84928610933934e-06, | |
| "loss": 0.2069, | |
| "step": 5705 | |
| }, | |
| { | |
| "epoch": 3.294864396999423, | |
| "grad_norm": 0.4978213968888964, | |
| "learning_rate": 1.8347230368872227e-06, | |
| "loss": 0.2125, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 3.297749567224466, | |
| "grad_norm": 0.4619698699264187, | |
| "learning_rate": 1.8202117393870122e-06, | |
| "loss": 0.2048, | |
| "step": 5715 | |
| }, | |
| { | |
| "epoch": 3.3006347374495095, | |
| "grad_norm": 0.4439799217375983, | |
| "learning_rate": 1.805752308852945e-06, | |
| "loss": 0.2076, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 3.303519907674553, | |
| "grad_norm": 0.5050476605960446, | |
| "learning_rate": 1.7913448369703801e-06, | |
| "loss": 0.2148, | |
| "step": 5725 | |
| }, | |
| { | |
| "epoch": 3.306405077899596, | |
| "grad_norm": 0.4449280284195996, | |
| "learning_rate": 1.776989415095206e-06, | |
| "loss": 0.1961, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 3.3092902481246393, | |
| "grad_norm": 0.5023493469028406, | |
| "learning_rate": 1.7626861342532764e-06, | |
| "loss": 0.2108, | |
| "step": 5735 | |
| }, | |
| { | |
| "epoch": 3.3121754183496828, | |
| "grad_norm": 0.4899997511129661, | |
| "learning_rate": 1.748435085139818e-06, | |
| "loss": 0.2041, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 3.3150605885747257, | |
| "grad_norm": 0.620987392864056, | |
| "learning_rate": 1.7342363581188716e-06, | |
| "loss": 0.1967, | |
| "step": 5745 | |
| }, | |
| { | |
| "epoch": 3.317945758799769, | |
| "grad_norm": 0.48947716401563934, | |
| "learning_rate": 1.720090043222704e-06, | |
| "loss": 0.2093, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 3.3208309290248126, | |
| "grad_norm": 0.49193608057425164, | |
| "learning_rate": 1.705996230151239e-06, | |
| "loss": 0.2143, | |
| "step": 5755 | |
| }, | |
| { | |
| "epoch": 3.3237160992498556, | |
| "grad_norm": 0.5410390908943622, | |
| "learning_rate": 1.6919550082715108e-06, | |
| "loss": 0.2101, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 3.326601269474899, | |
| "grad_norm": 0.44748685099184915, | |
| "learning_rate": 1.6779664666170626e-06, | |
| "loss": 0.2154, | |
| "step": 5765 | |
| }, | |
| { | |
| "epoch": 3.3294864396999424, | |
| "grad_norm": 0.49202824127233535, | |
| "learning_rate": 1.6640306938874052e-06, | |
| "loss": 0.2084, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 3.3323716099249854, | |
| "grad_norm": 0.5129313111006824, | |
| "learning_rate": 1.650147778447444e-06, | |
| "loss": 0.2043, | |
| "step": 5775 | |
| }, | |
| { | |
| "epoch": 3.335256780150029, | |
| "grad_norm": 0.5011687713242771, | |
| "learning_rate": 1.6363178083269381e-06, | |
| "loss": 0.2137, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 3.338141950375072, | |
| "grad_norm": 0.5125645791853729, | |
| "learning_rate": 1.6225408712199097e-06, | |
| "loss": 0.213, | |
| "step": 5785 | |
| }, | |
| { | |
| "epoch": 3.341027120600115, | |
| "grad_norm": 0.496426542842325, | |
| "learning_rate": 1.608817054484113e-06, | |
| "loss": 0.2036, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 3.3439122908251586, | |
| "grad_norm": 0.573443010809008, | |
| "learning_rate": 1.5951464451404775e-06, | |
| "loss": 0.1975, | |
| "step": 5795 | |
| }, | |
| { | |
| "epoch": 3.346797461050202, | |
| "grad_norm": 0.4777011846082858, | |
| "learning_rate": 1.5815291298725387e-06, | |
| "loss": 0.2278, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 3.349682631275245, | |
| "grad_norm": 0.4994285452680278, | |
| "learning_rate": 1.5679651950259178e-06, | |
| "loss": 0.2019, | |
| "step": 5805 | |
| }, | |
| { | |
| "epoch": 3.3525678015002884, | |
| "grad_norm": 0.47305655838061333, | |
| "learning_rate": 1.5544547266077425e-06, | |
| "loss": 0.2121, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 3.355452971725332, | |
| "grad_norm": 0.5117964943614841, | |
| "learning_rate": 1.54099781028613e-06, | |
| "loss": 0.2117, | |
| "step": 5815 | |
| }, | |
| { | |
| "epoch": 3.358338141950375, | |
| "grad_norm": 0.47969243336386125, | |
| "learning_rate": 1.52759453138962e-06, | |
| "loss": 0.2103, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 3.3612233121754183, | |
| "grad_norm": 0.46863801053654247, | |
| "learning_rate": 1.514244974906649e-06, | |
| "loss": 0.2162, | |
| "step": 5825 | |
| }, | |
| { | |
| "epoch": 3.3641084824004617, | |
| "grad_norm": 0.4937087047605848, | |
| "learning_rate": 1.5009492254850056e-06, | |
| "loss": 0.2074, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 3.366993652625505, | |
| "grad_norm": 0.5349150058728949, | |
| "learning_rate": 1.4877073674313004e-06, | |
| "loss": 0.2092, | |
| "step": 5835 | |
| }, | |
| { | |
| "epoch": 3.369878822850548, | |
| "grad_norm": 0.509590831524668, | |
| "learning_rate": 1.4745194847104184e-06, | |
| "loss": 0.2258, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 3.3727639930755915, | |
| "grad_norm": 0.5411363420436703, | |
| "learning_rate": 1.461385660944994e-06, | |
| "loss": 0.2026, | |
| "step": 5845 | |
| }, | |
| { | |
| "epoch": 3.375649163300635, | |
| "grad_norm": 0.4869231603073319, | |
| "learning_rate": 1.4483059794148869e-06, | |
| "loss": 0.212, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 3.378534333525678, | |
| "grad_norm": 0.5066221127114735, | |
| "learning_rate": 1.4352805230566458e-06, | |
| "loss": 0.2066, | |
| "step": 5855 | |
| }, | |
| { | |
| "epoch": 3.3814195037507213, | |
| "grad_norm": 0.49848266645770745, | |
| "learning_rate": 1.4223093744629802e-06, | |
| "loss": 0.2039, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 3.3843046739757647, | |
| "grad_norm": 0.46674023509056656, | |
| "learning_rate": 1.409392615882248e-06, | |
| "loss": 0.2056, | |
| "step": 5865 | |
| }, | |
| { | |
| "epoch": 3.3871898442008077, | |
| "grad_norm": 0.47153344056400004, | |
| "learning_rate": 1.3965303292179211e-06, | |
| "loss": 0.2035, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 3.390075014425851, | |
| "grad_norm": 0.5737791203551417, | |
| "learning_rate": 1.3837225960280777e-06, | |
| "loss": 0.2092, | |
| "step": 5875 | |
| }, | |
| { | |
| "epoch": 3.3929601846508946, | |
| "grad_norm": 0.5430927208050875, | |
| "learning_rate": 1.370969497524872e-06, | |
| "loss": 0.2102, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 3.3958453548759375, | |
| "grad_norm": 0.4648636735503783, | |
| "learning_rate": 1.3582711145740378e-06, | |
| "loss": 0.2036, | |
| "step": 5885 | |
| }, | |
| { | |
| "epoch": 3.398730525100981, | |
| "grad_norm": 0.4809453402273537, | |
| "learning_rate": 1.3456275276943543e-06, | |
| "loss": 0.206, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 3.4016156953260244, | |
| "grad_norm": 0.5512582459841553, | |
| "learning_rate": 1.3330388170571496e-06, | |
| "loss": 0.2057, | |
| "step": 5895 | |
| }, | |
| { | |
| "epoch": 3.4045008655510673, | |
| "grad_norm": 0.5125832192345321, | |
| "learning_rate": 1.3205050624857895e-06, | |
| "loss": 0.2065, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 3.4073860357761108, | |
| "grad_norm": 0.5558586586402942, | |
| "learning_rate": 1.3080263434551743e-06, | |
| "loss": 0.2166, | |
| "step": 5905 | |
| }, | |
| { | |
| "epoch": 3.410271206001154, | |
| "grad_norm": 0.48236963618343165, | |
| "learning_rate": 1.295602739091224e-06, | |
| "loss": 0.2054, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 3.413156376226197, | |
| "grad_norm": 0.47170396426714695, | |
| "learning_rate": 1.2832343281703852e-06, | |
| "loss": 0.2102, | |
| "step": 5915 | |
| }, | |
| { | |
| "epoch": 3.4160415464512406, | |
| "grad_norm": 0.4953986618513792, | |
| "learning_rate": 1.2709211891191331e-06, | |
| "loss": 0.2262, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 3.418926716676284, | |
| "grad_norm": 0.5094102635524805, | |
| "learning_rate": 1.2586634000134735e-06, | |
| "loss": 0.2052, | |
| "step": 5925 | |
| }, | |
| { | |
| "epoch": 3.4218118869013274, | |
| "grad_norm": 0.46465406913045626, | |
| "learning_rate": 1.2464610385784381e-06, | |
| "loss": 0.2166, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 3.4246970571263704, | |
| "grad_norm": 0.537941105319963, | |
| "learning_rate": 1.2343141821876015e-06, | |
| "loss": 0.2052, | |
| "step": 5935 | |
| }, | |
| { | |
| "epoch": 3.427582227351414, | |
| "grad_norm": 0.4708366082196696, | |
| "learning_rate": 1.2222229078625935e-06, | |
| "loss": 0.2045, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 3.4304673975764572, | |
| "grad_norm": 0.4910988413622275, | |
| "learning_rate": 1.2101872922725976e-06, | |
| "loss": 0.2131, | |
| "step": 5945 | |
| }, | |
| { | |
| "epoch": 3.4333525678015, | |
| "grad_norm": 0.4726816646220706, | |
| "learning_rate": 1.198207411733877e-06, | |
| "loss": 0.2126, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 3.4362377380265436, | |
| "grad_norm": 0.499819420315185, | |
| "learning_rate": 1.1862833422092878e-06, | |
| "loss": 0.2053, | |
| "step": 5955 | |
| }, | |
| { | |
| "epoch": 3.439122908251587, | |
| "grad_norm": 0.45811978260654357, | |
| "learning_rate": 1.1744151593077925e-06, | |
| "loss": 0.2166, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 3.44200807847663, | |
| "grad_norm": 0.49132678636515253, | |
| "learning_rate": 1.1626029382839798e-06, | |
| "loss": 0.2093, | |
| "step": 5965 | |
| }, | |
| { | |
| "epoch": 3.4448932487016735, | |
| "grad_norm": 0.5178705009636367, | |
| "learning_rate": 1.1508467540376e-06, | |
| "loss": 0.1978, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 3.447778418926717, | |
| "grad_norm": 0.4771433582042318, | |
| "learning_rate": 1.1391466811130813e-06, | |
| "loss": 0.2055, | |
| "step": 5975 | |
| }, | |
| { | |
| "epoch": 3.45066358915176, | |
| "grad_norm": 0.4849667450545781, | |
| "learning_rate": 1.127502793699049e-06, | |
| "loss": 0.2114, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 3.4535487593768033, | |
| "grad_norm": 0.5323744822003759, | |
| "learning_rate": 1.1159151656278677e-06, | |
| "loss": 0.2246, | |
| "step": 5985 | |
| }, | |
| { | |
| "epoch": 3.4564339296018467, | |
| "grad_norm": 0.517295087718039, | |
| "learning_rate": 1.104383870375172e-06, | |
| "loss": 0.2033, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 3.4593190998268897, | |
| "grad_norm": 0.49882873868199235, | |
| "learning_rate": 1.0929089810593974e-06, | |
| "loss": 0.2072, | |
| "step": 5995 | |
| }, | |
| { | |
| "epoch": 3.462204270051933, | |
| "grad_norm": 0.5007022327120706, | |
| "learning_rate": 1.0814905704413093e-06, | |
| "loss": 0.2257, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 3.4650894402769765, | |
| "grad_norm": 0.46291584199269026, | |
| "learning_rate": 1.070128710923556e-06, | |
| "loss": 0.2086, | |
| "step": 6005 | |
| }, | |
| { | |
| "epoch": 3.4679746105020195, | |
| "grad_norm": 0.5269534068567228, | |
| "learning_rate": 1.0588234745502012e-06, | |
| "loss": 0.2112, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 3.470859780727063, | |
| "grad_norm": 0.4910407597306869, | |
| "learning_rate": 1.047574933006268e-06, | |
| "loss": 0.2177, | |
| "step": 6015 | |
| }, | |
| { | |
| "epoch": 3.4737449509521063, | |
| "grad_norm": 0.47260522012547096, | |
| "learning_rate": 1.0363831576172834e-06, | |
| "loss": 0.2001, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 3.4766301211771493, | |
| "grad_norm": 0.46807651107457027, | |
| "learning_rate": 1.0252482193488345e-06, | |
| "loss": 0.1936, | |
| "step": 6025 | |
| }, | |
| { | |
| "epoch": 3.4795152914021927, | |
| "grad_norm": 0.4496096381304417, | |
| "learning_rate": 1.0141701888061018e-06, | |
| "loss": 0.2051, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 3.482400461627236, | |
| "grad_norm": 0.48460781075602494, | |
| "learning_rate": 1.0031491362334289e-06, | |
| "loss": 0.2025, | |
| "step": 6035 | |
| }, | |
| { | |
| "epoch": 3.485285631852279, | |
| "grad_norm": 0.5182489885112946, | |
| "learning_rate": 9.92185131513862e-07, | |
| "loss": 0.2167, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 3.4881708020773226, | |
| "grad_norm": 0.48427730291008736, | |
| "learning_rate": 9.812782441687263e-07, | |
| "loss": 0.2024, | |
| "step": 6045 | |
| }, | |
| { | |
| "epoch": 3.491055972302366, | |
| "grad_norm": 0.5074747506126092, | |
| "learning_rate": 9.704285433571637e-07, | |
| "loss": 0.2143, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 3.493941142527409, | |
| "grad_norm": 0.4838143459849086, | |
| "learning_rate": 9.596360978757025e-07, | |
| "loss": 0.2139, | |
| "step": 6055 | |
| }, | |
| { | |
| "epoch": 3.4968263127524524, | |
| "grad_norm": 0.47554491034770197, | |
| "learning_rate": 9.489009761578294e-07, | |
| "loss": 0.206, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 3.499711482977496, | |
| "grad_norm": 0.5171439253070426, | |
| "learning_rate": 9.382232462735463e-07, | |
| "loss": 0.2044, | |
| "step": 6065 | |
| }, | |
| { | |
| "epoch": 3.5025966532025388, | |
| "grad_norm": 0.5115632720888774, | |
| "learning_rate": 9.276029759289384e-07, | |
| "loss": 0.2195, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 3.505481823427582, | |
| "grad_norm": 0.5128873018700219, | |
| "learning_rate": 9.17040232465748e-07, | |
| "loss": 0.208, | |
| "step": 6075 | |
| }, | |
| { | |
| "epoch": 3.5083669936526256, | |
| "grad_norm": 0.47047254608971883, | |
| "learning_rate": 9.065350828609509e-07, | |
| "loss": 0.2166, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 3.5112521638776686, | |
| "grad_norm": 0.47331390143331864, | |
| "learning_rate": 8.960875937263214e-07, | |
| "loss": 0.1926, | |
| "step": 6085 | |
| }, | |
| { | |
| "epoch": 3.514137334102712, | |
| "grad_norm": 0.5526849091013493, | |
| "learning_rate": 8.856978313080244e-07, | |
| "loss": 0.208, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 3.5170225043277554, | |
| "grad_norm": 0.5056598583710792, | |
| "learning_rate": 8.753658614861782e-07, | |
| "loss": 0.202, | |
| "step": 6095 | |
| }, | |
| { | |
| "epoch": 3.5199076745527984, | |
| "grad_norm": 0.45972967026892514, | |
| "learning_rate": 8.650917497744548e-07, | |
| "loss": 0.2052, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 3.522792844777842, | |
| "grad_norm": 0.9905171948522371, | |
| "learning_rate": 8.548755613196491e-07, | |
| "loss": 0.1975, | |
| "step": 6105 | |
| }, | |
| { | |
| "epoch": 3.5256780150028852, | |
| "grad_norm": 0.487725697200175, | |
| "learning_rate": 8.447173609012693e-07, | |
| "loss": 0.2003, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 3.5285631852279282, | |
| "grad_norm": 0.436842348667491, | |
| "learning_rate": 8.346172129311436e-07, | |
| "loss": 0.2002, | |
| "step": 6115 | |
| }, | |
| { | |
| "epoch": 3.5314483554529716, | |
| "grad_norm": 0.5324360594518225, | |
| "learning_rate": 8.245751814529812e-07, | |
| "loss": 0.2042, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 3.534333525678015, | |
| "grad_norm": 0.49501211496626846, | |
| "learning_rate": 8.145913301419894e-07, | |
| "loss": 0.2046, | |
| "step": 6125 | |
| }, | |
| { | |
| "epoch": 3.537218695903058, | |
| "grad_norm": 0.5636032763069877, | |
| "learning_rate": 8.046657223044618e-07, | |
| "loss": 0.2065, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 3.5401038661281015, | |
| "grad_norm": 0.47764396959998434, | |
| "learning_rate": 7.947984208773796e-07, | |
| "loss": 0.201, | |
| "step": 6135 | |
| }, | |
| { | |
| "epoch": 3.542989036353145, | |
| "grad_norm": 0.47110105270453406, | |
| "learning_rate": 7.849894884280085e-07, | |
| "loss": 0.1939, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 3.545874206578188, | |
| "grad_norm": 0.5089791473431078, | |
| "learning_rate": 7.752389871535027e-07, | |
| "loss": 0.2237, | |
| "step": 6145 | |
| }, | |
| { | |
| "epoch": 3.5487593768032313, | |
| "grad_norm": 0.4693514320901827, | |
| "learning_rate": 7.65546978880517e-07, | |
| "loss": 0.1918, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 3.5516445470282747, | |
| "grad_norm": 0.5339146402486223, | |
| "learning_rate": 7.55913525064802e-07, | |
| "loss": 0.2076, | |
| "step": 6155 | |
| }, | |
| { | |
| "epoch": 3.5545297172533177, | |
| "grad_norm": 0.510340675728564, | |
| "learning_rate": 7.463386867908318e-07, | |
| "loss": 0.1947, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 3.557414887478361, | |
| "grad_norm": 0.5063108850560543, | |
| "learning_rate": 7.368225247713978e-07, | |
| "loss": 0.2293, | |
| "step": 6165 | |
| }, | |
| { | |
| "epoch": 3.5603000577034045, | |
| "grad_norm": 0.5151597458874256, | |
| "learning_rate": 7.273650993472414e-07, | |
| "loss": 0.2027, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 3.563185227928448, | |
| "grad_norm": 0.6646624866313264, | |
| "learning_rate": 7.179664704866551e-07, | |
| "loss": 0.1955, | |
| "step": 6175 | |
| }, | |
| { | |
| "epoch": 3.566070398153491, | |
| "grad_norm": 0.4766136250823664, | |
| "learning_rate": 7.086266977851141e-07, | |
| "loss": 0.2012, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 3.5689555683785343, | |
| "grad_norm": 0.4840543281228228, | |
| "learning_rate": 6.993458404648945e-07, | |
| "loss": 0.2069, | |
| "step": 6185 | |
| }, | |
| { | |
| "epoch": 3.5718407386035778, | |
| "grad_norm": 0.49635401265932044, | |
| "learning_rate": 6.901239573746987e-07, | |
| "loss": 0.2054, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 3.5747259088286207, | |
| "grad_norm": 0.5178071858503266, | |
| "learning_rate": 6.809611069892785e-07, | |
| "loss": 0.2114, | |
| "step": 6195 | |
| }, | |
| { | |
| "epoch": 3.577611079053664, | |
| "grad_norm": 0.4805032184640311, | |
| "learning_rate": 6.718573474090673e-07, | |
| "loss": 0.2088, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 3.5804962492787076, | |
| "grad_norm": 0.5312256937793656, | |
| "learning_rate": 6.628127363598125e-07, | |
| "loss": 0.218, | |
| "step": 6205 | |
| }, | |
| { | |
| "epoch": 3.583381419503751, | |
| "grad_norm": 0.5038683897704207, | |
| "learning_rate": 6.538273311922105e-07, | |
| "loss": 0.2036, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 3.586266589728794, | |
| "grad_norm": 0.48973488364704354, | |
| "learning_rate": 6.449011888815359e-07, | |
| "loss": 0.2059, | |
| "step": 6215 | |
| }, | |
| { | |
| "epoch": 3.5891517599538374, | |
| "grad_norm": 0.46502727902053664, | |
| "learning_rate": 6.360343660272905e-07, | |
| "loss": 0.2032, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 3.592036930178881, | |
| "grad_norm": 0.4693993644882787, | |
| "learning_rate": 6.272269188528346e-07, | |
| "loss": 0.1993, | |
| "step": 6225 | |
| }, | |
| { | |
| "epoch": 3.594922100403924, | |
| "grad_norm": 0.4759028304516473, | |
| "learning_rate": 6.184789032050408e-07, | |
| "loss": 0.1937, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 3.597807270628967, | |
| "grad_norm": 0.5255955515623099, | |
| "learning_rate": 6.097903745539258e-07, | |
| "loss": 0.2051, | |
| "step": 6235 | |
| }, | |
| { | |
| "epoch": 3.6006924408540106, | |
| "grad_norm": 0.48476442425010263, | |
| "learning_rate": 6.011613879923161e-07, | |
| "loss": 0.1946, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 3.6035776110790536, | |
| "grad_norm": 0.5048238073660125, | |
| "learning_rate": 5.925919982354844e-07, | |
| "loss": 0.2122, | |
| "step": 6245 | |
| }, | |
| { | |
| "epoch": 3.606462781304097, | |
| "grad_norm": 0.5137209328897905, | |
| "learning_rate": 5.840822596208073e-07, | |
| "loss": 0.2078, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 3.6093479515291405, | |
| "grad_norm": 0.46932455507505844, | |
| "learning_rate": 5.756322261074232e-07, | |
| "loss": 0.2105, | |
| "step": 6255 | |
| }, | |
| { | |
| "epoch": 3.6122331217541834, | |
| "grad_norm": 0.47438897533582397, | |
| "learning_rate": 5.672419512758909e-07, | |
| "loss": 0.2046, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 3.615118291979227, | |
| "grad_norm": 0.49163449448043284, | |
| "learning_rate": 5.589114883278423e-07, | |
| "loss": 0.2106, | |
| "step": 6265 | |
| }, | |
| { | |
| "epoch": 3.6180034622042703, | |
| "grad_norm": 0.4948571949148204, | |
| "learning_rate": 5.506408900856508e-07, | |
| "loss": 0.2019, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 3.6208886324293132, | |
| "grad_norm": 0.5239387280536255, | |
| "learning_rate": 5.424302089920974e-07, | |
| "loss": 0.2152, | |
| "step": 6275 | |
| }, | |
| { | |
| "epoch": 3.6237738026543567, | |
| "grad_norm": 0.4645502539433645, | |
| "learning_rate": 5.342794971100374e-07, | |
| "loss": 0.2013, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 3.6266589728794, | |
| "grad_norm": 0.5644959774991677, | |
| "learning_rate": 5.261888061220643e-07, | |
| "loss": 0.2022, | |
| "step": 6285 | |
| }, | |
| { | |
| "epoch": 3.629544143104443, | |
| "grad_norm": 0.48442583058736516, | |
| "learning_rate": 5.181581873301944e-07, | |
| "loss": 0.2036, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 3.6324293133294865, | |
| "grad_norm": 0.515528269609854, | |
| "learning_rate": 5.101876916555271e-07, | |
| "loss": 0.204, | |
| "step": 6295 | |
| }, | |
| { | |
| "epoch": 3.63531448355453, | |
| "grad_norm": 0.5338898797051952, | |
| "learning_rate": 5.022773696379313e-07, | |
| "loss": 0.2033, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 3.638199653779573, | |
| "grad_norm": 0.4794106672832799, | |
| "learning_rate": 4.944272714357223e-07, | |
| "loss": 0.2061, | |
| "step": 6305 | |
| }, | |
| { | |
| "epoch": 3.6410848240046163, | |
| "grad_norm": 0.49326404127777695, | |
| "learning_rate": 4.866374468253487e-07, | |
| "loss": 0.1928, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 3.6439699942296597, | |
| "grad_norm": 0.5247706559589024, | |
| "learning_rate": 4.789079452010648e-07, | |
| "loss": 0.2058, | |
| "step": 6315 | |
| }, | |
| { | |
| "epoch": 3.6468551644547027, | |
| "grad_norm": 0.4781901107121829, | |
| "learning_rate": 4.7123881557462834e-07, | |
| "loss": 0.2045, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 3.649740334679746, | |
| "grad_norm": 0.4743243681177699, | |
| "learning_rate": 4.6363010657498597e-07, | |
| "loss": 0.2118, | |
| "step": 6325 | |
| }, | |
| { | |
| "epoch": 3.6526255049047895, | |
| "grad_norm": 0.5180788089096849, | |
| "learning_rate": 4.560818664479671e-07, | |
| "loss": 0.1981, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 3.6555106751298325, | |
| "grad_norm": 0.479204470483921, | |
| "learning_rate": 4.485941430559726e-07, | |
| "loss": 0.2242, | |
| "step": 6335 | |
| }, | |
| { | |
| "epoch": 3.658395845354876, | |
| "grad_norm": 0.49576155614359946, | |
| "learning_rate": 4.411669838776733e-07, | |
| "loss": 0.1973, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 3.6612810155799194, | |
| "grad_norm": 0.48614229875693404, | |
| "learning_rate": 4.338004360077164e-07, | |
| "loss": 0.1984, | |
| "step": 6345 | |
| }, | |
| { | |
| "epoch": 3.6641661858049623, | |
| "grad_norm": 0.46848988610190684, | |
| "learning_rate": 4.2649454615641737e-07, | |
| "loss": 0.2052, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 3.6670513560300058, | |
| "grad_norm": 0.5277331889272244, | |
| "learning_rate": 4.1924936064946633e-07, | |
| "loss": 0.1966, | |
| "step": 6355 | |
| }, | |
| { | |
| "epoch": 3.669936526255049, | |
| "grad_norm": 0.48392147766138405, | |
| "learning_rate": 4.120649254276321e-07, | |
| "loss": 0.2009, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 3.672821696480092, | |
| "grad_norm": 0.48495642382601156, | |
| "learning_rate": 4.04941286046483e-07, | |
| "loss": 0.2158, | |
| "step": 6365 | |
| }, | |
| { | |
| "epoch": 3.6757068667051356, | |
| "grad_norm": 0.49853651987090997, | |
| "learning_rate": 3.9787848767607997e-07, | |
| "loss": 0.2097, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 3.678592036930179, | |
| "grad_norm": 0.49229974849216307, | |
| "learning_rate": 3.908765751007038e-07, | |
| "loss": 0.2174, | |
| "step": 6375 | |
| }, | |
| { | |
| "epoch": 3.681477207155222, | |
| "grad_norm": 0.49794648252416773, | |
| "learning_rate": 3.839355927185662e-07, | |
| "loss": 0.2119, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 3.6843623773802654, | |
| "grad_norm": 0.7006970650191741, | |
| "learning_rate": 3.770555845415258e-07, | |
| "loss": 0.2038, | |
| "step": 6385 | |
| }, | |
| { | |
| "epoch": 3.687247547605309, | |
| "grad_norm": 0.5062300259095932, | |
| "learning_rate": 3.7023659419481275e-07, | |
| "loss": 0.2026, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 3.690132717830352, | |
| "grad_norm": 0.49987127156630407, | |
| "learning_rate": 3.6347866491675323e-07, | |
| "loss": 0.2016, | |
| "step": 6395 | |
| }, | |
| { | |
| "epoch": 3.693017888055395, | |
| "grad_norm": 0.5019423482974231, | |
| "learning_rate": 3.5678183955849323e-07, | |
| "loss": 0.2123, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 3.6959030582804386, | |
| "grad_norm": 0.49420552946041524, | |
| "learning_rate": 3.5014616058372306e-07, | |
| "loss": 0.2066, | |
| "step": 6405 | |
| }, | |
| { | |
| "epoch": 3.6987882285054816, | |
| "grad_norm": 0.483613460902034, | |
| "learning_rate": 3.435716700684133e-07, | |
| "loss": 0.2142, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 3.701673398730525, | |
| "grad_norm": 0.8317874119600014, | |
| "learning_rate": 3.3705840970054916e-07, | |
| "loss": 0.1973, | |
| "step": 6415 | |
| }, | |
| { | |
| "epoch": 3.7045585689555685, | |
| "grad_norm": 0.4583040079124961, | |
| "learning_rate": 3.3060642077986313e-07, | |
| "loss": 0.2027, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 3.7074437391806114, | |
| "grad_norm": 0.4698006325080622, | |
| "learning_rate": 3.242157442175686e-07, | |
| "loss": 0.1926, | |
| "step": 6425 | |
| }, | |
| { | |
| "epoch": 3.710328909405655, | |
| "grad_norm": 0.4935876897031322, | |
| "learning_rate": 3.1788642053610984e-07, | |
| "loss": 0.1982, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 3.7132140796306983, | |
| "grad_norm": 0.4871065386845346, | |
| "learning_rate": 3.1161848986890117e-07, | |
| "loss": 0.201, | |
| "step": 6435 | |
| }, | |
| { | |
| "epoch": 3.7160992498557412, | |
| "grad_norm": 0.5134006486764362, | |
| "learning_rate": 3.054119919600673e-07, | |
| "loss": 0.2103, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 3.7189844200807847, | |
| "grad_norm": 0.5247799897213314, | |
| "learning_rate": 2.9926696616420227e-07, | |
| "loss": 0.2158, | |
| "step": 6445 | |
| }, | |
| { | |
| "epoch": 3.721869590305828, | |
| "grad_norm": 0.45147630286318213, | |
| "learning_rate": 2.9318345144610627e-07, | |
| "loss": 0.2007, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 3.724754760530871, | |
| "grad_norm": 0.480778845212167, | |
| "learning_rate": 2.8716148638055166e-07, | |
| "loss": 0.216, | |
| "step": 6455 | |
| }, | |
| { | |
| "epoch": 3.7276399307559145, | |
| "grad_norm": 0.49023422691624635, | |
| "learning_rate": 2.8120110915202945e-07, | |
| "loss": 0.2131, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 3.730525100980958, | |
| "grad_norm": 0.5078622584490374, | |
| "learning_rate": 2.7530235755450886e-07, | |
| "loss": 0.2085, | |
| "step": 6465 | |
| }, | |
| { | |
| "epoch": 3.733410271206001, | |
| "grad_norm": 0.527238182528253, | |
| "learning_rate": 2.6946526899120693e-07, | |
| "loss": 0.2076, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 3.7362954414310443, | |
| "grad_norm": 0.4829671940741909, | |
| "learning_rate": 2.6368988047433373e-07, | |
| "loss": 0.1978, | |
| "step": 6475 | |
| }, | |
| { | |
| "epoch": 3.7391806116560877, | |
| "grad_norm": 0.5127022353864252, | |
| "learning_rate": 2.5797622862487104e-07, | |
| "loss": 0.2249, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 3.742065781881131, | |
| "grad_norm": 0.4378947388730543, | |
| "learning_rate": 2.523243496723382e-07, | |
| "loss": 0.1999, | |
| "step": 6485 | |
| }, | |
| { | |
| "epoch": 3.744950952106174, | |
| "grad_norm": 0.4730083265761743, | |
| "learning_rate": 2.467342794545613e-07, | |
| "loss": 0.2036, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 3.7478361223312175, | |
| "grad_norm": 0.5083579379173573, | |
| "learning_rate": 2.412060534174421e-07, | |
| "loss": 0.2134, | |
| "step": 6495 | |
| }, | |
| { | |
| "epoch": 3.750721292556261, | |
| "grad_norm": 0.5014960077674261, | |
| "learning_rate": 2.35739706614736e-07, | |
| "loss": 0.2066, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 3.753606462781304, | |
| "grad_norm": 0.505196404447519, | |
| "learning_rate": 2.3033527370783459e-07, | |
| "loss": 0.2118, | |
| "step": 6505 | |
| }, | |
| { | |
| "epoch": 3.7564916330063474, | |
| "grad_norm": 0.4720849386750261, | |
| "learning_rate": 2.2499278896553678e-07, | |
| "loss": 0.2176, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 3.759376803231391, | |
| "grad_norm": 0.49946759486920755, | |
| "learning_rate": 2.1971228626384233e-07, | |
| "loss": 0.2073, | |
| "step": 6515 | |
| }, | |
| { | |
| "epoch": 3.762261973456434, | |
| "grad_norm": 0.4836314820421101, | |
| "learning_rate": 2.1449379908572431e-07, | |
| "loss": 0.2079, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 3.765147143681477, | |
| "grad_norm": 0.4848667800466061, | |
| "learning_rate": 2.0933736052092812e-07, | |
| "loss": 0.2244, | |
| "step": 6525 | |
| }, | |
| { | |
| "epoch": 3.7680323139065206, | |
| "grad_norm": 0.4735713481776197, | |
| "learning_rate": 2.0424300326575497e-07, | |
| "loss": 0.2018, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 3.770917484131564, | |
| "grad_norm": 0.5176491575721175, | |
| "learning_rate": 1.992107596228554e-07, | |
| "loss": 0.2109, | |
| "step": 6535 | |
| }, | |
| { | |
| "epoch": 3.773802654356607, | |
| "grad_norm": 0.49048701673925643, | |
| "learning_rate": 1.942406615010295e-07, | |
| "loss": 0.2048, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 3.7766878245816504, | |
| "grad_norm": 0.4951411517869068, | |
| "learning_rate": 1.8933274041501915e-07, | |
| "loss": 0.1982, | |
| "step": 6545 | |
| }, | |
| { | |
| "epoch": 3.779572994806694, | |
| "grad_norm": 0.4615884792028935, | |
| "learning_rate": 1.8448702748530723e-07, | |
| "loss": 0.2105, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 3.782458165031737, | |
| "grad_norm": 0.503114785095133, | |
| "learning_rate": 1.7970355343792434e-07, | |
| "loss": 0.1995, | |
| "step": 6555 | |
| }, | |
| { | |
| "epoch": 3.7853433352567802, | |
| "grad_norm": 0.4531549414892555, | |
| "learning_rate": 1.7498234860425567e-07, | |
| "loss": 0.1954, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 3.7882285054818237, | |
| "grad_norm": 0.4476621701726881, | |
| "learning_rate": 1.7032344292084112e-07, | |
| "loss": 0.207, | |
| "step": 6565 | |
| }, | |
| { | |
| "epoch": 3.7911136757068666, | |
| "grad_norm": 0.5301338039777519, | |
| "learning_rate": 1.6572686592919107e-07, | |
| "loss": 0.2047, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 3.79399884593191, | |
| "grad_norm": 0.5215830928997757, | |
| "learning_rate": 1.6119264677559977e-07, | |
| "loss": 0.2164, | |
| "step": 6575 | |
| }, | |
| { | |
| "epoch": 3.7968840161569535, | |
| "grad_norm": 0.4581052464950232, | |
| "learning_rate": 1.5672081421095552e-07, | |
| "loss": 0.1954, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 3.7997691863819965, | |
| "grad_norm": 0.4658151305069496, | |
| "learning_rate": 1.5231139659056426e-07, | |
| "loss": 0.2065, | |
| "step": 6585 | |
| }, | |
| { | |
| "epoch": 3.80265435660704, | |
| "grad_norm": 0.5087191746251587, | |
| "learning_rate": 1.4796442187396397e-07, | |
| "loss": 0.2027, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 3.8055395268320833, | |
| "grad_norm": 0.5505757142185078, | |
| "learning_rate": 1.4367991762475497e-07, | |
| "loss": 0.2095, | |
| "step": 6595 | |
| }, | |
| { | |
| "epoch": 3.8084246970571263, | |
| "grad_norm": 0.5414562907573153, | |
| "learning_rate": 1.394579110104144e-07, | |
| "loss": 0.2143, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 3.8113098672821697, | |
| "grad_norm": 0.46524927624850704, | |
| "learning_rate": 1.352984288021353e-07, | |
| "loss": 0.1978, | |
| "step": 6605 | |
| }, | |
| { | |
| "epoch": 3.814195037507213, | |
| "grad_norm": 0.4628457981686407, | |
| "learning_rate": 1.3120149737464893e-07, | |
| "loss": 0.201, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 3.817080207732256, | |
| "grad_norm": 0.4966243391627818, | |
| "learning_rate": 1.2716714270606057e-07, | |
| "loss": 0.2002, | |
| "step": 6615 | |
| }, | |
| { | |
| "epoch": 3.8199653779572995, | |
| "grad_norm": 0.46771609095809985, | |
| "learning_rate": 1.2319539037768614e-07, | |
| "loss": 0.2009, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 3.822850548182343, | |
| "grad_norm": 0.47874741702870294, | |
| "learning_rate": 1.192862655738858e-07, | |
| "loss": 0.2091, | |
| "step": 6625 | |
| }, | |
| { | |
| "epoch": 3.825735718407386, | |
| "grad_norm": 0.48329839687108117, | |
| "learning_rate": 1.1543979308190844e-07, | |
| "loss": 0.2005, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 3.8286208886324293, | |
| "grad_norm": 0.4891511461415484, | |
| "learning_rate": 1.1165599729173193e-07, | |
| "loss": 0.2088, | |
| "step": 6635 | |
| }, | |
| { | |
| "epoch": 3.8315060588574728, | |
| "grad_norm": 0.52591217962733, | |
| "learning_rate": 1.0793490219591085e-07, | |
| "loss": 0.2048, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 3.8343912290825157, | |
| "grad_norm": 0.49626455120247676, | |
| "learning_rate": 1.0427653138942118e-07, | |
| "loss": 0.2033, | |
| "step": 6645 | |
| }, | |
| { | |
| "epoch": 3.837276399307559, | |
| "grad_norm": 0.4673741618418967, | |
| "learning_rate": 1.0068090806951369e-07, | |
| "loss": 0.221, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 3.8401615695326026, | |
| "grad_norm": 0.47678344134659506, | |
| "learning_rate": 9.714805503556413e-08, | |
| "loss": 0.2276, | |
| "step": 6655 | |
| }, | |
| { | |
| "epoch": 3.8430467397576455, | |
| "grad_norm": 1.143709435388536, | |
| "learning_rate": 9.367799468893102e-08, | |
| "loss": 0.1927, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 3.845931909982689, | |
| "grad_norm": 0.46684620921540665, | |
| "learning_rate": 9.027074903281364e-08, | |
| "loss": 0.1986, | |
| "step": 6665 | |
| }, | |
| { | |
| "epoch": 3.8488170802077324, | |
| "grad_norm": 0.5307872034727703, | |
| "learning_rate": 8.692633967210872e-08, | |
| "loss": 0.2188, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 3.8517022504327754, | |
| "grad_norm": 0.4735022804301449, | |
| "learning_rate": 8.364478781327956e-08, | |
| "loss": 0.2141, | |
| "step": 6675 | |
| }, | |
| { | |
| "epoch": 3.854587420657819, | |
| "grad_norm": 0.45935591835994294, | |
| "learning_rate": 8.042611426421598e-08, | |
| "loss": 0.2084, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 3.857472590882862, | |
| "grad_norm": 0.515095531011111, | |
| "learning_rate": 7.727033943410678e-08, | |
| "loss": 0.2099, | |
| "step": 6685 | |
| }, | |
| { | |
| "epoch": 3.860357761107905, | |
| "grad_norm": 0.48719846805820766, | |
| "learning_rate": 7.417748333330533e-08, | |
| "loss": 0.2108, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 3.8632429313329486, | |
| "grad_norm": 0.4535374141566299, | |
| "learning_rate": 7.114756557320745e-08, | |
| "loss": 0.2101, | |
| "step": 6695 | |
| }, | |
| { | |
| "epoch": 3.866128101557992, | |
| "grad_norm": 0.48820441041592805, | |
| "learning_rate": 6.818060536612381e-08, | |
| "loss": 0.1999, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 3.869013271783035, | |
| "grad_norm": 0.4752993276059485, | |
| "learning_rate": 6.527662152516323e-08, | |
| "loss": 0.1855, | |
| "step": 6705 | |
| }, | |
| { | |
| "epoch": 3.8718984420080784, | |
| "grad_norm": 0.4986229869294503, | |
| "learning_rate": 6.243563246410622e-08, | |
| "loss": 0.2105, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 3.874783612233122, | |
| "grad_norm": 0.49185252551299946, | |
| "learning_rate": 5.965765619728947e-08, | |
| "loss": 0.2239, | |
| "step": 6715 | |
| }, | |
| { | |
| "epoch": 3.877668782458165, | |
| "grad_norm": 0.5713339630980684, | |
| "learning_rate": 5.694271033950038e-08, | |
| "loss": 0.2193, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 3.8805539526832082, | |
| "grad_norm": 0.5344135726437659, | |
| "learning_rate": 5.429081210585274e-08, | |
| "loss": 0.2163, | |
| "step": 6725 | |
| }, | |
| { | |
| "epoch": 3.8834391229082517, | |
| "grad_norm": 0.46373171517071093, | |
| "learning_rate": 5.170197831168677e-08, | |
| "loss": 0.2072, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 3.8863242931332946, | |
| "grad_norm": 0.5108545524577744, | |
| "learning_rate": 4.917622537245925e-08, | |
| "loss": 0.1981, | |
| "step": 6735 | |
| }, | |
| { | |
| "epoch": 3.889209463358338, | |
| "grad_norm": 0.5134285782488883, | |
| "learning_rate": 4.671356930363691e-08, | |
| "loss": 0.2115, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 3.8920946335833815, | |
| "grad_norm": 0.5112672014805328, | |
| "learning_rate": 4.431402572060095e-08, | |
| "loss": 0.2067, | |
| "step": 6745 | |
| }, | |
| { | |
| "epoch": 3.8949798038084245, | |
| "grad_norm": 0.5093768129247295, | |
| "learning_rate": 4.197760983854271e-08, | |
| "loss": 0.2094, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 3.897864974033468, | |
| "grad_norm": 0.49789509988732483, | |
| "learning_rate": 3.970433647236926e-08, | |
| "loss": 0.2109, | |
| "step": 6755 | |
| }, | |
| { | |
| "epoch": 3.9007501442585113, | |
| "grad_norm": 0.45027822200576423, | |
| "learning_rate": 3.749422003661018e-08, | |
| "loss": 0.1992, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 3.9036353144835543, | |
| "grad_norm": 0.4956800653304501, | |
| "learning_rate": 3.5347274545322055e-08, | |
| "loss": 0.2255, | |
| "step": 6765 | |
| }, | |
| { | |
| "epoch": 3.9065204847085977, | |
| "grad_norm": 0.5131612404624672, | |
| "learning_rate": 3.3263513612006305e-08, | |
| "loss": 0.2123, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 3.909405654933641, | |
| "grad_norm": 0.4801277473215345, | |
| "learning_rate": 3.124295044951931e-08, | |
| "loss": 0.2064, | |
| "step": 6775 | |
| }, | |
| { | |
| "epoch": 3.912290825158684, | |
| "grad_norm": 0.47685091887520487, | |
| "learning_rate": 2.928559786998575e-08, | |
| "loss": 0.2026, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 3.9151759953837275, | |
| "grad_norm": 0.4948070608182228, | |
| "learning_rate": 2.7391468284722056e-08, | |
| "loss": 0.2015, | |
| "step": 6785 | |
| }, | |
| { | |
| "epoch": 3.918061165608771, | |
| "grad_norm": 0.49666575468745433, | |
| "learning_rate": 2.5560573704157543e-08, | |
| "loss": 0.2127, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 3.9209463358338144, | |
| "grad_norm": 0.517584053580229, | |
| "learning_rate": 2.379292573775338e-08, | |
| "loss": 0.2103, | |
| "step": 6795 | |
| }, | |
| { | |
| "epoch": 3.9238315060588573, | |
| "grad_norm": 0.4715466573891517, | |
| "learning_rate": 2.2088535593933756e-08, | |
| "loss": 0.2034, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 3.9267166762839008, | |
| "grad_norm": 0.48184559815395395, | |
| "learning_rate": 2.0447414080014826e-08, | |
| "loss": 0.2163, | |
| "step": 6805 | |
| }, | |
| { | |
| "epoch": 3.929601846508944, | |
| "grad_norm": 0.47730934367308564, | |
| "learning_rate": 1.8869571602132543e-08, | |
| "loss": 0.2055, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 3.932487016733987, | |
| "grad_norm": 0.509248139581449, | |
| "learning_rate": 1.735501816517937e-08, | |
| "loss": 0.2001, | |
| "step": 6815 | |
| }, | |
| { | |
| "epoch": 3.9353721869590306, | |
| "grad_norm": 0.5087728221195549, | |
| "learning_rate": 1.59037633727388e-08, | |
| "loss": 0.2079, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 3.938257357184074, | |
| "grad_norm": 0.4986147495940189, | |
| "learning_rate": 1.4515816427029816e-08, | |
| "loss": 0.2173, | |
| "step": 6825 | |
| }, | |
| { | |
| "epoch": 3.9411425274091174, | |
| "grad_norm": 0.5024656322502186, | |
| "learning_rate": 1.3191186128841404e-08, | |
| "loss": 0.1988, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 3.9440276976341604, | |
| "grad_norm": 0.49534516720622374, | |
| "learning_rate": 1.192988087748037e-08, | |
| "loss": 0.2081, | |
| "step": 6835 | |
| }, | |
| { | |
| "epoch": 3.946912867859204, | |
| "grad_norm": 0.479255083905943, | |
| "learning_rate": 1.073190867071805e-08, | |
| "loss": 0.2017, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 3.9497980380842472, | |
| "grad_norm": 0.4632454498993253, | |
| "learning_rate": 9.597277104739233e-09, | |
| "loss": 0.2044, | |
| "step": 6845 | |
| }, | |
| { | |
| "epoch": 3.95268320830929, | |
| "grad_norm": 0.5211397474253632, | |
| "learning_rate": 8.525993374095543e-09, | |
| "loss": 0.2058, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 3.9555683785343336, | |
| "grad_norm": 0.48574173728886655, | |
| "learning_rate": 7.518064271654357e-09, | |
| "loss": 0.215, | |
| "step": 6855 | |
| }, | |
| { | |
| "epoch": 3.958453548759377, | |
| "grad_norm": 0.4668619180516313, | |
| "learning_rate": 6.573496188565509e-09, | |
| "loss": 0.1986, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 3.96133871898442, | |
| "grad_norm": 0.47768436258947716, | |
| "learning_rate": 5.6922951142079904e-09, | |
| "loss": 0.1948, | |
| "step": 6865 | |
| }, | |
| { | |
| "epoch": 3.9642238892094634, | |
| "grad_norm": 0.5138599831276915, | |
| "learning_rate": 4.874466636164421e-09, | |
| "loss": 0.2113, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 3.967109059434507, | |
| "grad_norm": 0.5000185830706366, | |
| "learning_rate": 4.120015940177746e-09, | |
| "loss": 0.2256, | |
| "step": 6875 | |
| }, | |
| { | |
| "epoch": 3.96999422965955, | |
| "grad_norm": 0.4866192841385196, | |
| "learning_rate": 3.4289478101201536e-09, | |
| "loss": 0.2063, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 3.9728793998845933, | |
| "grad_norm": 0.47456211919715674, | |
| "learning_rate": 2.8012666279653155e-09, | |
| "loss": 0.2031, | |
| "step": 6885 | |
| }, | |
| { | |
| "epoch": 3.9757645701096367, | |
| "grad_norm": 0.4973916489751191, | |
| "learning_rate": 2.2369763737573046e-09, | |
| "loss": 0.2013, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 3.9786497403346797, | |
| "grad_norm": 0.50633061407573, | |
| "learning_rate": 1.7360806255861673e-09, | |
| "loss": 0.2108, | |
| "step": 6895 | |
| }, | |
| { | |
| "epoch": 3.981534910559723, | |
| "grad_norm": 0.4761645591299464, | |
| "learning_rate": 1.29858255956683e-09, | |
| "loss": 0.2145, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 3.9844200807847665, | |
| "grad_norm": 0.4667942094809914, | |
| "learning_rate": 9.244849498168951e-10, | |
| "loss": 0.1998, | |
| "step": 6905 | |
| }, | |
| { | |
| "epoch": 3.9873052510098095, | |
| "grad_norm": 0.4758585527520601, | |
| "learning_rate": 6.137901684399871e-10, | |
| "loss": 0.2103, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 3.990190421234853, | |
| "grad_norm": 0.5058399856877976, | |
| "learning_rate": 3.665001855113204e-10, | |
| "loss": 0.2037, | |
| "step": 6915 | |
| }, | |
| { | |
| "epoch": 3.9930755914598963, | |
| "grad_norm": 0.4840062336803948, | |
| "learning_rate": 1.8261656906437553e-10, | |
| "loss": 0.2114, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 3.9959607616849393, | |
| "grad_norm": 0.49250096214422046, | |
| "learning_rate": 6.21404850809082e-11, | |
| "loss": 0.2046, | |
| "step": 6925 | |
| }, | |
| { | |
| "epoch": 3.9988459319099827, | |
| "grad_norm": 0.521991079354229, | |
| "learning_rate": 5.0726974842874035e-12, | |
| "loss": 0.2225, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 0.5301879048347473, | |
| "eval_runtime": 0.5871, | |
| "eval_samples_per_second": 131.147, | |
| "eval_steps_per_second": 3.406, | |
| "step": 6932 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "step": 6932, | |
| "total_flos": 340525408321536.0, | |
| "train_loss": 0.3252074327502578, | |
| "train_runtime": 5343.4474, | |
| "train_samples_per_second": 41.505, | |
| "train_steps_per_second": 1.297 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 6932, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 4, | |
| "save_steps": 694, | |
| "total_flos": 340525408321536.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |